mirror of
https://gitee.com/infiniflow/ragflow.git
synced 2025-12-06 15:29:03 +08:00
### What problem does this PR solve? As title ### Type of change - [x] Other (please describe): --------- Signed-off-by: Jin Hai <haijin.chn@gmail.com>
280 lines
14 KiB
YAML
280 lines
14 KiB
YAML
name: tests
|
|
|
|
on:
|
|
push:
|
|
branches:
|
|
- 'main'
|
|
- '*.*.*'
|
|
paths-ignore:
|
|
- 'docs/**'
|
|
- '*.md'
|
|
- '*.mdx'
|
|
# The only difference between pull_request and pull_request_target is the context in which the workflow runs:
|
|
# — pull_request_target workflows use the workflow files from the default branch, and secrets are available.
|
|
# — pull_request workflows use the workflow files from the pull request branch, and secrets are unavailable.
|
|
pull_request:
|
|
types: [ synchronize, ready_for_review ]
|
|
paths-ignore:
|
|
- 'docs/**'
|
|
- '*.md'
|
|
- '*.mdx'
|
|
schedule:
|
|
- cron: '0 16 * * *' # This schedule runs every 16:00:00Z(00:00:00+08:00)
|
|
|
|
# https://docs.github.com/en/actions/using-jobs/using-concurrency
|
|
concurrency:
|
|
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
|
|
cancel-in-progress: true
|
|
|
|
jobs:
|
|
ragflow_tests:
|
|
name: ragflow_tests
|
|
# https://docs.github.com/en/actions/using-jobs/using-conditions-to-control-job-execution
|
|
# https://github.com/orgs/community/discussions/26261
|
|
if: ${{ github.event_name != 'pull_request' || (github.event.pull_request.draft == false && contains(github.event.pull_request.labels.*.name, 'ci')) }}
|
|
runs-on: [ "self-hosted", "ragflow-test" ]
|
|
steps:
|
|
# https://github.com/hmarr/debug-action
|
|
#- uses: hmarr/debug-action@v2
|
|
|
|
- name: Ensure workspace ownership
|
|
run: |
|
|
echo "Workflow triggered by ${{ github.event_name }}"
|
|
echo "chown -R ${USER} ${GITHUB_WORKSPACE}" && sudo chown -R ${USER} ${GITHUB_WORKSPACE}
|
|
|
|
# https://github.com/actions/checkout/issues/1781
|
|
- name: Check out code
|
|
uses: actions/checkout@v4
|
|
with:
|
|
ref: ${{ (github.event_name == 'pull_request' || github.event_name == 'pull_request_target') && format('refs/pull/{0}/merge', github.event.pull_request.number) || github.sha }}
|
|
fetch-depth: 0
|
|
fetch-tags: true
|
|
|
|
- name: Check workflow duplication
|
|
if: ${{ !cancelled() && !failure() }}
|
|
run: |
|
|
if [[ ${GITHUB_EVENT_NAME} != "pull_request" && ${GITHUB_EVENT_NAME} != "schedule" ]]; then
|
|
HEAD=$(git rev-parse HEAD)
|
|
# Find a PR that introduced a given commit
|
|
gh auth login --with-token <<< "${{ secrets.GITHUB_TOKEN }}"
|
|
PR_NUMBER=$(gh pr list --search ${HEAD} --state merged --json number --jq .[0].number)
|
|
echo "HEAD=${HEAD}"
|
|
echo "PR_NUMBER=${PR_NUMBER}"
|
|
if [[ -n "${PR_NUMBER}" ]]; then
|
|
PR_SHA_FP=${RUNNER_WORKSPACE_PREFIX}/artifacts/${GITHUB_REPOSITORY}/PR_${PR_NUMBER}
|
|
if [[ -f "${PR_SHA_FP}" ]]; then
|
|
read -r PR_SHA PR_RUN_ID < "${PR_SHA_FP}"
|
|
# Calculate the hash of the current workspace content
|
|
HEAD_SHA=$(git rev-parse HEAD^{tree})
|
|
if [[ "${HEAD_SHA}" == "${PR_SHA}" ]]; then
|
|
echo "Cancel myself since the workspace content hash is the same with PR #${PR_NUMBER} merged. See ${GITHUB_SERVER_URL}/${GITHUB_REPOSITORY}/actions/runs/${PR_RUN_ID} for details."
|
|
gh run cancel ${GITHUB_RUN_ID}
|
|
while true; do
|
|
status=$(gh run view ${GITHUB_RUN_ID} --json status -q .status)
|
|
[ "${status}" = "completed" ] && break
|
|
sleep 5
|
|
done
|
|
exit 1
|
|
fi
|
|
fi
|
|
fi
|
|
elif [[ ${GITHUB_EVENT_NAME} == "pull_request" ]]; then
|
|
PR_NUMBER=${{ github.event.pull_request.number }}
|
|
PR_SHA_FP=${RUNNER_WORKSPACE_PREFIX}/artifacts/${GITHUB_REPOSITORY}/PR_${PR_NUMBER}
|
|
# Calculate the hash of the current workspace content
|
|
PR_SHA=$(git rev-parse HEAD^{tree})
|
|
echo "PR #${PR_NUMBER} workspace content hash: ${PR_SHA}"
|
|
mkdir -p ${RUNNER_WORKSPACE_PREFIX}/artifacts/${GITHUB_REPOSITORY}
|
|
echo "${PR_SHA} ${GITHUB_RUN_ID}" > ${PR_SHA_FP}
|
|
fi
|
|
|
|
# https://github.com/astral-sh/ruff-action
|
|
- name: Static check with Ruff
|
|
uses: astral-sh/ruff-action@v3
|
|
with:
|
|
version: ">=0.11.x"
|
|
args: "check"
|
|
|
|
- name: Check comments of changed Python files
|
|
if: ${{ false }}
|
|
run: |
|
|
if [[ ${{ github.event_name }} == 'pull_request' || ${{ github.event_name }} == 'pull_request_target' ]]; then
|
|
CHANGED_FILES=$(git diff --name-only ${{ github.event.pull_request.base.sha }}...${{ github.event.pull_request.head.sha }} \
|
|
| grep -E '\.(py)$' || true)
|
|
|
|
if [ -n "$CHANGED_FILES" ]; then
|
|
echo "Check comments of changed Python files with check_comment_ascii.py"
|
|
|
|
readarray -t files <<< "$CHANGED_FILES"
|
|
HAS_ERROR=0
|
|
|
|
for file in "${files[@]}"; do
|
|
if [ -f "$file" ]; then
|
|
if python3 check_comment_ascii.py "$file"; then
|
|
echo "✅ $file"
|
|
else
|
|
echo "❌ $file"
|
|
HAS_ERROR=1
|
|
fi
|
|
fi
|
|
done
|
|
|
|
if [ $HAS_ERROR -ne 0 ]; then
|
|
exit 1
|
|
fi
|
|
else
|
|
echo "No Python files changed"
|
|
fi
|
|
fi
|
|
|
|
- name: Run unit test
|
|
run: |
|
|
uv sync --python 3.10 --group test --frozen
|
|
source .venv/bin/activate
|
|
which pytest || echo "pytest not in PATH"
|
|
echo "Start to run unit test"
|
|
python3 run_tests.py
|
|
|
|
- name: Build ragflow:nightly
|
|
run: |
|
|
RUNNER_WORKSPACE_PREFIX=${RUNNER_WORKSPACE_PREFIX:-${HOME}}
|
|
RAGFLOW_IMAGE=infiniflow/ragflow:${GITHUB_RUN_ID}
|
|
echo "RAGFLOW_IMAGE=${RAGFLOW_IMAGE}" >> ${GITHUB_ENV}
|
|
sudo docker pull ubuntu:22.04
|
|
sudo DOCKER_BUILDKIT=1 docker build --build-arg NEED_MIRROR=1 -f Dockerfile -t ${RAGFLOW_IMAGE} .
|
|
if [[ ${GITHUB_EVENT_NAME} == "schedule" ]]; then
|
|
export HTTP_API_TEST_LEVEL=p3
|
|
else
|
|
export HTTP_API_TEST_LEVEL=p2
|
|
fi
|
|
echo "HTTP_API_TEST_LEVEL=${HTTP_API_TEST_LEVEL}" >> ${GITHUB_ENV}
|
|
echo "RAGFLOW_CONTAINER=${GITHUB_RUN_ID}-ragflow-cpu-1" >> ${GITHUB_ENV}
|
|
|
|
- name: Start ragflow:nightly
|
|
run: |
|
|
# Determine runner number (default to 1 if not found)
|
|
RUNNER_NUM=$(sudo docker inspect $(hostname) --format '{{index .Config.Labels "com.docker.compose.container-number"}}' 2>/dev/null || true)
|
|
RUNNER_NUM=${RUNNER_NUM:-1}
|
|
|
|
# Compute port numbers using bash arithmetic
|
|
ES_PORT=$((1200 + RUNNER_NUM * 10))
|
|
OS_PORT=$((1201 + RUNNER_NUM * 10))
|
|
INFINITY_THRIFT_PORT=$((23817 + RUNNER_NUM * 10))
|
|
INFINITY_HTTP_PORT=$((23820 + RUNNER_NUM * 10))
|
|
INFINITY_PSQL_PORT=$((5432 + RUNNER_NUM * 10))
|
|
MYSQL_PORT=$((5455 + RUNNER_NUM * 10))
|
|
MINIO_PORT=$((9000 + RUNNER_NUM * 10))
|
|
MINIO_CONSOLE_PORT=$((9001 + RUNNER_NUM * 10))
|
|
REDIS_PORT=$((6379 + RUNNER_NUM * 10))
|
|
TEI_PORT=$((6380 + RUNNER_NUM * 10))
|
|
KIBANA_PORT=$((6601 + RUNNER_NUM * 10))
|
|
SVR_HTTP_PORT=$((9380 + RUNNER_NUM * 10))
|
|
ADMIN_SVR_HTTP_PORT=$((9381 + RUNNER_NUM * 10))
|
|
SVR_MCP_PORT=$((9382 + RUNNER_NUM * 10))
|
|
SANDBOX_EXECUTOR_MANAGER_PORT=$((9385 + RUNNER_NUM * 10))
|
|
SVR_WEB_HTTP_PORT=$((80 + RUNNER_NUM * 10))
|
|
SVR_WEB_HTTPS_PORT=$((443 + RUNNER_NUM * 10))
|
|
|
|
# Persist computed ports into docker/.env so docker-compose uses the correct host bindings
|
|
echo "" >> docker/.env
|
|
echo -e "ES_PORT=${ES_PORT}" >> docker/.env
|
|
echo -e "OS_PORT=${OS_PORT}" >> docker/.env
|
|
echo -e "INFINITY_THRIFT_PORT=${INFINITY_THRIFT_PORT}" >> docker/.env
|
|
echo -e "INFINITY_HTTP_PORT=${INFINITY_HTTP_PORT}" >> docker/.env
|
|
echo -e "INFINITY_PSQL_PORT=${INFINITY_PSQL_PORT}" >> docker/.env
|
|
echo -e "MYSQL_PORT=${MYSQL_PORT}" >> docker/.env
|
|
echo -e "MINIO_PORT=${MINIO_PORT}" >> docker/.env
|
|
echo -e "MINIO_CONSOLE_PORT=${MINIO_CONSOLE_PORT}" >> docker/.env
|
|
echo -e "REDIS_PORT=${REDIS_PORT}" >> docker/.env
|
|
echo -e "TEI_PORT=${TEI_PORT}" >> docker/.env
|
|
echo -e "KIBANA_PORT=${KIBANA_PORT}" >> docker/.env
|
|
echo -e "SVR_HTTP_PORT=${SVR_HTTP_PORT}" >> docker/.env
|
|
echo -e "ADMIN_SVR_HTTP_PORT=${ADMIN_SVR_HTTP_PORT}" >> docker/.env
|
|
echo -e "SVR_MCP_PORT=${SVR_MCP_PORT}" >> docker/.env
|
|
echo -e "SANDBOX_EXECUTOR_MANAGER_PORT=${SANDBOX_EXECUTOR_MANAGER_PORT}" >> docker/.env
|
|
echo -e "SVR_WEB_HTTP_PORT=${SVR_WEB_HTTP_PORT}" >> docker/.env
|
|
echo -e "SVR_WEB_HTTPS_PORT=${SVR_WEB_HTTPS_PORT}" >> docker/.env
|
|
|
|
echo -e "COMPOSE_PROFILES=\${COMPOSE_PROFILES},tei-cpu" >> docker/.env
|
|
echo -e "TEI_MODEL=BAAI/bge-small-en-v1.5" >> docker/.env
|
|
echo -e "RAGFLOW_IMAGE=${RAGFLOW_IMAGE}" >> docker/.env
|
|
echo "HOST_ADDRESS=http://host.docker.internal:${SVR_HTTP_PORT}" >> ${GITHUB_ENV}
|
|
|
|
sudo docker compose -f docker/docker-compose.yml -p ${GITHUB_RUN_ID} up -d
|
|
uv sync --python 3.10 --only-group test --no-default-groups --frozen && uv pip install sdk/python --group test
|
|
|
|
- name: Run sdk tests against Elasticsearch
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && pytest -s --tb=short --level=${HTTP_API_TEST_LEVEL} test/testcases/test_sdk_api
|
|
|
|
- name: Run frontend api tests against Elasticsearch
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && pytest -s --tb=short sdk/python/test/test_frontend_api/get_email.py sdk/python/test/test_frontend_api/test_dataset.py
|
|
|
|
- name: Run http api tests against Elasticsearch
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && pytest -s --tb=short --level=${HTTP_API_TEST_LEVEL} test/testcases/test_http_api
|
|
|
|
- name: Stop ragflow:nightly
|
|
if: always() # always run this step even if previous steps failed
|
|
run: |
|
|
sudo docker compose -f docker/docker-compose.yml -p ${GITHUB_RUN_ID} down -v || true
|
|
sudo docker ps -a --filter "label=com.docker.compose.project=${GITHUB_RUN_ID}" -q | xargs -r sudo docker rm -f
|
|
|
|
- name: Start ragflow:nightly
|
|
run: |
|
|
sed -i '1i DOC_ENGINE=infinity' docker/.env
|
|
sudo docker compose -f docker/docker-compose.yml -p ${GITHUB_RUN_ID} up -d
|
|
|
|
- name: Run sdk tests against Infinity
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && DOC_ENGINE=infinity pytest -s --tb=short --level=${HTTP_API_TEST_LEVEL} test/testcases/test_sdk_api
|
|
|
|
- name: Run frontend api tests against Infinity
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && DOC_ENGINE=infinity pytest -s --tb=short sdk/python/test/test_frontend_api/get_email.py sdk/python/test/test_frontend_api/test_dataset.py
|
|
|
|
- name: Run http api tests against Infinity
|
|
run: |
|
|
export http_proxy=""; export https_proxy=""; export no_proxy=""; export HTTP_PROXY=""; export HTTPS_PROXY=""; export NO_PROXY=""
|
|
until sudo docker exec ${RAGFLOW_CONTAINER} curl -s --connect-timeout 5 ${HOST_ADDRESS} > /dev/null; do
|
|
echo "Waiting for service to be available..."
|
|
sleep 5
|
|
done
|
|
source .venv/bin/activate && DOC_ENGINE=infinity pytest -s --tb=short --level=${HTTP_API_TEST_LEVEL} test/testcases/test_http_api
|
|
|
|
- name: Stop ragflow:nightly
|
|
if: always() # always run this step even if previous steps failed
|
|
run: |
|
|
# Sometimes `docker compose down` fail due to hang container, heavy load etc. Need to remove such containers to release resources(for example, listen ports).
|
|
sudo docker compose -f docker/docker-compose.yml -p ${GITHUB_RUN_ID} down -v || true
|
|
sudo docker ps -a --filter "label=com.docker.compose.project=${GITHUB_RUN_ID}" -q | xargs -r sudo docker rm -f
|
|
if [[ -n ${RAGFLOW_IMAGE} ]]; then
|
|
sudo docker rmi -f ${RAGFLOW_IMAGE}
|
|
fi
|