From 9b52f8d5b5131dd0a7e8f2069a25719daec51dbc Mon Sep 17 00:00:00 2001 From: zhulinJulia24 <145004780+zhulinJulia24@users.noreply.github.com> Date: Sat, 12 Oct 2024 17:06:19 +0800 Subject: [PATCH] [ci] use local requirements for test workflow (#2569) * update * update --- .github/workflows/benchmark.yml | 8 +------- .github/workflows/daily_ete_test.yml | 16 +++++----------- .github/workflows/evaluate.yml | 8 +------- .github/workflows/stable.yml | 8 +------- 4 files changed, 8 insertions(+), 32 deletions(-) diff --git a/.github/workflows/benchmark.yml b/.github/workflows/benchmark.yml index cf8283bbd..bd3876f9e 100644 --- a/.github/workflows/benchmark.yml +++ b/.github/workflows/benchmark.yml @@ -23,11 +23,6 @@ on: description: 'Whether start a offline mode, if true, you should prepare code and whl package by yourself' type: boolean default: false - dependency_pkgs: - required: true - description: 'Dependency packages, you can also set a specific version' - type: string - default: 'packaging transformers_stream_generator transformers datasets matplotlib jmespath' env: HOST_PIP_CACHE_DIR: /nvme/github-actions/pip-cache @@ -35,7 +30,6 @@ env: OUTPUT_FOLDER: cuda11.8_dist_${{ github.run_id }} REPORT_DIR: /nvme/qa_test_models/benchmark-reports/${{ github.run_id }} ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true - dependency_pkgs: ${{inputs.dependency_pkgs || 'packaging transformers_stream_generator transformers datasets matplotlib jmespath'}} FAIL_CONFIG: ${{ github.run_attempt != 1 && '--lf --lfnf none' || '--lf'}} jobs: @@ -115,7 +109,7 @@ jobs: python3 -m pip install -e /root/packages/AutoAWQ_kernels python3 -m pip install /root/packages/autoawq-0.2.6-cp310-cp310-manylinux2014_x86_64.whl --no-deps python3 -m pip install /root/packages/xformers-0.0.27+cu118-cp310-cp310-manylinux2014_x86_64.whl --no-deps - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | diff --git a/.github/workflows/daily_ete_test.yml b/.github/workflows/daily_ete_test.yml index 16202656e..1cb242a74 100644 --- a/.github/workflows/daily_ete_test.yml +++ b/.github/workflows/daily_ete_test.yml @@ -28,11 +28,6 @@ on: description: 'Whether start a offline mode, if true, you should prepare code and whl package by yourself' type: boolean default: false - dependency_pkgs: - required: true - description: 'Dependency packages, you can also set a specific version' - type: string - default: 'packaging transformers_stream_generator transformers datasets matplotlib openai attrdict timm modelscope jmespath decord auto_gptq qwen_vl_utils mmengine-lite==0.10.5' regression_func: required: true description: 'regression functions' @@ -43,7 +38,6 @@ on: env: HOST_PIP_CACHE_DIR: /nvme/github-actions/pip-cache - dependency_pkgs: ${{inputs.dependency_pkgs || 'packaging transformers_stream_generator transformers datasets matplotlib openai attrdict timm modelscope jmespath decord auto_gptq qwen_vl_utils mmengine-lite==0.10.5'}} HOST_LOCALTIME: /usr/share/zoneinfo/Asia/Shanghai OUTPUT_FOLDER: cuda11.8_dist_${{ github.run_id }} ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true @@ -132,7 +126,7 @@ jobs: python3 -m pip install -e /root/packages/AutoAWQ_kernels python3 -m pip install /root/packages/autoawq-0.2.6-cp310-cp310-manylinux2014_x86_64.whl --no-deps python3 -m pip install /root/packages/xformers-0.0.27+cu118-cp310-cp310-manylinux2014_x86_64.whl --no-deps - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | @@ -306,7 +300,7 @@ jobs: # manually install flash attn # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | @@ -410,7 +404,7 @@ jobs: # manually install flash attn # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | @@ -483,7 +477,7 @@ jobs: # manually install flash attn # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | @@ -559,7 +553,7 @@ jobs: # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl python3 -m pip install /root/packages/xformers-0.0.27+cu118-cp310-cp310-manylinux2014_x86_64.whl --no-deps - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | diff --git a/.github/workflows/evaluate.yml b/.github/workflows/evaluate.yml index 433f432ef..1e8d78d14 100644 --- a/.github/workflows/evaluate.yml +++ b/.github/workflows/evaluate.yml @@ -43,15 +43,9 @@ on: description: 'Whether start a offline mode, if true, you should prepare code and whl package by yourself' type: boolean default: false - dependency_pkgs: - required: true - description: 'Dependency packages, you can also set a specific version' - type: string - default: 'pynvml packaging protobuf transformers_stream_generator transformers human_eval mmengine-lite==0.10.5' env: ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true - dependency_pkgs: ${{inputs.dependency_pkgs || 'pynvml packaging protobuf transformers_stream_generator transformers human_eval mmengine-lite==0.10.5'}} jobs: linux-build: @@ -136,7 +130,7 @@ jobs: # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl python3 -m pip install /root/packages/xformers-0.0.27+cu118-cp310-cp310-manylinux2014_x86_64.whl --no-deps - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /root/models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: | diff --git a/.github/workflows/stable.yml b/.github/workflows/stable.yml index 97a9df782..85daed8e2 100644 --- a/.github/workflows/stable.yml +++ b/.github/workflows/stable.yml @@ -18,11 +18,6 @@ on: description: 'Whether start a offline mode, if true, you should prepare code and whl package by yourself' type: boolean default: false - dependency_pkgs: - required: true - description: 'Dependency packages, you can also set a specific version' - type: string - default: 'packaging transformers_stream_generator transformers datasets matplotlib jmespath mmengine-lite==0.10.5' schedule: - cron: '00 8 * * 1' @@ -32,7 +27,6 @@ env: OUTPUT_FOLDER: cuda11.8_dist_${{ github.run_id }} REPORT_DIR: /nvme/qa_test_models/stable_reports/${{ github.run_id }} ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true - dependency_pkgs: ${{inputs.dependency_pkgs || 'packaging transformers_stream_generator transformers datasets matplotlib jmespath mmengine-lite==0.10.5'}} COMPASS_DATA_CACHE: /nvme/qa_test_models/dataset jobs: @@ -110,7 +104,7 @@ jobs: # the install packeage from. https://github.com/Dao-AILab/flash-attention/releases python3 -m pip install /root/packages/flash_attn-2.6.3+cu118torch2.3cxx11abiFALSE-cp310-cp310-linux_x86_64.whl python3 -m pip install /root/packages/xformers-0.0.27+cu118-cp310-cp310-manylinux2014_x86_64.whl --no-deps - python3 -m pip install ${{env.dependency_pkgs}} + python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt - name: Install lmdeploy if: ${{github.event_name == 'schedule' || !inputs.offline_mode}} run: |