Spaces:
Runtime error
Runtime error
| name: Nightly and release tests on main/release branch | |
| on: | |
| workflow_dispatch: | |
| schedule: | |
| - cron: "0 0 * * *" # every day at midnight | |
| env: | |
| DIFFUSERS_IS_CI: yes | |
| HF_HUB_ENABLE_HF_TRANSFER: 1 | |
| OMP_NUM_THREADS: 8 | |
| MKL_NUM_THREADS: 8 | |
| PYTEST_TIMEOUT: 600 | |
| RUN_SLOW: yes | |
| RUN_NIGHTLY: yes | |
| PIPELINE_USAGE_CUTOFF: 0 | |
| SLACK_API_TOKEN: ${{ secrets.SLACK_CIFEEDBACK_BOT_TOKEN }} | |
| CONSOLIDATED_REPORT_PATH: consolidated_test_report.md | |
| jobs: | |
| setup_torch_cuda_pipeline_matrix: | |
| name: Setup Torch Pipelines CUDA Slow Tests Matrix | |
| runs-on: | |
| group: aws-general-8-plus | |
| container: | |
| image: diffusers/diffusers-pytorch-cpu | |
| outputs: | |
| pipeline_test_matrix: ${{ steps.fetch_pipeline_matrix.outputs.pipeline_test_matrix }} | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: Install dependencies | |
| run: | | |
| pip install -e .[test] | |
| pip install huggingface_hub | |
| - name: Fetch Pipeline Matrix | |
| id: fetch_pipeline_matrix | |
| run: | | |
| matrix=$(python utils/fetch_torch_cuda_pipeline_test_matrix.py) | |
| echo $matrix | |
| echo "pipeline_test_matrix=$matrix" >> $GITHUB_OUTPUT | |
| - name: Pipeline Tests Artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: test-pipelines.json | |
| path: reports | |
| run_nightly_tests_for_torch_pipelines: | |
| name: Nightly Torch Pipelines CUDA Tests | |
| needs: setup_torch_cuda_pipeline_matrix | |
| strategy: | |
| fail-fast: false | |
| max-parallel: 8 | |
| matrix: | |
| module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} | |
| runs-on: | |
| group: aws-g4dn-2xlarge | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --shm-size "16gb" --ipc host --gpus 0 | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: Pipeline CUDA Test | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -s -v -k "not Flax and not Onnx" \ | |
| --make-reports=tests_pipeline_${{ matrix.module }}_cuda \ | |
| --report-log=tests_pipeline_${{ matrix.module }}_cuda.log \ | |
| tests/pipelines/${{ matrix.module }} | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_pipeline_${{ matrix.module }}_cuda_stats.txt | |
| cat reports/tests_pipeline_${{ matrix.module }}_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: pipeline_${{ matrix.module }}_test_reports | |
| path: reports | |
| run_nightly_tests_for_other_torch_modules: | |
| name: Nightly Torch CUDA Tests | |
| runs-on: | |
| group: aws-g4dn-2xlarge | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --shm-size "16gb" --ipc host --gpus 0 | |
| defaults: | |
| run: | |
| shell: bash | |
| strategy: | |
| fail-fast: false | |
| max-parallel: 2 | |
| matrix: | |
| module: [models, schedulers, lora, others, single_file, examples] | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| python -m uv pip install peft@git+https://github.com/huggingface/peft.git | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: python utils/print_env.py | |
| - name: Run nightly PyTorch CUDA tests for non-pipeline modules | |
| if: ${{ matrix.module != 'examples'}} | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -s -v -k "not Flax and not Onnx" \ | |
| --make-reports=tests_torch_${{ matrix.module }}_cuda \ | |
| --report-log=tests_torch_${{ matrix.module }}_cuda.log \ | |
| tests/${{ matrix.module }} | |
| - name: Run nightly example tests with Torch | |
| if: ${{ matrix.module == 'examples' }} | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -s -v --make-reports=examples_torch_cuda \ | |
| --report-log=examples_torch_cuda.log \ | |
| examples/ | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_torch_${{ matrix.module }}_cuda_stats.txt | |
| cat reports/tests_torch_${{ matrix.module }}_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_${{ matrix.module }}_cuda_test_reports | |
| path: reports | |
| run_torch_compile_tests: | |
| name: PyTorch Compile CUDA tests | |
| runs-on: | |
| group: aws-g4dn-2xlarge | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --gpus 0 --shm-size "16gb" --ipc host | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: | | |
| nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test,training] | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: Run torch compile tests on GPU | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| RUN_COMPILE: yes | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v -k "compile" --make-reports=tests_torch_compile_cuda tests/ | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: cat reports/tests_torch_compile_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_compile_test_reports | |
| path: reports | |
| run_big_gpu_torch_tests: | |
| name: Torch tests on big GPU | |
| strategy: | |
| fail-fast: false | |
| max-parallel: 2 | |
| runs-on: | |
| group: aws-g6e-xlarge-plus | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --shm-size "16gb" --ipc host --gpus 0 | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| python -m uv pip install peft@git+https://github.com/huggingface/peft.git | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: Selected Torch CUDA Test on big GPU | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| BIG_GPU_MEMORY: 40 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -m "big_gpu_with_torch_cuda" \ | |
| --make-reports=tests_big_gpu_torch_cuda \ | |
| --report-log=tests_big_gpu_torch_cuda.log \ | |
| tests/ | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_big_gpu_torch_cuda_stats.txt | |
| cat reports/tests_big_gpu_torch_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_cuda_big_gpu_test_reports | |
| path: reports | |
| torch_minimum_version_cuda_tests: | |
| name: Torch Minimum Version CUDA Tests | |
| runs-on: | |
| group: aws-g4dn-2xlarge | |
| container: | |
| image: diffusers/diffusers-pytorch-minimum-cuda | |
| options: --shm-size "16gb" --ipc host --gpus 0 | |
| defaults: | |
| run: | |
| shell: bash | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| python -m uv pip install peft@git+https://github.com/huggingface/peft.git | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: Run PyTorch CUDA tests | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -s -v -k "not Flax and not Onnx" \ | |
| --make-reports=tests_torch_minimum_version_cuda \ | |
| tests/models/test_modeling_common.py \ | |
| tests/pipelines/test_pipelines_common.py \ | |
| tests/pipelines/test_pipeline_utils.py \ | |
| tests/pipelines/test_pipelines.py \ | |
| tests/pipelines/test_pipelines_auto.py \ | |
| tests/schedulers/test_schedulers.py \ | |
| tests/others | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_torch_minimum_version_cuda_stats.txt | |
| cat reports/tests_torch_minimum_version_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_minimum_version_cuda_test_reports | |
| path: reports | |
| run_nightly_onnx_tests: | |
| name: Nightly ONNXRuntime CUDA tests on Ubuntu | |
| runs-on: | |
| group: aws-g4dn-2xlarge | |
| container: | |
| image: diffusers/diffusers-onnxruntime-cuda | |
| options: --gpus 0 --shm-size "16gb" --ipc host | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: python utils/print_env.py | |
| - name: Run Nightly ONNXRuntime CUDA tests | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| -s -v -k "Onnx" \ | |
| --make-reports=tests_onnx_cuda \ | |
| --report-log=tests_onnx_cuda.log \ | |
| tests/ | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_onnx_cuda_stats.txt | |
| cat reports/tests_onnx_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: tests_onnx_cuda_reports | |
| path: reports | |
| run_nightly_quantization_tests: | |
| name: Torch quantization nightly tests | |
| strategy: | |
| fail-fast: false | |
| max-parallel: 2 | |
| matrix: | |
| config: | |
| - backend: "bitsandbytes" | |
| test_location: "bnb" | |
| additional_deps: ["peft"] | |
| - backend: "gguf" | |
| test_location: "gguf" | |
| additional_deps: ["peft"] | |
| - backend: "torchao" | |
| test_location: "torchao" | |
| additional_deps: [] | |
| - backend: "optimum_quanto" | |
| test_location: "quanto" | |
| additional_deps: [] | |
| runs-on: | |
| group: aws-g6e-xlarge-plus | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --shm-size "20gb" --ipc host --gpus 0 | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| python -m uv pip install -U ${{ matrix.config.backend }} | |
| if [ "${{ join(matrix.config.additional_deps, ' ') }}" != "" ]; then | |
| python -m uv pip install ${{ join(matrix.config.additional_deps, ' ') }} | |
| fi | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: ${{ matrix.config.backend }} quantization tests on GPU | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| BIG_GPU_MEMORY: 40 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| --make-reports=tests_${{ matrix.config.backend }}_torch_cuda \ | |
| --report-log=tests_${{ matrix.config.backend }}_torch_cuda.log \ | |
| tests/quantization/${{ matrix.config.test_location }} | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_${{ matrix.config.backend }}_torch_cuda_stats.txt | |
| cat reports/tests_${{ matrix.config.backend }}_torch_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_cuda_${{ matrix.config.backend }}_reports | |
| path: reports | |
| run_nightly_pipeline_level_quantization_tests: | |
| name: Torch quantization nightly tests | |
| strategy: | |
| fail-fast: false | |
| max-parallel: 2 | |
| runs-on: | |
| group: aws-g6e-xlarge-plus | |
| container: | |
| image: diffusers/diffusers-pytorch-cuda | |
| options: --shm-size "20gb" --ipc host --gpus 0 | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: NVIDIA-SMI | |
| run: nvidia-smi | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| python -m uv pip install -U bitsandbytes optimum_quanto | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: | | |
| python utils/print_env.py | |
| - name: Pipeline-level quantization tests on GPU | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
| CUBLAS_WORKSPACE_CONFIG: :16:8 | |
| BIG_GPU_MEMORY: 40 | |
| run: | | |
| python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
| --make-reports=tests_pipeline_level_quant_torch_cuda \ | |
| --report-log=tests_pipeline_level_quant_torch_cuda.log \ | |
| tests/quantization/test_pipeline_level_quantization.py | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_pipeline_level_quant_torch_cuda_stats.txt | |
| cat reports/tests_pipeline_level_quant_torch_cuda_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: torch_cuda_pipeline_level_quant_reports | |
| path: reports | |
| run_flax_tpu_tests: | |
| name: Nightly Flax TPU Tests | |
| runs-on: | |
| group: gcp-ct5lp-hightpu-8t | |
| if: github.event_name == 'schedule' | |
| container: | |
| image: diffusers/diffusers-flax-tpu | |
| options: --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} -v /mnt/hf_cache:/mnt/hf_cache | |
| defaults: | |
| run: | |
| shell: bash | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: Install dependencies | |
| run: | | |
| python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
| python -m uv pip install -e [quality,test] | |
| pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git | |
| python -m uv pip install pytest-reportlog | |
| - name: Environment | |
| run: python utils/print_env.py | |
| - name: Run nightly Flax TPU tests | |
| env: | |
| HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| run: | | |
| python -m pytest -n 0 \ | |
| -s -v -k "Flax" \ | |
| --make-reports=tests_flax_tpu \ | |
| --report-log=tests_flax_tpu.log \ | |
| tests/ | |
| - name: Failure short reports | |
| if: ${{ failure() }} | |
| run: | | |
| cat reports/tests_flax_tpu_stats.txt | |
| cat reports/tests_flax_tpu_failures_short.txt | |
| - name: Test suite reports artifacts | |
| if: ${{ always() }} | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: flax_tpu_test_reports | |
| path: reports | |
| generate_consolidated_report: | |
| name: Generate Consolidated Test Report | |
| needs: [ | |
| run_nightly_tests_for_torch_pipelines, | |
| run_nightly_tests_for_other_torch_modules, | |
| run_torch_compile_tests, | |
| run_big_gpu_torch_tests, | |
| run_nightly_quantization_tests, | |
| run_nightly_pipeline_level_quantization_tests, | |
| run_nightly_onnx_tests, | |
| torch_minimum_version_cuda_tests, | |
| run_flax_tpu_tests | |
| ] | |
| if: always() | |
| runs-on: | |
| group: aws-general-8-plus | |
| container: | |
| image: diffusers/diffusers-pytorch-cpu | |
| steps: | |
| - name: Checkout diffusers | |
| uses: actions/checkout@v3 | |
| with: | |
| fetch-depth: 2 | |
| - name: Create reports directory | |
| run: mkdir -p combined_reports | |
| - name: Download all test reports | |
| uses: actions/download-artifact@v4 | |
| with: | |
| path: artifacts | |
| - name: Prepare reports | |
| run: | | |
| # Move all report files to a single directory for processing | |
| find artifacts -name "*.txt" -exec cp {} combined_reports/ \; | |
| - name: Install dependencies | |
| run: | | |
| pip install -e .[test] | |
| pip install slack_sdk tabulate | |
| - name: Generate consolidated report | |
| run: | | |
| python utils/consolidated_test_report.py \ | |
| --reports_dir combined_reports \ | |
| --output_file $CONSOLIDATED_REPORT_PATH \ | |
| --slack_channel_name diffusers-ci-nightly | |
| - name: Show consolidated report | |
| run: | | |
| cat $CONSOLIDATED_REPORT_PATH >> $GITHUB_STEP_SUMMARY | |
| - name: Upload consolidated report | |
| uses: actions/upload-artifact@v4 | |
| with: | |
| name: consolidated_test_report | |
| path: ${{ env.CONSOLIDATED_REPORT_PATH }} | |
| # M1 runner currently not well supported | |
| # TODO: (Dhruv) add these back when we setup better testing for Apple Silicon | |
| # run_nightly_tests_apple_m1: | |
| # name: Nightly PyTorch MPS tests on MacOS | |
| # runs-on: [ self-hosted, apple-m1 ] | |
| # if: github.event_name == 'schedule' | |
| # | |
| # steps: | |
| # - name: Checkout diffusers | |
| # uses: actions/checkout@v3 | |
| # with: | |
| # fetch-depth: 2 | |
| # | |
| # - name: Clean checkout | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # git clean -fxd | |
| # - name: Setup miniconda | |
| # uses: ./.github/actions/setup-miniconda | |
| # with: | |
| # python-version: 3.9 | |
| # | |
| # - name: Install dependencies | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # ${CONDA_RUN} python -m pip install --upgrade pip uv | |
| # ${CONDA_RUN} python -m uv pip install -e [quality,test] | |
| # ${CONDA_RUN} python -m uv pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cpu | |
| # ${CONDA_RUN} python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate | |
| # ${CONDA_RUN} python -m uv pip install pytest-reportlog | |
| # - name: Environment | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # ${CONDA_RUN} python utils/print_env.py | |
| # - name: Run nightly PyTorch tests on M1 (MPS) | |
| # shell: arch -arch arm64 bash {0} | |
| # env: | |
| # HF_HOME: /System/Volumes/Data/mnt/cache | |
| # HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # run: | | |
| # ${CONDA_RUN} python -m pytest -n 1 -s -v --make-reports=tests_torch_mps \ | |
| # --report-log=tests_torch_mps.log \ | |
| # tests/ | |
| # - name: Failure short reports | |
| # if: ${{ failure() }} | |
| # run: cat reports/tests_torch_mps_failures_short.txt | |
| # | |
| # - name: Test suite reports artifacts | |
| # if: ${{ always() }} | |
| # uses: actions/upload-artifact@v4 | |
| # with: | |
| # name: torch_mps_test_reports | |
| # path: reports | |
| # | |
| # - name: Generate Report and Notify Channel | |
| # if: always() | |
| # run: | | |
| # pip install slack_sdk tabulate | |
| # python utils/log_reports.py >> $GITHUB_STEP_SUMMARY run_nightly_tests_apple_m1: | |
| # name: Nightly PyTorch MPS tests on MacOS | |
| # runs-on: [ self-hosted, apple-m1 ] | |
| # if: github.event_name == 'schedule' | |
| # | |
| # steps: | |
| # - name: Checkout diffusers | |
| # uses: actions/checkout@v3 | |
| # with: | |
| # fetch-depth: 2 | |
| # | |
| # - name: Clean checkout | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # git clean -fxd | |
| # - name: Setup miniconda | |
| # uses: ./.github/actions/setup-miniconda | |
| # with: | |
| # python-version: 3.9 | |
| # | |
| # - name: Install dependencies | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # ${CONDA_RUN} python -m pip install --upgrade pip uv | |
| # ${CONDA_RUN} python -m uv pip install -e [quality,test] | |
| # ${CONDA_RUN} python -m uv pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cpu | |
| # ${CONDA_RUN} python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate | |
| # ${CONDA_RUN} python -m uv pip install pytest-reportlog | |
| # - name: Environment | |
| # shell: arch -arch arm64 bash {0} | |
| # run: | | |
| # ${CONDA_RUN} python utils/print_env.py | |
| # - name: Run nightly PyTorch tests on M1 (MPS) | |
| # shell: arch -arch arm64 bash {0} | |
| # env: | |
| # HF_HOME: /System/Volumes/Data/mnt/cache | |
| # HF_TOKEN: ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }} | |
| # run: | | |
| # ${CONDA_RUN} python -m pytest -n 1 -s -v --make-reports=tests_torch_mps \ | |
| # --report-log=tests_torch_mps.log \ | |
| # tests/ | |
| # - name: Failure short reports | |
| # if: ${{ failure() }} | |
| # run: cat reports/tests_torch_mps_failures_short.txt | |
| # | |
| # - name: Test suite reports artifacts | |
| # if: ${{ always() }} | |
| # uses: actions/upload-artifact@v4 | |
| # with: | |
| # name: torch_mps_test_reports | |
| # path: reports | |
| # | |
| # - name: Generate Report and Notify Channel | |
| # if: always() | |
| # run: | | |
| # pip install slack_sdk tabulate | |
| # python utils/log_reports.py >> $GITHUB_STEP_SUMMARY | |