Merge pull request #143 from huggingface/main #179
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Slow Tests on main | |
on: | |
push: | |
branches: | |
- main | |
env: | |
DIFFUSERS_IS_CI: yes | |
HF_HOME: /mnt/cache | |
OMP_NUM_THREADS: 8 | |
MKL_NUM_THREADS: 8 | |
PYTEST_TIMEOUT: 600 | |
RUN_SLOW: yes | |
PIPELINE_USAGE_CUTOFF: 50000 | |
jobs: | |
setup_torch_cuda_pipeline_matrix: | |
name: Setup Torch Pipelines CUDA Slow Tests Matrix | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-cpu # this is a CPU image, but we need it to fetch the matrix | |
options: --shm-size "16gb" --ipc host | |
outputs: | |
pipeline_test_matrix: ${{ steps.fetch_pipeline_matrix.outputs.pipeline_test_matrix }} | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Fetch Pipeline Matrix | |
id: fetch_pipeline_matrix | |
run: | | |
matrix=$(python utils/fetch_torch_cuda_pipeline_test_matrix.py) | |
echo $matrix | |
echo "pipeline_test_matrix=$matrix" >> $GITHUB_OUTPUT | |
- name: Pipeline Tests Artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: test-pipelines.json | |
path: reports | |
torch_pipelines_cuda_tests: | |
name: Torch Pipelines CUDA Slow Tests | |
needs: setup_torch_cuda_pipeline_matrix | |
strategy: | |
fail-fast: false | |
max-parallel: 1 | |
matrix: | |
module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: | | |
nvidia-smi | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Slow PyTorch CUDA checkpoint tests on Ubuntu | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx" \ | |
--make-reports=tests_pipeline_${{ matrix.module }}_cuda \ | |
tests/pipelines/${{ matrix.module }} | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_pipeline_${{ matrix.module }}_cuda_stats.txt | |
cat reports/tests_pipeline_${{ matrix.module }}_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: pipeline_${{ matrix.module }}_test_reports | |
path: reports | |
torch_cuda_tests: | |
name: Torch CUDA Tests | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
defaults: | |
run: | |
shell: bash | |
strategy: | |
matrix: | |
module: [models, schedulers, lora, others] | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run slow PyTorch CUDA tests | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx" \ | |
--make-reports=tests_torch_cuda \ | |
tests/${{ matrix.module }} | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_torch_cuda_stats.txt | |
cat reports/tests_torch_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_cuda_test_reports | |
path: reports | |
peft_cuda_tests: | |
name: PEFT CUDA Tests | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
defaults: | |
run: | |
shell: bash | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
python -m pip install git+https://github.com/huggingface/peft.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run slow PEFT CUDA tests | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx and not PEFTLoRALoading" \ | |
--make-reports=tests_peft_cuda \ | |
tests/lora/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_peft_cuda_stats.txt | |
cat reports/tests_peft_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_peft_test_reports | |
path: reports | |
flax_tpu_tests: | |
name: Flax TPU Tests | |
runs-on: docker-tpu | |
container: | |
image: diffusers/diffusers-flax-tpu | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --privileged | |
defaults: | |
run: | |
shell: bash | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run slow Flax TPU tests | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
run: | | |
python -m pytest -n 0 \ | |
-s -v -k "Flax" \ | |
--make-reports=tests_flax_tpu \ | |
tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_flax_tpu_stats.txt | |
cat reports/tests_flax_tpu_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: flax_tpu_test_reports | |
path: reports | |
onnx_cuda_tests: | |
name: ONNX CUDA Tests | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-onnxruntime-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
defaults: | |
run: | |
shell: bash | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
apt-get update && apt-get install libsndfile1-dev libgl1 -y | |
python -m pip install -e .[quality,test] | |
python -m pip install git+https://github.com/huggingface/accelerate.git | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run slow ONNXRuntime CUDA tests | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "Onnx" \ | |
--make-reports=tests_onnx_cuda \ | |
tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_onnx_cuda_stats.txt | |
cat reports/tests_onnx_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: onnx_cuda_test_reports | |
path: reports | |
run_torch_compile_tests: | |
name: PyTorch Compile CUDA tests | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-compile-cuda | |
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: | | |
nvidia-smi | |
- name: Install dependencies | |
run: | | |
python -m pip install -e .[quality,test,training] | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run example tests on GPU | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v -k "compile" --make-reports=tests_torch_compile_cuda tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: cat reports/tests_torch_compile_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_compile_test_reports | |
path: reports | |
run_xformers_tests: | |
name: PyTorch xformers CUDA tests | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-xformers-cuda | |
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: | | |
nvidia-smi | |
- name: Install dependencies | |
run: | | |
python -m pip install -e .[quality,test,training] | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run example tests on GPU | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v -k "xformers" --make-reports=tests_torch_xformers_cuda tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: cat reports/tests_torch_xformers_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_xformers_test_reports | |
path: reports | |
run_examples_tests: | |
name: Examples PyTorch CUDA tests on Ubuntu | |
runs-on: docker-gpu | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: | | |
nvidia-smi | |
- name: Install dependencies | |
run: | | |
python -m pip install -e .[quality,test,training] | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Run example tests on GPU | |
env: | |
HUGGING_FACE_HUB_TOKEN: ${{ secrets.HUGGING_FACE_HUB_TOKEN }} | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v --make-reports=examples_torch_cuda examples/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/examples_torch_cuda_stats.txt | |
cat reports/examples_torch_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: examples_test_reports | |
path: reports |