Unverified Commit bfc9369f authored by Adrien's avatar Adrien Committed by GitHub
Browse files

[CI] Update runner configuration for setup and nightly tests (#9005)



* [CI] Update runner configuration for setup and nightly tests
Signed-off-by: default avatarAdrien <adrien@huggingface.co>

* fix group
Signed-off-by: default avatarAdrien <adrien@huggingface.co>

* update for t4
Signed-off-by: default avatarAdrien <adrien@huggingface.co>

---------
Signed-off-by: default avatarAdrien <adrien@huggingface.co>
parent 73acebb8
...@@ -20,7 +20,8 @@ env: ...@@ -20,7 +20,8 @@ env:
jobs: jobs:
test-build-docker-images: test-build-docker-images:
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
if: github.event_name == 'pull_request' if: github.event_name == 'pull_request'
steps: steps:
- name: Set up Docker Buildx - name: Set up Docker Buildx
...@@ -50,7 +51,8 @@ jobs: ...@@ -50,7 +51,8 @@ jobs:
if: steps.file_changes.outputs.all != '' if: steps.file_changes.outputs.all != ''
build-and-push-docker-images: build-and-push-docker-images:
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
if: github.event_name != 'pull_request' if: github.event_name != 'pull_request'
permissions: permissions:
...@@ -98,4 +100,4 @@ jobs: ...@@ -98,4 +100,4 @@ jobs:
slack_channel: ${{ env.CI_SLACK_CHANNEL }} slack_channel: ${{ env.CI_SLACK_CHANNEL }}
title: "🤗 Results of the ${{ matrix.image-name }} Docker Image build" title: "🤗 Results of the ${{ matrix.image-name }} Docker Image build"
status: ${{ job.status }} status: ${{ job.status }}
slack_token: ${{ secrets.SLACK_CIFEEDBACK_BOT_TOKEN }} slack_token: ${{ secrets.SLACK_CIFEEDBACK_BOT_TOKEN }}
\ No newline at end of file
...@@ -19,7 +19,8 @@ env: ...@@ -19,7 +19,8 @@ env:
jobs: jobs:
setup_torch_cuda_pipeline_matrix: setup_torch_cuda_pipeline_matrix:
name: Setup Torch Pipelines CUDA Slow Tests Matrix name: Setup Torch Pipelines CUDA Slow Tests Matrix
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
container: container:
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
outputs: outputs:
...@@ -55,7 +56,8 @@ jobs: ...@@ -55,7 +56,8 @@ jobs:
max-parallel: 8 max-parallel: 8
matrix: matrix:
module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }}
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-cuda image: diffusers/diffusers-pytorch-cuda
options: --shm-size "16gb" --ipc host --gpus 0 options: --shm-size "16gb" --ipc host --gpus 0
...@@ -105,7 +107,8 @@ jobs: ...@@ -105,7 +107,8 @@ jobs:
run_nightly_tests_for_other_torch_modules: run_nightly_tests_for_other_torch_modules:
name: Nightly Torch CUDA Tests name: Nightly Torch CUDA Tests
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-cuda image: diffusers/diffusers-pytorch-cuda
options: --shm-size "16gb" --ipc host --gpus 0 options: --shm-size "16gb" --ipc host --gpus 0
...@@ -234,7 +237,8 @@ jobs: ...@@ -234,7 +237,8 @@ jobs:
run_nightly_onnx_tests: run_nightly_onnx_tests:
name: Nightly ONNXRuntime CUDA tests on Ubuntu name: Nightly ONNXRuntime CUDA tests on Ubuntu
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-onnxruntime-cuda image: diffusers/diffusers-onnxruntime-cuda
options: --gpus 0 --shm-size "16gb" --ipc host options: --gpus 0 --shm-size "16gb" --ipc host
......
...@@ -15,7 +15,8 @@ concurrency: ...@@ -15,7 +15,8 @@ concurrency:
jobs: jobs:
setup_pr_tests: setup_pr_tests:
name: Setup PR Tests name: Setup PR Tests
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
container: container:
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/
...@@ -73,7 +74,8 @@ jobs: ...@@ -73,7 +74,8 @@ jobs:
max-parallel: 2 max-parallel: 2
matrix: matrix:
modules: ${{ fromJson(needs.setup_pr_tests.outputs.matrix) }} modules: ${{ fromJson(needs.setup_pr_tests.outputs.matrix) }}
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
container: container:
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/
...@@ -123,12 +125,13 @@ jobs: ...@@ -123,12 +125,13 @@ jobs:
config: config:
- name: Hub tests for models, schedulers, and pipelines - name: Hub tests for models, schedulers, and pipelines
framework: hub_tests_pytorch framework: hub_tests_pytorch
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_hub report: torch_hub
name: ${{ matrix.config.name }} name: ${{ matrix.config.name }}
runs-on: ${{ matrix.config.runner }} runs-on:
group: ${{ matrix.config.runner }}
container: container:
image: ${{ matrix.config.image }} image: ${{ matrix.config.image }}
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/
......
...@@ -71,7 +71,8 @@ jobs: ...@@ -71,7 +71,8 @@ jobs:
name: LoRA - ${{ matrix.lib-versions }} name: LoRA - ${{ matrix.lib-versions }}
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
container: container:
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
...@@ -128,4 +129,4 @@ jobs: ...@@ -128,4 +129,4 @@ jobs:
uses: actions/upload-artifact@v2 uses: actions/upload-artifact@v2
with: with:
name: pr_${{ matrix.config.report }}_test_reports name: pr_${{ matrix.config.report }}_test_reports
path: reports path: reports
\ No newline at end of file
...@@ -77,28 +77,29 @@ jobs: ...@@ -77,28 +77,29 @@ jobs:
config: config:
- name: Fast PyTorch Pipeline CPU tests - name: Fast PyTorch Pipeline CPU tests
framework: pytorch_pipelines framework: pytorch_pipelines
runner: [ self-hosted, intel-cpu, 32-cpu, 256-ram, ci ] runner: aws-highmemory-32-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_cpu_pipelines report: torch_cpu_pipelines
- name: Fast PyTorch Models & Schedulers CPU tests - name: Fast PyTorch Models & Schedulers CPU tests
framework: pytorch_models framework: pytorch_models
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_cpu_models_schedulers report: torch_cpu_models_schedulers
- name: Fast Flax CPU tests - name: Fast Flax CPU tests
framework: flax framework: flax
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-flax-cpu image: diffusers/diffusers-flax-cpu
report: flax_cpu report: flax_cpu
- name: PyTorch Example CPU tests - name: PyTorch Example CPU tests
framework: pytorch_examples framework: pytorch_examples
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_example_cpu report: torch_example_cpu
name: ${{ matrix.config.name }} name: ${{ matrix.config.name }}
runs-on: ${{ matrix.config.runner }} runs-on:
group: ${{ matrix.config.runner }}
container: container:
image: ${{ matrix.config.image }} image: ${{ matrix.config.image }}
...@@ -180,7 +181,8 @@ jobs: ...@@ -180,7 +181,8 @@ jobs:
config: config:
- name: Hub tests for models, schedulers, and pipelines - name: Hub tests for models, schedulers, and pipelines
framework: hub_tests_pytorch framework: hub_tests_pytorch
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner:
group: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_hub report: torch_hub
......
...@@ -19,7 +19,8 @@ env: ...@@ -19,7 +19,8 @@ env:
jobs: jobs:
setup_torch_cuda_pipeline_matrix: setup_torch_cuda_pipeline_matrix:
name: Setup Torch Pipelines CUDA Slow Tests Matrix name: Setup Torch Pipelines CUDA Slow Tests Matrix
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] runs-on:
group: aws-general-8-plus
container: container:
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
outputs: outputs:
...@@ -57,7 +58,8 @@ jobs: ...@@ -57,7 +58,8 @@ jobs:
max-parallel: 8 max-parallel: 8
matrix: matrix:
module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }}
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-cuda image: diffusers/diffusers-pytorch-cuda
options: --shm-size "16gb" --ipc host --gpus 0 options: --shm-size "16gb" --ipc host --gpus 0
...@@ -101,7 +103,8 @@ jobs: ...@@ -101,7 +103,8 @@ jobs:
torch_cuda_tests: torch_cuda_tests:
name: Torch CUDA Tests name: Torch CUDA Tests
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-cuda image: diffusers/diffusers-pytorch-cuda
options: --shm-size "16gb" --ipc host --gpus 0 options: --shm-size "16gb" --ipc host --gpus 0
...@@ -201,7 +204,8 @@ jobs: ...@@ -201,7 +204,8 @@ jobs:
onnx_cuda_tests: onnx_cuda_tests:
name: ONNX CUDA Tests name: ONNX CUDA Tests
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-onnxruntime-cuda image: diffusers/diffusers-onnxruntime-cuda
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0 options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
...@@ -249,7 +253,8 @@ jobs: ...@@ -249,7 +253,8 @@ jobs:
run_torch_compile_tests: run_torch_compile_tests:
name: PyTorch Compile CUDA tests name: PyTorch Compile CUDA tests
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-compile-cuda image: diffusers/diffusers-pytorch-compile-cuda
...@@ -291,7 +296,8 @@ jobs: ...@@ -291,7 +296,8 @@ jobs:
run_xformers_tests: run_xformers_tests:
name: PyTorch xformers CUDA tests name: PyTorch xformers CUDA tests
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-xformers-cuda image: diffusers/diffusers-pytorch-xformers-cuda
...@@ -332,7 +338,8 @@ jobs: ...@@ -332,7 +338,8 @@ jobs:
run_examples_tests: run_examples_tests:
name: Examples PyTorch CUDA tests on Ubuntu name: Examples PyTorch CUDA tests on Ubuntu
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: diffusers/diffusers-pytorch-cuda image: diffusers/diffusers-pytorch-cuda
......
...@@ -29,28 +29,29 @@ jobs: ...@@ -29,28 +29,29 @@ jobs:
config: config:
- name: Fast PyTorch CPU tests on Ubuntu - name: Fast PyTorch CPU tests on Ubuntu
framework: pytorch framework: pytorch
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_cpu report: torch_cpu
- name: Fast Flax CPU tests on Ubuntu - name: Fast Flax CPU tests on Ubuntu
framework: flax framework: flax
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-flax-cpu image: diffusers/diffusers-flax-cpu
report: flax_cpu report: flax_cpu
- name: Fast ONNXRuntime CPU tests on Ubuntu - name: Fast ONNXRuntime CPU tests on Ubuntu
framework: onnxruntime framework: onnxruntime
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-onnxruntime-cpu image: diffusers/diffusers-onnxruntime-cpu
report: onnx_cpu report: onnx_cpu
- name: PyTorch Example CPU tests on Ubuntu - name: PyTorch Example CPU tests on Ubuntu
framework: pytorch_examples framework: pytorch_examples
runner: [ self-hosted, intel-cpu, 8-cpu, ci ] runner: aws-general-8-plus
image: diffusers/diffusers-pytorch-cpu image: diffusers/diffusers-pytorch-cpu
report: torch_example_cpu report: torch_example_cpu
name: ${{ matrix.config.name }} name: ${{ matrix.config.name }}
runs-on: ${{ matrix.config.runner }} runs-on:
group: ${{ matrix.config.runner }}
container: container:
image: ${{ matrix.config.image }} image: ${{ matrix.config.image }}
......
...@@ -26,7 +26,8 @@ env: ...@@ -26,7 +26,8 @@ env:
jobs: jobs:
run_tests: run_tests:
name: "Run a test on our runner from a PR" name: "Run a test on our runner from a PR"
runs-on: [single-gpu, nvidia-gpu, t4, ci] runs-on:
group: aws-g4dn-2xlarge
container: container:
image: ${{ github.event.inputs.docker_image }} image: ${{ github.event.inputs.docker_image }}
options: --gpus 0 --privileged --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ options: --gpus 0 --privileged --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/
...@@ -70,4 +71,4 @@ jobs: ...@@ -70,4 +71,4 @@ jobs:
env: env:
PY_TEST: ${{ github.event.inputs.test }} PY_TEST: ${{ github.event.inputs.test }}
run: | run: |
pytest "$PY_TEST" pytest "$PY_TEST"
\ No newline at end of file
...@@ -19,7 +19,8 @@ env: ...@@ -19,7 +19,8 @@ env:
jobs: jobs:
ssh_runner: ssh_runner:
name: "SSH" name: "SSH"
runs-on: [self-hosted, intel-cpu, 32-cpu, 256-ram, ci] runs-on:
group: aws-highmemory-32-plus
container: container:
image: ${{ github.event.inputs.docker_image }} image: ${{ github.event.inputs.docker_image }}
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --privileged options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --privileged
......
...@@ -22,7 +22,8 @@ env: ...@@ -22,7 +22,8 @@ env:
jobs: jobs:
ssh_runner: ssh_runner:
name: "SSH" name: "SSH"
runs-on: [single-gpu, nvidia-gpu, "${{ github.event.inputs.runner_type }}", ci] runs-on:
group: "${{ github.event.inputs.runner_type }}"
container: container:
image: ${{ github.event.inputs.docker_image }} image: ${{ github.event.inputs.docker_image }}
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 --privileged options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 --privileged
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment