"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "f176e707234ab4d6f2a44179066d71658cc40056"
Unverified Commit 95e2e14f authored by Sylvain Gugger's avatar Sylvain Gugger Committed by GitHub
Browse files

Revert to all tests whil we debug what's wrong (#13072)

parent 477480ce
...@@ -50,21 +50,22 @@ jobs: ...@@ -50,21 +50,22 @@ jobs:
python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())" python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())"
python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())" python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())"
- name: Fetch the tests to run # - name: Fetch the tests to run
run: | # run: |
python utils/tests_fetcher.py | tee test_preparation.txt # python utils/tests_fetcher.py | tee test_preparation.txt
#
- name: Report fetched tests # - name: Report fetched tests
uses: actions/upload-artifact@v2 # uses: actions/upload-artifact@v2
with: # with:
name: test_fetched # name: test_fetched
path: test_preparation.txt # path: test_preparation.txt
#
- name: Run all non-slow tests on GPU - name: Run all non-slow tests on GPU
run: | run: |
if [ -f test_list.txt ]; then python -m pytest -n 2 --dist=loadfile -v --make-reports=tests_torch_gpu tests
python -m pytest -n 2 --dist=loadfile -v --make-reports=tests_torch_gpu $(cat test_list.txt) # if [ -f test_list.txt ]; then
fi # python -m pytest -n 2 --dist=loadfile -v --make-reports=tests_torch_gpu $(cat test_list.txt)
# fi
- name: Failure short reports - name: Failure short reports
if: ${{ always() }} if: ${{ always() }}
...@@ -162,15 +163,15 @@ jobs: ...@@ -162,15 +163,15 @@ jobs:
python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())" python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())"
python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())" python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())"
- name: Fetch the tests to run # - name: Fetch the tests to run
run: | # run: |
python utils/tests_fetcher.py | tee test_preparation.txt # python utils/tests_fetcher.py | tee test_preparation.txt
#
- name: Report fetched tests # - name: Report fetched tests
uses: actions/upload-artifact@v2 # uses: actions/upload-artifact@v2
with: # with:
name: test_fetched # name: test_fetched
path: test_preparation.txt # path: test_preparation.txt
- name: Run all non-slow tests on GPU - name: Run all non-slow tests on GPU
env: env:
...@@ -275,21 +276,22 @@ jobs: ...@@ -275,21 +276,22 @@ jobs:
python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())" python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())"
python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())" python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())"
- name: Fetch the tests to run # - name: Fetch the tests to run
run: | # run: |
python utils/tests_fetcher.py --filters tests/deepspeed tests/extended | tee test_preparation.txt # python utils/tests_fetcher.py --filters tests/deepspeed tests/extended | tee test_preparation.txt
#
- name: Report fetched tests # - name: Report fetched tests
uses: actions/upload-artifact@v2 # uses: actions/upload-artifact@v2
with: # with:
name: test_fetched # name: test_fetched
path: test_preparation.txt # path: test_preparation.txt
- name: Run all tests on GPU - name: Run all tests on GPU
run: | run: |
if [ -f test_list.txt ]; then python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_gpu tests/deepspeed tests/extended
python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_gpu $(cat test_list.txt) # if [ -f test_list.txt ]; then
fi # python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_gpu $(cat test_list.txt)
# fi
- name: Failure short reports - name: Failure short reports
if: ${{ always() }} if: ${{ always() }}
...@@ -330,21 +332,22 @@ jobs: ...@@ -330,21 +332,22 @@ jobs:
python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())" python -c "import torch; print('CuDNN version:', torch.backends.cudnn.version())"
python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())" python -c "import torch; print('Number of GPUs available:', torch.cuda.device_count())"
- name: Fetch the tests to run # - name: Fetch the tests to run
run: | # run: |
python utils/tests_fetcher.py --filters tests/deepspeed tests/extended | tee test_preparation.txt # python utils/tests_fetcher.py --filters tests/deepspeed tests/extended | tee test_preparation.txt
#
- name: Report fetched tests # - name: Report fetched tests
uses: actions/upload-artifact@v2 # uses: actions/upload-artifact@v2
with: # with:
name: test_fetched # name: test_fetched
path: test_preparation.txt # path: test_preparation.txt
- name: Run all tests on GPU - name: Run all tests on GPU
run: | run: |
if [ -f test_list.txt ]; then python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_multi_gpu tests/deepspeed tests/extended
python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_multi_gpu $(cat test_list.txt) # if [ -f test_list.txt ]; then
fi # python -m pytest -n 1 --dist=loadfile -v --make-reports=tests_torch_cuda_extensions_multi_gpu $(cat test_list.txt)
# fi
- name: Failure short reports - name: Failure short reports
if: ${{ always() }} if: ${{ always() }}
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment