Unverified Commit a929f81e authored by Sylvain Gugger's avatar Sylvain Gugger Committed by GitHub
Browse files

Repo utils test (#19696)

* Create repo utils test job

* Last occurence

* Add tests for tests_fetcher

* Better filtering

* Let's learn more

* Should fix

* Should fix

* Remove debug

* Style

* WiP

WiP

WiP

WiP

WiP

WiP

WiP

WiP

WiP

* Quality

* address review comments

* Fix link
parent a23819ed
...@@ -30,7 +30,13 @@ jobs: ...@@ -30,7 +30,13 @@ jobs:
else else
touch test_preparation/test_list.txt touch test_preparation/test_list.txt
fi fi
- run: python utils/tests_fetcher.py --filter_pipeline_tests - run: |
if [ -f test_repo_utils.txt ]; then
mv test_repo_utils.txt test_preparation/test_repo_utils.txt
else
touch test_preparation/test_repo_utils.txt
fi
- run: python utils/tests_fetcher.py --filter_tests
- run: | - run: |
if [ -f test_list.txt ]; then if [ -f test_list.txt ]; then
mv test_list.txt test_preparation/filtered_test_list.txt mv test_list.txt test_preparation/filtered_test_list.txt
...@@ -75,8 +81,9 @@ jobs: ...@@ -75,8 +81,9 @@ jobs:
mkdir test_preparation mkdir test_preparation
echo "tests" > test_preparation/test_list.txt echo "tests" > test_preparation/test_list.txt
echo "tests" > test_preparation/examples_test_list.txt echo "tests" > test_preparation/examples_test_list.txt
- run: python utils/tests_fetcher.py --filter_pipeline_tests - run: python utils/tests_fetcher.py --filter_tests
- run: mv test_list.txt test_preparation/filtered_test_list.txt - run: mv test_list.txt test_preparation/filtered_test_list.txt
- run: mv test_repo_utils.txt test_preparation/test_repo_utils.txt
- run: python .circleci/create_circleci_config.py --fetcher_folder test_preparation - run: python .circleci/create_circleci_config.py --fetcher_folder test_preparation
- run: cp test_preparation/generated_config.yml test_preparation/generated_config.txt - run: cp test_preparation/generated_config.yml test_preparation/generated_config.txt
- store_artifacts: - store_artifacts:
......
...@@ -324,6 +324,18 @@ layoutlm_job = CircleCIJob( ...@@ -324,6 +324,18 @@ layoutlm_job = CircleCIJob(
) )
repo_utils_job = CircleCIJob(
"repo_utils",
install_steps=[
"pip install --upgrade pip",
"pip install .[all,quality,testing]",
],
parallelism=None,
pytest_num_workers=1,
resource_class=None,
tests_to_run="tests/repo_utils",
)
REGULAR_TESTS = [ REGULAR_TESTS = [
torch_and_tf_job, torch_and_tf_job,
torch_and_flax_job, torch_and_flax_job,
...@@ -344,7 +356,7 @@ PIPELINE_TESTS = [ ...@@ -344,7 +356,7 @@ PIPELINE_TESTS = [
pipelines_torch_job, pipelines_torch_job,
pipelines_tf_job, pipelines_tf_job,
] ]
REPO_UTIL_TESTS = [repo_utils_job]
def create_circleci_config(folder=None): def create_circleci_config(folder=None):
if folder is None: if folder is None:
...@@ -371,6 +383,10 @@ def create_circleci_config(folder=None): ...@@ -371,6 +383,10 @@ def create_circleci_config(folder=None):
example_file = os.path.join(folder, "examples_test_list.txt") example_file = os.path.join(folder, "examples_test_list.txt")
if os.path.exists(example_file) and os.path.getsize(example_file) > 0: if os.path.exists(example_file) and os.path.getsize(example_file) > 0:
jobs.extend(EXAMPLES_TESTS) jobs.extend(EXAMPLES_TESTS)
repo_util_file = os.path.join(folder, "test_repo_utils.txt")
if os.path.exists(repo_util_file) and os.path.getsize(repo_util_file) > 0:
jobs.extend(REPO_UTIL_TESTS)
if len(jobs) > 0: if len(jobs) > 0:
config = {"version": "2.1"} config = {"version": "2.1"}
......
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
import sys
import unittest
from git import Repo
git_repo_path = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
sys.path.append(os.path.join(git_repo_path, "utils"))
transformers_path = os.path.join(git_repo_path, "src", "transformers")
# Tests are run against this specific commit for reproducibility
# https://github.com/huggingface/transformers/tree/07f6690206e39ed7a4d9dbc58824314f7089bb38
GIT_TEST_SHA = "07f6690206e39ed7a4d9dbc58824314f7089bb38"
from tests_fetcher import checkout_commit, clean_code, get_module_dependencies # noqa: E402
class CheckDummiesTester(unittest.TestCase):
def test_clean_code(self):
# Clean code removes all strings in triple quotes
self.assertEqual(clean_code('"""\nDocstring\n"""\ncode\n"""Long string"""\ncode\n'), "code\ncode")
self.assertEqual(clean_code("'''\nDocstring\n'''\ncode\n'''Long string'''\ncode\n'''"), "code\ncode")
# Clean code removes all comments
self.assertEqual(clean_code("code\n# Comment\ncode"), "code\ncode")
self.assertEqual(clean_code("code # inline comment\ncode"), "code \ncode")
def test_checkout_commit(self):
repo = Repo(git_repo_path)
self.assertNotEqual(repo.head.commit.hexsha, GIT_TEST_SHA)
with checkout_commit(repo, GIT_TEST_SHA):
self.assertEqual(repo.head.commit.hexsha, GIT_TEST_SHA)
self.assertNotEqual(repo.head.commit.hexsha, GIT_TEST_SHA)
def test_get_module_dependencies(self):
bert_module = os.path.join(transformers_path, "models", "bert", "modeling_bert.py")
expected_deps = [
"activations.py",
"modeling_outputs.py",
"modeling_utils.py",
"pytorch_utils.py",
"models/bert/configuration_bert.py",
]
expected_deps = set(os.path.join(transformers_path, f) for f in expected_deps)
repo = Repo(git_repo_path)
with checkout_commit(repo, GIT_TEST_SHA):
deps = get_module_dependencies(bert_module)
deps = set(os.path.expanduser(f) for f in deps)
self.assertEqual(deps, expected_deps)
...@@ -547,6 +547,7 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j ...@@ -547,6 +547,7 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j
# Grab the corresponding test files: # Grab the corresponding test files:
if "setup.py" in impacted_files: if "setup.py" in impacted_files:
test_files_to_run = ["tests"] test_files_to_run = ["tests"]
repo_utils_launch = True
else: else:
# Grab the corresponding test files: # Grab the corresponding test files:
test_files_to_run = [] test_files_to_run = []
...@@ -577,6 +578,12 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j ...@@ -577,6 +578,12 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j
for filter in filters: for filter in filters:
filtered_files.extend([f for f in test_files_to_run if f.startswith(filter)]) filtered_files.extend([f for f in test_files_to_run if f.startswith(filter)])
test_files_to_run = filtered_files test_files_to_run = filtered_files
repo_utils_launch = any(f.split(os.path.sep)[1] == "repo_utils" for f in test_files_to_run)
if repo_utils_launch:
repo_util_file = Path(output_file).parent / "test_repo_utils.txt"
with open(repo_util_file, "w", encoding="utf-8") as f:
f.write("tests/repo_utils")
print(f"\n### TEST TO RUN ###\n{_print_list(test_files_to_run)}") print(f"\n### TEST TO RUN ###\n{_print_list(test_files_to_run)}")
if len(test_files_to_run) > 0: if len(test_files_to_run) > 0:
...@@ -620,20 +627,29 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j ...@@ -620,20 +627,29 @@ def infer_tests_to_run(output_file, diff_with_last_commit=False, filters=None, j
json.dump(test_map, fp, ensure_ascii=False) json.dump(test_map, fp, ensure_ascii=False)
def filter_pipeline_tests(output_file): def filter_tests(output_file, filters):
"""
Reads the content of the output file and filters out all the tests in a list of given folders.
Args:
output_file (`str` or `os.PathLike`): The path to the output file of the tests fetcher.
filters (`List[str]`): A list of folders to filter.
"""
if not os.path.isfile(output_file): if not os.path.isfile(output_file):
print("No test file found.") print("No test file found.")
return return
with open(output_file, "r", encoding="utf-8") as f: with open(output_file, "r", encoding="utf-8") as f:
test_files = f.read().split(" ") test_files = f.read().split(" ")
if len(test_files) == 0: if len(test_files) == 0 or test_files == [""]:
print("No tests to filter.") print("No tests to filter.")
return return
print(test_files)
if test_files == ["tests"]: if test_files == ["tests"]:
test_files = [os.path.join("tests", f) for f in os.listdir("tests") if f not in ["__init__.py", "pipelines"]] test_files = [os.path.join("tests", f) for f in os.listdir("tests") if f not in ["__init__.py"] + filters]
else: else:
test_files = [f for f in test_files if not f.startswith(os.path.join("tests", "pipelines"))] test_files = [f for f in test_files if f.split(os.path.sep)[1] not in filters]
with open(output_file, "w", encoding="utf-8") as f: with open(output_file, "w", encoding="utf-8") as f:
f.write(" ".join(test_files)) f.write(" ".join(test_files))
...@@ -666,9 +682,9 @@ if __name__ == "__main__": ...@@ -666,9 +682,9 @@ if __name__ == "__main__":
help="Only keep the test files matching one of those filters.", help="Only keep the test files matching one of those filters.",
) )
parser.add_argument( parser.add_argument(
"--filter_pipeline_tests", "--filter_tests",
action="store_true", action="store_true",
help="Will filter the pipeline tests outside of the generated list of tests.", help="Will filter the pipeline/repo utils tests outside of the generated list of tests.",
) )
parser.add_argument( parser.add_argument(
"--print_dependencies_of", "--print_dependencies_of",
...@@ -681,8 +697,8 @@ if __name__ == "__main__": ...@@ -681,8 +697,8 @@ if __name__ == "__main__":
print_tree_deps_of(args.print_dependencies_of) print_tree_deps_of(args.print_dependencies_of)
elif args.sanity_check: elif args.sanity_check:
sanity_check() sanity_check()
elif args.filter_pipeline_tests: elif args.filter_tests:
filter_pipeline_tests(args.output_file) filter_tests(args.output_file, ["pipelines", "repo_utils"])
else: else:
repo = Repo(PATH_TO_TRANFORMERS) repo = Repo(PATH_TO_TRANFORMERS)
...@@ -698,6 +714,7 @@ if __name__ == "__main__": ...@@ -698,6 +714,7 @@ if __name__ == "__main__":
filters=args.filters, filters=args.filters,
json_output_file=args.json_output_file, json_output_file=args.json_output_file,
) )
filter_tests(args.output_file, ["repo_utils"])
except Exception as e: except Exception as e:
print(f"\nError when trying to grab the relevant tests: {e}\n\nRunning all tests.") print(f"\nError when trying to grab the relevant tests: {e}\n\nRunning all tests.")
with open(args.output_file, "w", encoding="utf-8") as f: with open(args.output_file, "w", encoding="utf-8") as f:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment