create_circleci_config.py 19.9 KB
Newer Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
# coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import argparse
import copy
import os
19
import random
20
21
from dataclasses import dataclass
from typing import Any, Dict, List, Optional
22
import glob
23
import yaml
24
25


26
27
28
29
30
31
32
33
COMMON_ENV_VARIABLES = {
    "OMP_NUM_THREADS": 1,
    "TRANSFORMERS_IS_CI": True,
    "PYTEST_TIMEOUT": 120,
    "RUN_PIPELINE_TESTS": False,
    "RUN_PT_TF_CROSS_TESTS": False,
    "RUN_PT_FLAX_CROSS_TESTS": False,
}
Yih-Dar's avatar
Yih-Dar committed
34
# Disable the use of {"s": None} as the output is way too long, causing the navigation on CircleCI impractical
35
COMMON_PYTEST_OPTIONS = {"max-worker-restart": 0, "dist": "loadfile", "v": None}
36
DEFAULT_DOCKER_IMAGE = [{"image": "cimg/python:3.8.12"}]
37
38


39
40
41
42
43
44
45
46
47
48
class EmptyJob:
    job_name = "empty"

    def to_dict(self):
        return {
            "docker": copy.deepcopy(DEFAULT_DOCKER_IMAGE),
            "steps":["checkout"],
        }


49
50
51
52
53
@dataclass
class CircleCIJob:
    name: str
    additional_env: Dict[str, Any] = None
    cache_name: str = None
Arthur's avatar
Arthur committed
54
    cache_version: str = "0.8.2"
55
56
57
58
    docker_image: List[Dict[str, str]] = None
    install_steps: List[str] = None
    marker: Optional[str] = None
    parallelism: Optional[int] = 1
Yih-Dar's avatar
Yih-Dar committed
59
    pytest_num_workers: int = 12
60
    pytest_options: Dict[str, Any] = None
Yih-Dar's avatar
Yih-Dar committed
61
    resource_class: Optional[str] = "2xlarge"
62
    tests_to_run: Optional[List[str]] = None
63
64
    # This should be only used for doctest job!
    command_timeout: Optional[int] = None
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80

    def __post_init__(self):
        # Deal with defaults for mutable attributes.
        if self.additional_env is None:
            self.additional_env = {}
        if self.cache_name is None:
            self.cache_name = self.name
        if self.docker_image is None:
            # Let's avoid changing the default list and make a copy.
            self.docker_image = copy.deepcopy(DEFAULT_DOCKER_IMAGE)
        if self.install_steps is None:
            self.install_steps = []
        if self.pytest_options is None:
            self.pytest_options = {}
        if isinstance(self.tests_to_run, str):
            self.tests_to_run = [self.tests_to_run]
81
82
        if self.parallelism is None:
            self.parallelism = 1
83
84

    def to_dict(self):
85
86
        env = COMMON_ENV_VARIABLES.copy()
        env.update(self.additional_env)
87
88
89
90
91

        cache_branch_prefix = os.environ.get("CIRCLE_BRANCH", "pull")
        if cache_branch_prefix != "main":
            cache_branch_prefix = "pull"

92
93
        job = {
            "docker": self.docker_image,
94
            "environment": env,
95
96
97
98
99
100
101
        }
        if self.resource_class is not None:
            job["resource_class"] = self.resource_class
        if self.parallelism is not None:
            job["parallelism"] = self.parallelism
        steps = [
            "checkout",
102
            {"attach_workspace": {"at": "test_preparation"}},
103
104
        ]
        steps.extend([{"run": l} for l in self.install_steps])
105
106
107
108
109
        steps.append({"run": {"name": "Show installed libraries and their size", "command": """du -h -d 1 "$(pip -V | cut -d ' ' -f 4 | sed 's/pip//g')" | grep -vE "dist-info|_distutils_hack|__pycache__" | sort -h | tee installed.txt || true"""}})
        steps.append({"run": {"name": "Show installed libraries and their versions", "command": """pip list --format=freeze | tee installed.txt || true"""}})

        steps.append({"run":{"name":"Show biggest libraries","command":"""dpkg-query --show --showformat='${Installed-Size}\t${Package}\n' | sort -rh | head -25 | sort -h | awk '{ package=$2; sub(".*/", "", package); printf("%.5f GB %s\n", $1/1024/1024, package)}' || true"""}})
        steps.append({"store_artifacts": {"path": "installed.txt"}})
110
111

        all_options = {**COMMON_PYTEST_OPTIONS, **self.pytest_options}
112
        pytest_flags = [f"--{key}={value}" if (value is not None or key in ["doctest-modules"]) else f"-{key}" for key, value in all_options.items()]
113
114
115
        pytest_flags.append(
            f"--make-reports={self.name}" if "examples" in self.name else f"--make-reports=tests_{self.name}"
        )
116
117

        steps.append({"run": {"name": "Create `test-results` directory", "command": "mkdir test-results"}})
118
119
120
        test_command = ""
        if self.command_timeout:
            test_command = f"timeout {self.command_timeout} "
121
122
        # junit familiy xunit1 is necessary to support splitting on test name or class name with circleci split
        test_command += f"python3 -m pytest -rsfE -p no:warnings -o junit_family=xunit1 --tb=short --junitxml=test-results/junit.xml -n {self.pytest_num_workers} " + " ".join(pytest_flags)
amyeroberts's avatar
amyeroberts committed
123

124
125
126
127
128
        if self.parallelism == 1:
            if self.tests_to_run is None:
                test_command += " << pipeline.parameters.tests_to_run >>"
            else:
                test_command += " " + " ".join(self.tests_to_run)
129
        else:
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
            # We need explicit list instead of `pipeline.parameters.tests_to_run` (only available at job runtime)
            tests = self.tests_to_run
            if tests is None:
                folder = os.environ["test_preparation_dir"]
                test_file = os.path.join(folder, "filtered_test_list.txt")
                if os.path.exists(test_file):
                    with open(test_file) as f:
                        tests = f.read().split(" ")

            # expand the test list
            if tests == ["tests"]:
                tests = [os.path.join("tests", x) for x in os.listdir("tests")]
            expanded_tests = []
            for test in tests:
                if test.endswith(".py"):
                    expanded_tests.append(test)
                elif test == "tests/models":
147
                    expanded_tests.extend(glob.glob("tests/models/**/test*.py", recursive=True))
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
                elif test == "tests/pipelines":
                    expanded_tests.extend([os.path.join(test, x) for x in os.listdir(test)])
                else:
                    expanded_tests.append(test)
            tests = " ".join(expanded_tests)

            # Each executor to run ~10 tests
            n_executors = max(len(tests) // 10, 1)
            # Avoid empty test list on some executor(s) or launching too many executors
            if n_executors > self.parallelism:
                n_executors = self.parallelism
            job["parallelism"] = n_executors

            # Need to be newline separated for the command `circleci tests split` below
            command = f'echo {tests} | tr " " "\\n" >> tests.txt'
            steps.append({"run": {"name": "Get tests", "command": command}})

            command = 'TESTS=$(circleci tests split tests.txt) && echo $TESTS > splitted_tests.txt'
            steps.append({"run": {"name": "Split tests", "command": command}})

168
169
            steps.append({"store_artifacts": {"path": "tests.txt"}})
            steps.append({"store_artifacts": {"path": "splitted_tests.txt"}})
170

171
            test_command = ""
172
173
174
            if self.command_timeout:
                test_command = f"timeout {self.command_timeout} "
            test_command += f"python3 -m pytest -rsfE -p no:warnings --tb=short  -o junit_family=xunit1 --junitxml=test-results/junit.xml -n {self.pytest_num_workers} " + " ".join(pytest_flags)
175
            test_command += " $(cat splitted_tests.txt)"
176
177
        if self.marker is not None:
            test_command += f" -m {self.marker}"
178
179
180
181
182
183
184
185
186
187
188

        if self.name == "pr_documentation_tests":
            # can't use ` | tee tee tests_output.txt` as usual
            test_command += " > tests_output.txt"
            # Save the return code, so we can check if it is timeout in the next step.
            test_command += '; touch "$?".txt'
            # Never fail the test step for the doctest job. We will check the results in the next step, and fail that
            # step instead if the actual test failures are found. This is to avoid the timeout being reported as test
            # failure.
            test_command = f"({test_command}) || true"
        else:
189
            test_command = f"({test_command} | tee tests_output.txt)"
190
        steps.append({"run": {"name": "Run tests", "command": test_command}})
191

192
193
194
        steps.append({"run": {"name": "Skipped tests", "when": "always", "command": f"python3 .circleci/parse_test_outputs.py --file tests_output.txt --skip"}})
        steps.append({"run": {"name": "Failed tests",  "when": "always", "command": f"python3 .circleci/parse_test_outputs.py --file tests_output.txt --fail"}})
        steps.append({"run": {"name": "Errors",        "when": "always", "command": f"python3 .circleci/parse_test_outputs.py --file tests_output.txt --errors"}})
195

196
        steps.append({"store_test_results": {"path": "test-results"}})
197
198
199
        steps.append({"store_artifacts": {"path": "tests_output.txt"}})
        steps.append({"store_artifacts": {"path": "test-results/junit.xml"}})
        steps.append({"store_artifacts": {"path": "reports"}})
200

201
202
203
204
205
206
207
208
209
210
211
        job["steps"] = steps
        return job

    @property
    def job_name(self):
        return self.name if "examples" in self.name else f"tests_{self.name}"


# JOBS
torch_and_tf_job = CircleCIJob(
    "torch_and_tf",
212
213
    docker_image=[{"image":"huggingface/transformers-torch-tf-light"}],
    install_steps=["uv venv && uv pip install ."],
214
215
216
217
218
219
220
221
222
    additional_env={"RUN_PT_TF_CROSS_TESTS": True},
    marker="is_pt_tf_cross_test",
    pytest_options={"rA": None, "durations": 0},
)


torch_and_flax_job = CircleCIJob(
    "torch_and_flax",
    additional_env={"RUN_PT_FLAX_CROSS_TESTS": True},
223
224
    docker_image=[{"image":"huggingface/transformers-torch-jax-light"}],
    install_steps=["uv venv && uv pip install ."],
225
226
227
228
229
230
    marker="is_pt_flax_cross_test",
    pytest_options={"rA": None, "durations": 0},
)

torch_job = CircleCIJob(
    "torch",
231
232
233
234
    docker_image=[{"image": "huggingface/transformers-torch-light"}],
    install_steps=["uv venv && uv pip install ."],
    parallelism=6,
    pytest_num_workers=16
235
236
237
238
239
)


tf_job = CircleCIJob(
    "tf",
240
241
242
243
    docker_image=[{"image":"huggingface/transformers-tf-light"}],
    install_steps=["uv venv", "uv pip install -e."],
    parallelism=6,
    pytest_num_workers=16,
244
245
246
247
248
)


flax_job = CircleCIJob(
    "flax",
249
250
251
252
    docker_image=[{"image":"huggingface/transformers-jax-light"}],
    install_steps=["uv venv && uv pip install ."],
    parallelism=6,
    pytest_num_workers=16
253
254
255
256
257
)


pipelines_torch_job = CircleCIJob(
    "pipelines_torch",
258
    additional_env={"RUN_PIPELINE_TESTS": True},
259
260
    docker_image=[{"image":"huggingface/transformers-torch-light"}],
    install_steps=["uv venv && uv pip install ."],
261
    marker="is_pipeline_test",
262
263
264
265
266
)


pipelines_tf_job = CircleCIJob(
    "pipelines_tf",
267
    additional_env={"RUN_PIPELINE_TESTS": True},
268
269
    docker_image=[{"image":"huggingface/transformers-tf-light"}],
    install_steps=["uv venv && uv pip install ."],
270
    marker="is_pipeline_test",
271
272
273
274
275
276
)


custom_tokenizers_job = CircleCIJob(
    "custom_tokenizers",
    additional_env={"RUN_CUSTOM_TOKENIZERS": True},
277
278
    docker_image=[{"image": "huggingface/transformers-custom-tokenizers"}],
    install_steps=["uv venv","uv pip install -e ."],
279
280
281
282
283
284
285
286
287
288
289
290
    parallelism=None,
    resource_class=None,
    tests_to_run=[
        "./tests/models/bert_japanese/test_tokenization_bert_japanese.py",
        "./tests/models/openai/test_tokenization_openai.py",
        "./tests/models/clip/test_tokenization_clip.py",
    ],
)


examples_torch_job = CircleCIJob(
    "examples_torch",
291
    additional_env={"OMP_NUM_THREADS": 8},
292
    cache_name="torch_examples",
293
294
    docker_image=[{"image":"huggingface/transformers-examples-torch"}],
    install_steps=["uv venv && uv pip install ."],
295
    pytest_num_workers=1,
296
297
298
299
300
301
)


examples_tensorflow_job = CircleCIJob(
    "examples_tensorflow",
    cache_name="tensorflow_examples",
302
303
304
    docker_image=[{"image":"huggingface/transformers-examples-tf"}],
    install_steps=["uv venv && uv pip install ."],
    parallelism=8
305
306
307
308
309
)


hub_job = CircleCIJob(
    "hub",
Sylvain Gugger's avatar
Sylvain Gugger committed
310
    additional_env={"HUGGINGFACE_CO_STAGING": True},
311
    docker_image=[{"image":"huggingface/transformers-torch-light"}],
312
    install_steps=[
313
        "uv venv && uv pip install .",
314
315
316
317
318
319
320
321
322
323
        'git config --global user.email "ci@dummy.com"',
        'git config --global user.name "ci"',
    ],
    marker="is_staging_test",
    pytest_num_workers=1,
)


onnx_job = CircleCIJob(
    "onnx",
324
    docker_image=[{"image":"huggingface/transformers-torch-tf-light"}],
325
    install_steps=[
326
327
328
        "uv venv && uv pip install .",
        "uv pip install --upgrade eager pip",
        "uv pip install .[torch,tf,testing,sentencepiece,onnxruntime,vision,rjieba]",
329
330
331
332
333
334
    ],
    pytest_options={"k onnx": None},
    pytest_num_workers=1,
)


335
336
exotic_models_job = CircleCIJob(
    "exotic_models",
337
338
    install_steps=["uv venv && uv pip install ."],
    docker_image=[{"image":"huggingface/transformers-exotic-models"}],
339
340
341
    tests_to_run=[
        "tests/models/*layoutlmv*",
        "tests/models/*nat",
NielsRogge's avatar
NielsRogge committed
342
        "tests/models/deta",
NielsRogge's avatar
NielsRogge committed
343
        "tests/models/udop",
NielsRogge's avatar
NielsRogge committed
344
        "tests/models/nougat",
345
    ],
346
347
    pytest_num_workers=12,
    parallelism=4,
348
349
350
351
    pytest_options={"durations": 100},
)


Sylvain Gugger's avatar
Sylvain Gugger committed
352
353
repo_utils_job = CircleCIJob(
    "repo_utils",
354
355
    docker_image=[{"image":"huggingface/transformers-consistency"}],
    install_steps=["uv venv && uv pip install ."],
Sylvain Gugger's avatar
Sylvain Gugger committed
356
357
    parallelism=None,
    pytest_num_workers=1,
358
    resource_class="large",
Sylvain Gugger's avatar
Sylvain Gugger committed
359
360
361
    tests_to_run="tests/repo_utils",
)

362
363
364
365
366

# We also include a `dummy.py` file in the files to be doc-tested to prevent edge case failure. Otherwise, the pytest
# hangs forever during test collection while showing `collecting 0 items / 21 errors`. (To see this, we have to remove
# the bash output redirection.)
py_command = 'from utils.tests_fetcher import get_doctest_files; to_test = get_doctest_files() + ["dummy.py"]; to_test = " ".join(to_test); print(to_test)'
367
py_command = f"$(python3 -c '{py_command}')"
368
command = f'echo "{py_command}" > pr_documentation_tests_temp.txt'
369
370
doc_test_job = CircleCIJob(
    "pr_documentation_tests",
371
    docker_image=[{"image":"huggingface/transformers-consistency"}],
372
373
374
375
376
377
    additional_env={"TRANSFORMERS_VERBOSITY": "error", "DATASETS_VERBOSITY": "error", "SKIP_CUDA_DOCTEST": "1"},
    install_steps=[
        # Add an empty file to keep the test step running correctly even no file is selected to be tested.
        "touch dummy.py",
        {
            "name": "Get files to test",
378
            "command": command,
379
380
        },
        {
381
            "name": "Show information in `Get files to test`",
382
            "command":
383
                "cat pr_documentation_tests_temp.txt"
384
385
        },
        {
386
            "name": "Get the last line in `pr_documentation_tests.txt`",
387
            "command":
388
                "tail -n1 pr_documentation_tests_temp.txt | tee pr_documentation_tests.txt"
389
390
        },
    ],
391
    tests_to_run="$(cat pr_documentation_tests.txt)",  # noqa
392
    pytest_options={"-doctest-modules": None, "doctest-glob": "*.md", "dist": "loadfile", "rvsA": None},
393
394
395
396
    command_timeout=1200,  # test cannot run longer than 1200 seconds
    pytest_num_workers=1,
)

397
398
399
400
401
402
403
404
405
REGULAR_TESTS = [
    torch_and_tf_job,
    torch_and_flax_job,
    torch_job,
    tf_job,
    flax_job,
    custom_tokenizers_job,
    hub_job,
    onnx_job,
406
    exotic_models_job,
407
408
409
410
411
412
413
414
415
]
EXAMPLES_TESTS = [
    examples_torch_job,
    examples_tensorflow_job,
]
PIPELINE_TESTS = [
    pipelines_torch_job,
    pipelines_tf_job,
]
Sylvain Gugger's avatar
Sylvain Gugger committed
416
REPO_UTIL_TESTS = [repo_utils_job]
417
418
DOC_TESTS = [doc_test_job]

419
420
421
422

def create_circleci_config(folder=None):
    if folder is None:
        folder = os.getcwd()
423
424
    # Used in CircleCIJob.to_dict() to expand the test list (for using parallelism)
    os.environ["test_preparation_dir"] = folder
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
    jobs = []
    all_test_file = os.path.join(folder, "test_list.txt")
    if os.path.exists(all_test_file):
        with open(all_test_file) as f:
            all_test_list = f.read()
    else:
        all_test_list = []
    if len(all_test_list) > 0:
        jobs.extend(PIPELINE_TESTS)

    test_file = os.path.join(folder, "filtered_test_list.txt")
    if os.path.exists(test_file):
        with open(test_file) as f:
            test_list = f.read()
    else:
        test_list = []
    if len(test_list) > 0:
        jobs.extend(REGULAR_TESTS)

Yih-Dar's avatar
Yih-Dar committed
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
        extended_tests_to_run = set(test_list.split())
        # Extend the test files for cross test jobs
        for job in jobs:
            if job.job_name in ["tests_torch_and_tf", "tests_torch_and_flax"]:
                for test_path in copy.copy(extended_tests_to_run):
                    dir_path, fn = os.path.split(test_path)
                    if fn.startswith("test_modeling_tf_"):
                        fn = fn.replace("test_modeling_tf_", "test_modeling_")
                    elif fn.startswith("test_modeling_flax_"):
                        fn = fn.replace("test_modeling_flax_", "test_modeling_")
                    else:
                        if job.job_name == "test_torch_and_tf":
                            fn = fn.replace("test_modeling_", "test_modeling_tf_")
                        elif job.job_name == "test_torch_and_flax":
                            fn = fn.replace("test_modeling_", "test_modeling_flax_")
                    new_test_file = str(os.path.join(dir_path, fn))
                    if os.path.isfile(new_test_file):
                        if new_test_file not in extended_tests_to_run:
                            extended_tests_to_run.add(new_test_file)
        extended_tests_to_run = sorted(extended_tests_to_run)
        for job in jobs:
            if job.job_name in ["tests_torch_and_tf", "tests_torch_and_flax"]:
                job.tests_to_run = extended_tests_to_run
        fn = "filtered_test_list_cross_tests.txt"
        f_path = os.path.join(folder, fn)
        with open(f_path, "w") as fp:
            fp.write(" ".join(extended_tests_to_run))

472
473
    example_file = os.path.join(folder, "examples_test_list.txt")
    if os.path.exists(example_file) and os.path.getsize(example_file) > 0:
474
        with open(example_file, "r", encoding="utf-8") as f:
475
            example_tests = f.read()
476
477
478
479
480
        for job in EXAMPLES_TESTS:
            framework = job.name.replace("examples_", "").replace("torch", "pytorch")
            if example_tests == "all":
                job.tests_to_run = [f"examples/{framework}"]
            else:
481
                job.tests_to_run = [f for f in example_tests.split(" ") if f.startswith(f"examples/{framework}")]
482

483
484
            if len(job.tests_to_run) > 0:
                jobs.append(job)
485

486
487
488
489
490
491
492
493
494
    doctest_file = os.path.join(folder, "doctest_list.txt")
    if os.path.exists(doctest_file):
        with open(doctest_file) as f:
            doctest_list = f.read()
    else:
        doctest_list = []
    if len(doctest_list) > 0:
        jobs.extend(DOC_TESTS)

Sylvain Gugger's avatar
Sylvain Gugger committed
495
496
497
    repo_util_file = os.path.join(folder, "test_repo_utils.txt")
    if os.path.exists(repo_util_file) and os.path.getsize(repo_util_file) > 0:
        jobs.extend(REPO_UTIL_TESTS)
498

499
500
501
502
503
504
505
506
507
508
509
510
    if len(jobs) == 0:
        jobs = [EmptyJob()]
    config = {"version": "2.1"}
    config["parameters"] = {
        # Only used to accept the parameters from the trigger
        "nightly": {"type": "boolean", "default": False},
        "tests_to_run": {"type": "string", "default": test_list},
    }
    config["jobs"] = {j.job_name: j.to_dict() for j in jobs}
    config["workflows"] = {"version": 2, "run_tests": {"jobs": [j.job_name for j in jobs]}}
    with open(os.path.join(folder, "generated_config.yml"), "w") as f:
        f.write(yaml.dump(config, indent=2, width=1000000, sort_keys=False))
511
512
513
514
515
516
517
518
519
520


if __name__ == "__main__":
    parser = argparse.ArgumentParser()
    parser.add_argument(
        "--fetcher_folder", type=str, default=None, help="Only test that all tests and modules are accounted for."
    )
    args = parser.parse_args()

    create_circleci_config(args.fetcher_folder)