test_utils_check_copies.py 4.75 KB
Newer Older
Sylvain Gugger's avatar
Sylvain Gugger committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
import os
import re
import shutil
import sys
import tempfile
import unittest


git_repo_path = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))
sys.path.append(os.path.join(git_repo_path, "utils"))

import check_copies  # noqa: E402


# This is the reference code that will be used in the tests.
# If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated.
REFERENCE_CODE = """    def __init__(self, config):
        super().__init__()
        self.transform = BertPredictionHeadTransform(config)

        # The output weights are the same as the input embeddings, but there is
        # an output-only bias for each token.
        self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

        self.bias = nn.Parameter(torch.zeros(config.vocab_size))

        # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`
        self.decoder.bias = self.bias

    def forward(self, hidden_states):
        hidden_states = self.transform(hidden_states)
        hidden_states = self.decoder(hidden_states)
        return hidden_states
"""


class CopyCheckTester(unittest.TestCase):
    def setUp(self):
        self.transformer_dir = tempfile.mkdtemp()
Sylvain Gugger's avatar
Sylvain Gugger committed
54
        os.makedirs(os.path.join(self.transformer_dir, "models/bert/"))
55
56
        check_copies.TRANSFORMER_PATH = self.transformer_dir
        shutil.copy(
Sylvain Gugger's avatar
Sylvain Gugger committed
57
58
            os.path.join(git_repo_path, "src/transformers/models/bert/modeling_bert.py"),
            os.path.join(self.transformer_dir, "models/bert/modeling_bert.py"),
59
60
61
62
63
64
65
66
67
68
69
70
71
72
        )

    def tearDown(self):
        check_copies.TRANSFORMER_PATH = "src/transformers"
        shutil.rmtree(self.transformer_dir)

    def check_copy_consistency(self, comment, class_name, class_code, overwrite_result=None):
        code = comment + f"\nclass {class_name}(nn.Module):\n" + class_code
        if overwrite_result is not None:
            expected = comment + f"\nclass {class_name}(nn.Module):\n" + overwrite_result
        fname = os.path.join(self.transformer_dir, "new_code.py")
        with open(fname, "w") as f:
            f.write(code)
        if overwrite_result is None:
73
            self.assertTrue(len(check_copies.is_copy_consistent(fname)) == 0)
74
75
76
77
78
79
        else:
            check_copies.is_copy_consistent(f.name, overwrite=True)
            with open(fname, "r") as f:
                self.assertTrue(f.read(), expected)

    def test_find_code_in_transformers(self):
Sylvain Gugger's avatar
Sylvain Gugger committed
80
        code = check_copies.find_code_in_transformers("models.bert.modeling_bert.BertLMPredictionHead")
81
82
83
84
85
        self.assertEqual(code, REFERENCE_CODE)

    def test_is_copy_consistent(self):
        # Base copy consistency
        self.check_copy_consistency(
Sylvain Gugger's avatar
Sylvain Gugger committed
86
            "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead",
87
88
89
90
91
92
            "BertLMPredictionHead",
            REFERENCE_CODE + "\n",
        )

        # With no empty line at the end
        self.check_copy_consistency(
Sylvain Gugger's avatar
Sylvain Gugger committed
93
            "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead",
94
95
96
97
98
99
            "BertLMPredictionHead",
            REFERENCE_CODE,
        )

        # Copy consistency with rename
        self.check_copy_consistency(
Sylvain Gugger's avatar
Sylvain Gugger committed
100
            "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel",
101
102
103
104
105
106
107
            "TestModelLMPredictionHead",
            re.sub("Bert", "TestModel", REFERENCE_CODE),
        )

        # Copy consistency with a really long name
        long_class_name = "TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReasonIReallyDontUnderstand"
        self.check_copy_consistency(
Sylvain Gugger's avatar
Sylvain Gugger committed
108
            f"# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}",
109
110
111
112
113
114
            f"{long_class_name}LMPredictionHead",
            re.sub("Bert", long_class_name, REFERENCE_CODE),
        )

        # Copy consistency with overwrite
        self.check_copy_consistency(
Sylvain Gugger's avatar
Sylvain Gugger committed
115
            "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel",
116
117
118
119
            "TestModelLMPredictionHead",
            REFERENCE_CODE,
            overwrite_result=re.sub("Bert", "TestModel", REFERENCE_CODE),
        )