Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
a3c5883f
Commit
a3c5883f
authored
Dec 22, 2019
by
Aymeric Augustin
Browse files
Rename file for consistency.
parent
daf8bebc
Changes
16
Hide whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
15 additions
and
15 deletions
+15
-15
templates/adding_a_new_model/tests/test_tokenization_xxx.py
templates/adding_a_new_model/tests/test_tokenization_xxx.py
+1
-1
tests/test_configuration_common.py
tests/test_configuration_common.py
+1
-1
tests/test_model_card.py
tests/test_model_card.py
+1
-1
tests/test_optimization.py
tests/test_optimization.py
+1
-1
tests/test_tokenization_albert.py
tests/test_tokenization_albert.py
+1
-1
tests/test_tokenization_bert.py
tests/test_tokenization_bert.py
+1
-1
tests/test_tokenization_bert_japanese.py
tests/test_tokenization_bert_japanese.py
+1
-1
tests/test_tokenization_common.py
tests/test_tokenization_common.py
+0
-0
tests/test_tokenization_ctrl.py
tests/test_tokenization_ctrl.py
+1
-1
tests/test_tokenization_gpt2.py
tests/test_tokenization_gpt2.py
+1
-1
tests/test_tokenization_openai.py
tests/test_tokenization_openai.py
+1
-1
tests/test_tokenization_roberta.py
tests/test_tokenization_roberta.py
+1
-1
tests/test_tokenization_t5.py
tests/test_tokenization_t5.py
+1
-1
tests/test_tokenization_transfo_xl.py
tests/test_tokenization_transfo_xl.py
+1
-1
tests/test_tokenization_xlm.py
tests/test_tokenization_xlm.py
+1
-1
tests/test_tokenization_xlnet.py
tests/test_tokenization_xlnet.py
+1
-1
No files found.
templates/adding_a_new_model/tests/test_tokenization_xxx.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers.tokenization_bert
import
VOCAB_FILES_NAMES
,
XxxTokenizer
from
transformers.tokenization_bert
import
VOCAB_FILES_NAMES
,
XxxTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
XxxTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
XxxTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_configuration_common.py
View file @
a3c5883f
...
@@ -17,7 +17,7 @@ from __future__ import absolute_import, division, print_function
...
@@ -17,7 +17,7 @@ from __future__ import absolute_import, division, print_function
import
json
import
json
import
os
import
os
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
class
ConfigTester
(
object
):
class
ConfigTester
(
object
):
...
...
tests/test_model_card.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ import unittest
...
@@ -20,7 +20,7 @@ import unittest
from
transformers.modelcard
import
ModelCard
from
transformers.modelcard
import
ModelCard
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
class
ModelCardTester
(
unittest
.
TestCase
):
class
ModelCardTester
(
unittest
.
TestCase
):
...
...
tests/test_optimization.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import unittest
...
@@ -19,7 +19,7 @@ import unittest
from
transformers
import
is_torch_available
from
transformers
import
is_torch_available
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
from
.utils
import
require_torch
from
.utils
import
require_torch
...
...
tests/test_tokenization_albert.py
View file @
a3c5883f
...
@@ -18,7 +18,7 @@ import os
...
@@ -18,7 +18,7 @@ import os
from
transformers.tokenization_albert
import
AlbertTokenizer
from
transformers.tokenization_albert
import
AlbertTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/spiece.model"
)
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/spiece.model"
)
...
...
tests/test_tokenization_bert.py
View file @
a3c5883f
...
@@ -27,7 +27,7 @@ from transformers.tokenization_bert import (
...
@@ -27,7 +27,7 @@ from transformers.tokenization_bert import (
_is_whitespace
,
_is_whitespace
,
)
)
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_bert_japanese.py
View file @
a3c5883f
...
@@ -25,7 +25,7 @@ from transformers.tokenization_bert_japanese import (
...
@@ -25,7 +25,7 @@ from transformers.tokenization_bert_japanese import (
MecabTokenizer
,
MecabTokenizer
,
)
)
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
custom_tokenizers
,
slow
from
.utils
import
custom_tokenizers
,
slow
...
...
tests/tokenization_
tests_
common
s
.py
→
tests/
test_
tokenization_common.py
View file @
a3c5883f
File moved
tests/test_tokenization_ctrl.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers.tokenization_ctrl
import
VOCAB_FILES_NAMES
,
CTRLTokenizer
from
transformers.tokenization_ctrl
import
VOCAB_FILES_NAMES
,
CTRLTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
CTRLTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
CTRLTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_gpt2.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ from io import open
...
@@ -20,7 +20,7 @@ from io import open
from
transformers.tokenization_gpt2
import
VOCAB_FILES_NAMES
,
GPT2Tokenizer
from
transformers.tokenization_gpt2
import
VOCAB_FILES_NAMES
,
GPT2Tokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
GPT2TokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
GPT2TokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_openai.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_openai
import
VOCAB_FILES_NAMES
,
OpenAIGPTTokenizer
from
transformers.tokenization_openai
import
VOCAB_FILES_NAMES
,
OpenAIGPTTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
OpenAIGPTTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
OpenAIGPTTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_roberta.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ from io import open
...
@@ -20,7 +20,7 @@ from io import open
from
transformers.tokenization_roberta
import
VOCAB_FILES_NAMES
,
RobertaTokenizer
from
transformers.tokenization_roberta
import
VOCAB_FILES_NAMES
,
RobertaTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_t5.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_t5
import
T5Tokenizer
from
transformers.tokenization_t5
import
T5Tokenizer
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/test_sentencepiece.model"
)
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/test_sentencepiece.model"
)
...
...
tests/test_tokenization_transfo_xl.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers
import
is_torch_available
from
transformers
import
is_torch_available
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
require_torch
from
.utils
import
require_torch
...
...
tests/test_tokenization_xlm.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_xlm
import
VOCAB_FILES_NAMES
,
XLMTokenizer
from
transformers.tokenization_xlm
import
VOCAB_FILES_NAMES
,
XLMTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_xlnet.py
View file @
a3c5883f
...
@@ -18,7 +18,7 @@ import os
...
@@ -18,7 +18,7 @@ import os
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
,
XLNetTokenizer
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
,
XLNetTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment