Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
a3c5883f
"web/vscode:/vscode.git/clone" did not exist on "2d7ad41142a941f7d82e166869bb1e8372a559ac"
Commit
a3c5883f
authored
Dec 22, 2019
by
Aymeric Augustin
Browse files
Rename file for consistency.
parent
daf8bebc
Changes
16
Show whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
15 additions
and
15 deletions
+15
-15
templates/adding_a_new_model/tests/test_tokenization_xxx.py
templates/adding_a_new_model/tests/test_tokenization_xxx.py
+1
-1
tests/test_configuration_common.py
tests/test_configuration_common.py
+1
-1
tests/test_model_card.py
tests/test_model_card.py
+1
-1
tests/test_optimization.py
tests/test_optimization.py
+1
-1
tests/test_tokenization_albert.py
tests/test_tokenization_albert.py
+1
-1
tests/test_tokenization_bert.py
tests/test_tokenization_bert.py
+1
-1
tests/test_tokenization_bert_japanese.py
tests/test_tokenization_bert_japanese.py
+1
-1
tests/test_tokenization_common.py
tests/test_tokenization_common.py
+0
-0
tests/test_tokenization_ctrl.py
tests/test_tokenization_ctrl.py
+1
-1
tests/test_tokenization_gpt2.py
tests/test_tokenization_gpt2.py
+1
-1
tests/test_tokenization_openai.py
tests/test_tokenization_openai.py
+1
-1
tests/test_tokenization_roberta.py
tests/test_tokenization_roberta.py
+1
-1
tests/test_tokenization_t5.py
tests/test_tokenization_t5.py
+1
-1
tests/test_tokenization_transfo_xl.py
tests/test_tokenization_transfo_xl.py
+1
-1
tests/test_tokenization_xlm.py
tests/test_tokenization_xlm.py
+1
-1
tests/test_tokenization_xlnet.py
tests/test_tokenization_xlnet.py
+1
-1
No files found.
templates/adding_a_new_model/tests/test_tokenization_xxx.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers.tokenization_bert
import
VOCAB_FILES_NAMES
,
XxxTokenizer
from
transformers.tokenization_bert
import
VOCAB_FILES_NAMES
,
XxxTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
XxxTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
XxxTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_configuration_common.py
View file @
a3c5883f
...
@@ -17,7 +17,7 @@ from __future__ import absolute_import, division, print_function
...
@@ -17,7 +17,7 @@ from __future__ import absolute_import, division, print_function
import
json
import
json
import
os
import
os
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
class
ConfigTester
(
object
):
class
ConfigTester
(
object
):
...
...
tests/test_model_card.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ import unittest
...
@@ -20,7 +20,7 @@ import unittest
from
transformers.modelcard
import
ModelCard
from
transformers.modelcard
import
ModelCard
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
class
ModelCardTester
(
unittest
.
TestCase
):
class
ModelCardTester
(
unittest
.
TestCase
):
...
...
tests/test_optimization.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import unittest
...
@@ -19,7 +19,7 @@ import unittest
from
transformers
import
is_torch_available
from
transformers
import
is_torch_available
from
.tokenization_
tests_
commo
ns
import
TemporaryDirectory
from
.
test_
tokenization_commo
import
TemporaryDirectory
from
.utils
import
require_torch
from
.utils
import
require_torch
...
...
tests/test_tokenization_albert.py
View file @
a3c5883f
...
@@ -18,7 +18,7 @@ import os
...
@@ -18,7 +18,7 @@ import os
from
transformers.tokenization_albert
import
AlbertTokenizer
from
transformers.tokenization_albert
import
AlbertTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/spiece.model"
)
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/spiece.model"
)
...
...
tests/test_tokenization_bert.py
View file @
a3c5883f
...
@@ -27,7 +27,7 @@ from transformers.tokenization_bert import (
...
@@ -27,7 +27,7 @@ from transformers.tokenization_bert import (
_is_whitespace
,
_is_whitespace
,
)
)
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_bert_japanese.py
View file @
a3c5883f
...
@@ -25,7 +25,7 @@ from transformers.tokenization_bert_japanese import (
...
@@ -25,7 +25,7 @@ from transformers.tokenization_bert_japanese import (
MecabTokenizer
,
MecabTokenizer
,
)
)
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
custom_tokenizers
,
slow
from
.utils
import
custom_tokenizers
,
slow
...
...
tests/tokenization_
tests_
common
s
.py
→
tests/
test_
tokenization_common.py
View file @
a3c5883f
File moved
tests/test_tokenization_ctrl.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers.tokenization_ctrl
import
VOCAB_FILES_NAMES
,
CTRLTokenizer
from
transformers.tokenization_ctrl
import
VOCAB_FILES_NAMES
,
CTRLTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
CTRLTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
CTRLTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_gpt2.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ from io import open
...
@@ -20,7 +20,7 @@ from io import open
from
transformers.tokenization_gpt2
import
VOCAB_FILES_NAMES
,
GPT2Tokenizer
from
transformers.tokenization_gpt2
import
VOCAB_FILES_NAMES
,
GPT2Tokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
GPT2TokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
GPT2TokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_openai.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_openai
import
VOCAB_FILES_NAMES
,
OpenAIGPTTokenizer
from
transformers.tokenization_openai
import
VOCAB_FILES_NAMES
,
OpenAIGPTTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
class
OpenAIGPTTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
class
OpenAIGPTTokenizationTest
(
CommonTestCases
.
CommonTokenizerTester
):
...
...
tests/test_tokenization_roberta.py
View file @
a3c5883f
...
@@ -20,7 +20,7 @@ from io import open
...
@@ -20,7 +20,7 @@ from io import open
from
transformers.tokenization_roberta
import
VOCAB_FILES_NAMES
,
RobertaTokenizer
from
transformers.tokenization_roberta
import
VOCAB_FILES_NAMES
,
RobertaTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_t5.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_t5
import
T5Tokenizer
from
transformers.tokenization_t5
import
T5Tokenizer
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/test_sentencepiece.model"
)
SAMPLE_VOCAB
=
os
.
path
.
join
(
os
.
path
.
dirname
(
os
.
path
.
abspath
(
__file__
)),
"fixtures/test_sentencepiece.model"
)
...
...
tests/test_tokenization_transfo_xl.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ from io import open
...
@@ -19,7 +19,7 @@ from io import open
from
transformers
import
is_torch_available
from
transformers
import
is_torch_available
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
require_torch
from
.utils
import
require_torch
...
...
tests/test_tokenization_xlm.py
View file @
a3c5883f
...
@@ -19,7 +19,7 @@ import os
...
@@ -19,7 +19,7 @@ import os
from
transformers.tokenization_xlm
import
VOCAB_FILES_NAMES
,
XLMTokenizer
from
transformers.tokenization_xlm
import
VOCAB_FILES_NAMES
,
XLMTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
tests/test_tokenization_xlnet.py
View file @
a3c5883f
...
@@ -18,7 +18,7 @@ import os
...
@@ -18,7 +18,7 @@ import os
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
,
XLNetTokenizer
from
transformers.tokenization_xlnet
import
SPIECE_UNDERLINE
,
XLNetTokenizer
from
.tokenization_
tests_
commo
ns
import
CommonTestCases
from
.
test_
tokenization_commo
import
CommonTestCases
from
.utils
import
slow
from
.utils
import
slow
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment