Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
1b7de41a
Unverified
Commit
1b7de41a
authored
Apr 14, 2022
by
code-review-doctor
Committed by
GitHub
Apr 14, 2022
Browse files
Fix issue avoid-missing-comma found at
https://codereview.doctor
(#16768)
parent
de8b06f9
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
2 deletions
+2
-2
tests/bert_japanese/test_tokenization_bert_japanese.py
tests/bert_japanese/test_tokenization_bert_japanese.py
+2
-2
No files found.
tests/bert_japanese/test_tokenization_bert_japanese.py
View file @
1b7de41a
...
...
@@ -173,7 +173,7 @@ class BertJapaneseTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
)
def
test_wordpiece_tokenizer
(
self
):
vocab_tokens
=
[
"[UNK]"
,
"[CLS]"
,
"[SEP]"
,
"こんにちは"
,
"こん"
,
"にちは"
"ばんは"
,
"##こん"
,
"##にちは"
,
"##ばんは"
]
vocab_tokens
=
[
"[UNK]"
,
"[CLS]"
,
"[SEP]"
,
"こんにちは"
,
"こん"
,
"にちは"
,
"ばんは"
,
"##こん"
,
"##にちは"
,
"##ばんは"
]
vocab
=
{}
for
(
i
,
token
)
in
enumerate
(
vocab_tokens
):
...
...
@@ -246,7 +246,7 @@ class BertJapaneseCharacterTokenizationTest(TokenizerTesterMixin, unittest.TestC
)
def
test_character_tokenizer
(
self
):
vocab_tokens
=
[
"[UNK]"
,
"[CLS]"
,
"[SEP]"
,
"こ"
,
"ん"
,
"に"
,
"ち"
,
"は"
,
"ば"
,
"世"
,
"界"
"、"
,
"。"
]
vocab_tokens
=
[
"[UNK]"
,
"[CLS]"
,
"[SEP]"
,
"こ"
,
"ん"
,
"に"
,
"ち"
,
"は"
,
"ば"
,
"世"
,
"界"
,
"、"
,
"。"
]
vocab
=
{}
for
(
i
,
token
)
in
enumerate
(
vocab_tokens
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment