Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
1b652295
"examples/single_model_scripts/run_transfo_xl.py" did not exist on "6cd769957e5eadd59963bb02efd3b746e8d120af"
Unverified
Commit
1b652295
authored
May 21, 2021
by
Lysandre Debut
Committed by
GitHub
May 21, 2021
Browse files
Patch recursive import (#11812)
parent
bd987165
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
8 additions
and
2 deletions
+8
-2
src/transformers/convert_slow_tokenizer.py
src/transformers/convert_slow_tokenizer.py
+2
-1
tests/test_tokenization_utils.py
tests/test_tokenization_utils.py
+6
-1
No files found.
src/transformers/convert_slow_tokenizer.py
View file @
1b652295
...
@@ -25,7 +25,6 @@ from tokenizers import Regex, Tokenizer, decoders, normalizers, pre_tokenizers,
...
@@ -25,7 +25,6 @@ from tokenizers import Regex, Tokenizer, decoders, normalizers, pre_tokenizers,
from
tokenizers.models
import
BPE
,
Unigram
,
WordPiece
from
tokenizers.models
import
BPE
,
Unigram
,
WordPiece
from
.file_utils
import
requires_backends
from
.file_utils
import
requires_backends
from
.models.roformer.tokenization_utils
import
JiebaPreTokenizer
class
SentencePieceExtractor
:
class
SentencePieceExtractor
:
...
@@ -299,6 +298,8 @@ class RobertaConverter(Converter):
...
@@ -299,6 +298,8 @@ class RobertaConverter(Converter):
class
RoFormerConverter
(
Converter
):
class
RoFormerConverter
(
Converter
):
def
converted
(
self
)
->
Tokenizer
:
def
converted
(
self
)
->
Tokenizer
:
from
.models.roformer.tokenization_utils
import
JiebaPreTokenizer
vocab
=
self
.
original_tokenizer
.
vocab
vocab
=
self
.
original_tokenizer
.
vocab
tokenizer
=
Tokenizer
(
WordPiece
(
vocab
,
unk_token
=
str
(
self
.
original_tokenizer
.
unk_token
)))
tokenizer
=
Tokenizer
(
WordPiece
(
vocab
,
unk_token
=
str
(
self
.
original_tokenizer
.
unk_token
)))
...
...
tests/test_tokenization_utils.py
View file @
1b652295
...
@@ -12,6 +12,9 @@
...
@@ -12,6 +12,9 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# See the License for the specific language governing permissions and
# limitations under the License.
# limitations under the License.
"""
isort:skip_file
"""
import
os
import
os
import
pickle
import
pickle
import
tempfile
import
tempfile
...
@@ -20,12 +23,14 @@ from typing import Callable, Optional
...
@@ -20,12 +23,14 @@ from typing import Callable, Optional
import
numpy
as
np
import
numpy
as
np
# Ensure there are no circular imports when importing the parent class
from
transformers
import
PreTrainedTokenizerFast
from
transformers
import
(
from
transformers
import
(
BatchEncoding
,
BatchEncoding
,
BertTokenizer
,
BertTokenizer
,
BertTokenizerFast
,
BertTokenizerFast
,
PreTrainedTokenizer
,
PreTrainedTokenizer
,
PreTrainedTokenizerFast
,
TensorType
,
TensorType
,
TokenSpan
,
TokenSpan
,
is_tokenizers_available
,
is_tokenizers_available
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment