Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
885fda09
Commit
885fda09
authored
Jan 25, 2022
by
A. Unique TensorFlower
Browse files
Internal change
PiperOrigin-RevId: 424187808
parent
159697a2
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
13 additions
and
6 deletions
+13
-6
official/nlp/modeling/layers/text_layers.py
official/nlp/modeling/layers/text_layers.py
+13
-6
No files found.
official/nlp/modeling/layers/text_layers.py
View file @
885fda09
...
@@ -14,7 +14,7 @@
...
@@ -14,7 +14,7 @@
"""Keras Layers for BERT-specific preprocessing."""
"""Keras Layers for BERT-specific preprocessing."""
# pylint: disable=g-import-not-at-top
# pylint: disable=g-import-not-at-top
from
typing
import
Any
,
Dict
,
List
,
Optional
,
Union
from
typing
import
Any
,
Dict
,
List
,
Mapping
,
Optional
,
Text
,
Union
from
absl
import
logging
from
absl
import
logging
import
tensorflow
as
tf
import
tensorflow
as
tf
...
@@ -71,8 +71,9 @@ class BertTokenizer(tf.keras.layers.Layer):
...
@@ -71,8 +71,9 @@ class BertTokenizer(tf.keras.layers.Layer):
def
__init__
(
self
,
*
,
def
__init__
(
self
,
*
,
vocab_file
:
str
,
vocab_file
:
str
,
lower_case
:
bool
,
lower_case
:
Optional
[
bool
]
=
None
,
tokenize_with_offsets
:
bool
=
False
,
tokenize_with_offsets
:
bool
=
False
,
tokenizer_kwargs
:
Optional
[
Mapping
[
Text
,
Any
]]
=
None
,
**
kwargs
):
**
kwargs
):
"""Initialize a `BertTokenizer` layer.
"""Initialize a `BertTokenizer` layer.
...
@@ -81,15 +82,18 @@ class BertTokenizer(tf.keras.layers.Layer):
...
@@ -81,15 +82,18 @@ class BertTokenizer(tf.keras.layers.Layer):
This is a text file with newline-separated wordpiece tokens.
This is a text file with newline-separated wordpiece tokens.
This layer initializes a lookup table from it that gets used with
This layer initializes a lookup table from it that gets used with
`text.BertTokenizer`.
`text.BertTokenizer`.
lower_case:
A Python
boolean forwarded to `text.BertTokenizer`.
lower_case:
Optional
boolean forwarded to `text.BertTokenizer`.
If true, input text is converted to lower case (where applicable)
If true, input text is converted to lower case (where applicable)
before tokenization. This must be set to match the way in which
before tokenization. This must be set to match the way in which
the `vocab_file` was created.
the `vocab_file` was created. If passed, this overrides whatever value
may have been passed in `tokenizer_kwargs`.
tokenize_with_offsets: A Python boolean. If true, this layer calls
tokenize_with_offsets: A Python boolean. If true, this layer calls
`text.BertTokenizer.tokenize_with_offsets()` instead of plain
`text.BertTokenizer.tokenize_with_offsets()` instead of plain
`text.BertTokenizer.tokenize()` and outputs a triple of
`text.BertTokenizer.tokenize()` and outputs a triple of
`(tokens, start_offsets, limit_offsets)`
`(tokens, start_offsets, limit_offsets)`
insead of just tokens.
insead of just tokens.
tokenizer_kwargs: Optional mapping with keyword arguments to forward to
`text.BertTokenizer`'s constructor.
**kwargs: Standard arguments to `Layer()`.
**kwargs: Standard arguments to `Layer()`.
Raises:
Raises:
...
@@ -111,8 +115,11 @@ class BertTokenizer(tf.keras.layers.Layer):
...
@@ -111,8 +115,11 @@ class BertTokenizer(tf.keras.layers.Layer):
self
.
_special_tokens_dict
=
self
.
_create_special_tokens_dict
(
self
.
_special_tokens_dict
=
self
.
_create_special_tokens_dict
(
self
.
_vocab_table
,
vocab_file
)
self
.
_vocab_table
,
vocab_file
)
super
().
__init__
(
**
kwargs
)
super
().
__init__
(
**
kwargs
)
self
.
_bert_tokenizer
=
text
.
BertTokenizer
(
tokenizer_kwargs
=
dict
(
tokenizer_kwargs
or
{})
self
.
_vocab_table
,
lower_case
=
lower_case
)
if
lower_case
is
not
None
:
tokenizer_kwargs
[
"lower_case"
]
=
lower_case
self
.
_bert_tokenizer
=
text
.
BertTokenizer
(
self
.
_vocab_table
,
**
tokenizer_kwargs
)
@
property
@
property
def
vocab_size
(
self
):
def
vocab_size
(
self
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment