Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
011cc0be
Unverified
Commit
011cc0be
authored
Jun 16, 2020
by
Sylvain Gugger
Committed by
GitHub
Jun 16, 2020
Browse files
Fix all sphynx warnings (#5068)
parent
af497b56
Changes
25
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
17 additions
and
13 deletions
+17
-13
src/transformers/modeling_utils.py
src/transformers/modeling_utils.py
+1
-0
src/transformers/optimization_tf.py
src/transformers/optimization_tf.py
+12
-12
src/transformers/pipelines.py
src/transformers/pipelines.py
+1
-0
src/transformers/tokenization_auto.py
src/transformers/tokenization_auto.py
+2
-0
src/transformers/tokenization_utils_base.py
src/transformers/tokenization_utils_base.py
+1
-1
No files found.
src/transformers/modeling_utils.py
View file @
011cc0be
...
@@ -530,6 +530,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
...
@@ -530,6 +530,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
config: (`optional`) one of:
config: (`optional`) one of:
- an instance of a class derived from :class:`~transformers.PretrainedConfig`, or
- an instance of a class derived from :class:`~transformers.PretrainedConfig`, or
- a string valid as input to :func:`~transformers.PretrainedConfig.from_pretrained()`
- a string valid as input to :func:`~transformers.PretrainedConfig.from_pretrained()`
Configuration for the model to use instead of an automatically loaded configuation. Configuration can be automatically loaded when:
Configuration for the model to use instead of an automatically loaded configuation. Configuration can be automatically loaded when:
- the model is a model provided by the library (loaded with the ``shortcut-name`` string of a pretrained model), or
- the model is a model provided by the library (loaded with the ``shortcut-name`` string of a pretrained model), or
- the model was saved using :func:`~transformers.PreTrainedModel.save_pretrained` and is reloaded by suppling the save directory.
- the model was saved using :func:`~transformers.PreTrainedModel.save_pretrained` and is reloaded by suppling the save directory.
...
...
src/transformers/optimization_tf.py
View file @
011cc0be
...
@@ -97,13 +97,13 @@ def create_optimizer(
...
@@ -97,13 +97,13 @@ def create_optimizer(
class
AdamWeightDecay
(
tf
.
keras
.
optimizers
.
Adam
):
class
AdamWeightDecay
(
tf
.
keras
.
optimizers
.
Adam
):
"""Adam enables L2 weight decay and clip_by_global_norm on gradients.
"""Adam enables L2 weight decay and clip_by_global_norm on gradients.
Just adding the square of the weights to the loss function is *not* the
Just adding the square of the weights to the loss function is *not* the
correct way of using L2 regularization/weight decay with Adam, since that will
correct way of using L2 regularization/weight decay with Adam, since that will
interact with the m and v parameters in strange ways.
interact with the m and v parameters in strange ways.
Instead we want ot decay the weights in a manner that doesn't interact with
Instead we want ot decay the weights in a manner that doesn't interact with
the m/v parameters. This is equivalent to adding the square of the weights to
the m/v parameters. This is equivalent to adding the square of the weights to
the loss with plain (non-momentum) SGD.
the loss with plain (non-momentum) SGD.
"""
"""
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -198,11 +198,11 @@ class AdamWeightDecay(tf.keras.optimizers.Adam):
...
@@ -198,11 +198,11 @@ class AdamWeightDecay(tf.keras.optimizers.Adam):
# Extracted from https://github.com/OpenNMT/OpenNMT-tf/blob/master/opennmt/optimizers/utils.py
# Extracted from https://github.com/OpenNMT/OpenNMT-tf/blob/master/opennmt/optimizers/utils.py
class
GradientAccumulator
(
object
):
class
GradientAccumulator
(
object
):
"""Gradient accumulation utility.
"""Gradient accumulation utility.
When used with a distribution strategy, the accumulator should be called in a
When used with a distribution strategy, the accumulator should be called in a
replica context. Gradients will be accumulated locally on each replica and
replica context. Gradients will be accumulated locally on each replica and
without synchronization. Users should then call ``.gradients``, scale the
without synchronization. Users should then call ``.gradients``, scale the
gradients if required, and pass the result to ``apply_gradients``.
gradients if required, and pass the result to ``apply_gradients``.
"""
"""
# We use the ON_READ synchronization policy so that no synchronization is
# We use the ON_READ synchronization policy so that no synchronization is
# performed on assignment. To get the value, we call .value() which returns the
# performed on assignment. To get the value, we call .value() which returns the
...
...
src/transformers/pipelines.py
View file @
011cc0be
...
@@ -323,6 +323,7 @@ class Pipeline(_ScikitCompat):
...
@@ -323,6 +323,7 @@ class Pipeline(_ScikitCompat):
Base class implementing pipelined operations.
Base class implementing pipelined operations.
Pipeline workflow is defined as a sequence of the following operations:
Pipeline workflow is defined as a sequence of the following operations:
Input -> Tokenization -> Model Inference -> Post-Processing (Task dependent) -> Output
Input -> Tokenization -> Model Inference -> Post-Processing (Task dependent) -> Output
Pipeline supports running on CPU or GPU through the device argument. Users can specify
Pipeline supports running on CPU or GPU through the device argument. Users can specify
...
...
src/transformers/tokenization_auto.py
View file @
011cc0be
...
@@ -103,6 +103,7 @@ class AutoTokenizer:
...
@@ -103,6 +103,7 @@ class AutoTokenizer:
The `from_pretrained()` method takes care of returning the correct tokenizer class instance
The `from_pretrained()` method takes care of returning the correct tokenizer class instance
based on the `model_type` property of the config object, or when it's missing,
based on the `model_type` property of the config object, or when it's missing,
falling back to using pattern matching on the `pretrained_model_name_or_path` string:
falling back to using pattern matching on the `pretrained_model_name_or_path` string:
- `t5`: T5Tokenizer (T5 model)
- `t5`: T5Tokenizer (T5 model)
- `distilbert`: DistilBertTokenizer (DistilBert model)
- `distilbert`: DistilBertTokenizer (DistilBert model)
- `albert`: AlbertTokenizer (ALBERT model)
- `albert`: AlbertTokenizer (ALBERT model)
...
@@ -136,6 +137,7 @@ class AutoTokenizer:
...
@@ -136,6 +137,7 @@ class AutoTokenizer:
The tokenizer class to instantiate is selected
The tokenizer class to instantiate is selected
based on the `model_type` property of the config object, or when it's missing,
based on the `model_type` property of the config object, or when it's missing,
falling back to using pattern matching on the `pretrained_model_name_or_path` string:
falling back to using pattern matching on the `pretrained_model_name_or_path` string:
- `t5`: T5Tokenizer (T5 model)
- `t5`: T5Tokenizer (T5 model)
- `distilbert`: DistilBertTokenizer (DistilBert model)
- `distilbert`: DistilBertTokenizer (DistilBert model)
- `albert`: AlbertTokenizer (ALBERT model)
- `albert`: AlbertTokenizer (ALBERT model)
...
...
src/transformers/tokenization_utils_base.py
View file @
011cc0be
...
@@ -1408,7 +1408,7 @@ class PreTrainedTokenizerBase(SpecialTokensMixin):
...
@@ -1408,7 +1408,7 @@ class PreTrainedTokenizerBase(SpecialTokensMixin):
The sequence or batch of sequences to be encoded.
The sequence or batch of sequences to be encoded.
Each sequence can be a string or a list of strings (pre-tokenized string).
Each sequence can be a string or a list of strings (pre-tokenized string).
If the sequences are provided as list of strings (pretokenized), you must set `is_pretokenized=True`
If the sequences are provided as list of strings (pretokenized), you must set `is_pretokenized=True`
(to lift the ambiguity with a batch of sequences)
(to lift the ambiguity with a batch of sequences)
text_pair (:obj:`str`, :obj:`List[str]`, :obj:`List[List[str]]``):
text_pair (:obj:`str`, :obj:`List[str]`, :obj:`List[List[str]]``):
The sequence or batch of sequences to be encoded.
The sequence or batch of sequences to be encoded.
Each sequence can be a string or a list of strings (pre-tokenized string).
Each sequence can be a string or a list of strings (pre-tokenized string).
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment