Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
7a26307e
"vscode:/vscode.git/clone" did not exist on "3a2c4e6f63ecdf807490977d43f0b7bf867bc626"
Unverified
Commit
7a26307e
authored
Sep 01, 2021
by
Sylvain Gugger
Committed by
GitHub
Sep 01, 2021
Browse files
Fixes for the documentation (#13361)
parent
0b8c84e1
Changes
7
Show whitespace changes
Inline
Side-by-side
Showing
7 changed files
with
32 additions
and
14 deletions
+32
-14
src/transformers/configuration_utils.py
src/transformers/configuration_utils.py
+1
-1
src/transformers/data/__init__.py
src/transformers/data/__init__.py
+10
-0
src/transformers/deepspeed.py
src/transformers/deepspeed.py
+1
-1
src/transformers/generation_tf_utils.py
src/transformers/generation_tf_utils.py
+8
-4
src/transformers/generation_utils.py
src/transformers/generation_utils.py
+8
-4
src/transformers/models/speech_to_text/configuration_speech_to_text.py
...ers/models/speech_to_text/configuration_speech_to_text.py
+2
-2
src/transformers/pipelines/__init__.py
src/transformers/pipelines/__init__.py
+2
-2
No files found.
src/transformers/configuration_utils.py
View file @
7a26307e
...
@@ -87,7 +87,7 @@ class PretrainedConfig(PushToHubMixin):
...
@@ -87,7 +87,7 @@ class PretrainedConfig(PushToHubMixin):
Whether cross-attention layers should be added to the model. Note, this option is only relevant for models
Whether cross-attention layers should be added to the model. Note, this option is only relevant for models
that can be used as decoder models within the `:class:~transformers.EncoderDecoderModel` class, which
that can be used as decoder models within the `:class:~transformers.EncoderDecoderModel` class, which
consists of all models in ``AUTO_MODELS_FOR_CAUSAL_LM``.
consists of all models in ``AUTO_MODELS_FOR_CAUSAL_LM``.
tie_encoder_decoder (:obj:`bool`, `optional`, defaults to :obj:`False`)
tie_encoder_decoder (:obj:`bool`, `optional`, defaults to :obj:`False`)
:
Whether all encoder weights should be tied to their equivalent decoder weights. This requires the encoder
Whether all encoder weights should be tied to their equivalent decoder weights. This requires the encoder
and decoder model to have the exact same parameter names.
and decoder model to have the exact same parameter names.
prune_heads (:obj:`Dict[int, List[int]]`, `optional`, defaults to :obj:`{}`):
prune_heads (:obj:`Dict[int, List[int]]`, `optional`, defaults to :obj:`{}`):
...
...
src/transformers/data/__init__.py
View file @
7a26307e
...
@@ -16,6 +16,16 @@
...
@@ -16,6 +16,16 @@
# See the License for the specific language governing permissions and
# See the License for the specific language governing permissions and
# limitations under the License.
# limitations under the License.
from
.data_collator
import
(
DataCollatorForLanguageModeling
,
DataCollatorForPermutationLanguageModeling
,
DataCollatorForSeq2Seq
,
DataCollatorForSOP
,
DataCollatorForTokenClassification
,
DataCollatorForWholeWordMask
,
DataCollatorWithPadding
,
default_data_collator
,
)
from
.metrics
import
glue_compute_metrics
,
xnli_compute_metrics
from
.metrics
import
glue_compute_metrics
,
xnli_compute_metrics
from
.processors
import
(
from
.processors
import
(
DataProcessor
,
DataProcessor
,
...
...
src/transformers/deepspeed.py
View file @
7a26307e
...
@@ -50,7 +50,7 @@ class HfDeepSpeedConfig:
...
@@ -50,7 +50,7 @@ class HfDeepSpeedConfig:
values: ``"auto"``. Without this special logic the DeepSpeed configuration is not modified in any way.
values: ``"auto"``. Without this special logic the DeepSpeed configuration is not modified in any way.
Args:
Args:
config_file_or_dict (:obj:`Union[str, Dict]`)
-
path to DeepSpeed config file or dict.
config_file_or_dict (:obj:`Union[str, Dict]`)
:
path to DeepSpeed config file or dict.
"""
"""
...
...
src/transformers/generation_tf_utils.py
View file @
7a26307e
...
@@ -1535,10 +1535,14 @@ def tf_top_k_top_p_filtering(logits, top_k=0, top_p=1.0, filter_value=-float("In
...
@@ -1535,10 +1535,14 @@ def tf_top_k_top_p_filtering(logits, top_k=0, top_p=1.0, filter_value=-float("In
Args:
Args:
logits: logits distribution shape (batch size, vocabulary size)
logits: logits distribution shape (batch size, vocabulary size)
if top_k > 0: keep only top k tokens with highest probability (top-k filtering).
top_k (:obj:`int`, `optional`, defaults to 0):
if top_p < 1.0: keep the top tokens with cumulative probability >= top_p (nucleus filtering).
If > 0, only keep the top k tokens with highest probability (top-k filtering)
Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751)
top_p (:obj:`float`, `optional`, defaults to 1.0):
Make sure we keep at least min_tokens_to_keep per batch example in the output
If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus
filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751)
min_tokens_to_keep (:obj:`int`, `optional`, defaults to 1):
Minimumber of tokens we keep per batch example in the output.
From: https://gist.github.com/thomwolf/1a5a29f6962089e871b94cbd09daf317
From: https://gist.github.com/thomwolf/1a5a29f6962089e871b94cbd09daf317
"""
"""
logits_shape
=
shape_list
(
logits
)
logits_shape
=
shape_list
(
logits
)
...
...
src/transformers/generation_utils.py
View file @
7a26307e
...
@@ -2557,10 +2557,14 @@ def top_k_top_p_filtering(
...
@@ -2557,10 +2557,14 @@ def top_k_top_p_filtering(
Args:
Args:
logits: logits distribution shape (batch size, vocabulary size)
logits: logits distribution shape (batch size, vocabulary size)
if top_k > 0: keep only top k tokens with highest probability (top-k filtering).
top_k (:obj:`int`, `optional`, defaults to 0):
if top_p < 1.0: keep the top tokens with cumulative probability >= top_p (nucleus filtering).
If > 0, only keep the top k tokens with highest probability (top-k filtering)
Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751)
top_p (:obj:`float`, `optional`, defaults to 1.0):
Make sure we keep at least min_tokens_to_keep per batch example in the output
If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus
filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751)
min_tokens_to_keep (:obj:`int`, `optional`, defaults to 1):
Minimumber of tokens we keep per batch example in the output.
From: https://gist.github.com/thomwolf/1a5a29f6962089e871b94cbd09daf317
From: https://gist.github.com/thomwolf/1a5a29f6962089e871b94cbd09daf317
"""
"""
if
top_k
>
0
:
if
top_k
>
0
:
...
...
src/transformers/models/speech_to_text/configuration_speech_to_text.py
View file @
7a26307e
...
@@ -78,7 +78,7 @@ class Speech2TextConfig(PretrainedConfig):
...
@@ -78,7 +78,7 @@ class Speech2TextConfig(PretrainedConfig):
Whether or not the model should return the last key/values attentions (not used by all models).
Whether or not the model should return the last key/values attentions (not used by all models).
max_source_positions (:obj:`int`, `optional`, defaults to 6000):
max_source_positions (:obj:`int`, `optional`, defaults to 6000):
The maximum sequence length of log-mel filter-bank features that this model might ever be used with.
The maximum sequence length of log-mel filter-bank features that this model might ever be used with.
max_target_positions
:
(:obj:`int`, `optional`, defaults to 1024):
max_target_positions (:obj:`int`, `optional`, defaults to 1024):
The maximum sequence length that this model might ever be used with. Typically set this to something large
The maximum sequence length that this model might ever be used with. Typically set this to something large
just in case (e.g., 512 or 1024 or 2048).
just in case (e.g., 512 or 1024 or 2048).
num_conv_layers (:obj:`int`, `optional`, defaults to 2):
num_conv_layers (:obj:`int`, `optional`, defaults to 2):
...
...
src/transformers/pipelines/__init__.py
View file @
7a26307e
...
@@ -306,10 +306,10 @@ def pipeline(
...
@@ -306,10 +306,10 @@ def pipeline(
- :obj:`"feature-extraction"`: will return a :class:`~transformers.FeatureExtractionPipeline`.
- :obj:`"feature-extraction"`: will return a :class:`~transformers.FeatureExtractionPipeline`.
- :obj:`"text-classification"`: will return a :class:`~transformers.TextClassificationPipeline`.
- :obj:`"text-classification"`: will return a :class:`~transformers.TextClassificationPipeline`.
- :obj:`"sentiment-analysis"`: (alias of :obj:`"text-classification") will return a
- :obj:`"sentiment-analysis"`: (alias of :obj:`"text-classification"
`
) will return a
:class:`~transformers.TextClassificationPipeline`.
:class:`~transformers.TextClassificationPipeline`.
- :obj:`"token-classification"`: will return a :class:`~transformers.TokenClassificationPipeline`.
- :obj:`"token-classification"`: will return a :class:`~transformers.TokenClassificationPipeline`.
- :obj:`"ner"` (alias of :obj:`"token-classification"): will return a
- :obj:`"ner"` (alias of :obj:`"token-classification"
`
): will return a
:class:`~transformers.TokenClassificationPipeline`.
:class:`~transformers.TokenClassificationPipeline`.
- :obj:`"question-answering"`: will return a :class:`~transformers.QuestionAnsweringPipeline`.
- :obj:`"question-answering"`: will return a :class:`~transformers.QuestionAnsweringPipeline`.
- :obj:`"fill-mask"`: will return a :class:`~transformers.FillMaskPipeline`.
- :obj:`"fill-mask"`: will return a :class:`~transformers.FillMaskPipeline`.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment