"test/ut/vscode:/vscode.git/clone" did not exist on "055885d9a703ea645eb8d54ca893d5ed05c7d9f1"
- 30 Nov, 2019 2 commits
-
-
Rostislav Nedelchev authored
Whenever target_mapping is provided to the input, XLNet outputs two different attention streams. Based on that the attention output would be on of the two: - a list of tensors (usual case for most transformers) - a list of 2-tuples of tensors, one tesor for each of attention streams Docs and unit-tests have been updated
-
Rostislav Nedelchev authored
-
- 29 Nov, 2019 6 commits
-
-
maxvidal authored
-
Elad Segal authored
-
thomwolf authored
-
thomwolf authored
-
Thomas Wolf authored
Mean does not exist in TF2
-
Thomas Wolf authored
German DistilBERT
-
- 28 Nov, 2019 3 commits
-
-
Thomas Wolf authored
from_pretrained: convert DialoGPT format
-
Thomas Wolf authored
Only init encoder_attention_mask if stack is decoder
-
Thomas Wolf authored
Added Mish Activation Function
-
- 27 Nov, 2019 29 commits
-
-
Thomas Wolf authored
Fix CTRL past
-
thomwolf authored
-
Stefan Schweter authored
-
Yao Lu authored
-
Thomas Wolf authored
-
Thomas Wolf authored
Fix encode_plus
-
Thomas Wolf authored
Add support for resumable downloads for HTTP protocol.
-
Thomas Wolf authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
Julien Chaumond authored
-
Thomas Wolf authored
Consider do_lower_case in PreTrainedTokenizer
-
Thomas Wolf authored
-
R茅mi Louf authored
-
root authored
-