- 14 Jan, 2020 15 commits
-
-
Lysandre authored
-
Julien Chaumond authored
-
Lysandre authored
Created a link between the linear layer bias and the model attribute bias. This does not change anything for the user nor for the conversion scripts, but allows the `resize_token_embeddings` method to resize the bias as well as the weights of the decoder. Added a test.
-
Lysandre authored
-
Lysandre authored
-
Lysandre authored
-
Lysandre authored
-
Lysandre authored
-
Lysandre authored
-
Lysandre Debut authored
-
Thomas Wolf authored
Releasing file lock
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
- 13 Jan, 2020 9 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Thomas Wolf authored
Update pipelines.py
-
Morgan Funtowicz authored
-
- 11 Jan, 2020 8 commits
-
-
IWillPull authored
* Added repetition penalty * Default PPLM repetition_penalty to neutral * Minor modifications to comply with reviewer's suggestions. (j -> token_idx) * Formatted code with `make style`
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
VictorSanh authored
-
VictorSanh authored
-
- 10 Jan, 2020 8 commits
-
-
Julien Chaumond authored
-
Thomas Wolf authored
T5: move rp_bucket to relative_attention_bias' device
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Martin Schrimpf authored
otherwise, `rp_bucket` will always be on cpu and fail if `self.relative_attention_bias` is on cuda
-
Julien Chaumond authored
-
Victor SANH authored
-
Victor SANH authored
-