- 10 Dec, 2019 23 commits
-
-
thomwolf authored
-
Thomas Wolf authored
[WIP] Squad refactor
-
Thomas Wolf authored
create encoder attention mask from shape of hidden states
-
thomwolf authored
-
Julien Chaumond authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
-
Rémi Louf authored
We currently save the pretrained_weights of the encoder and decoder in two separate directories `encoder` and `decoder`. However, for the `from_pretrained` function to operate with automodels we need to specify the type of model in the path to the weights. The path to the encoder/decoder weights is handled by the `PreTrainedEncoderDecoder` class in the `save_pretrained` function. Sice there is no easy way to infer the type of model that was initialized for the encoder and decoder we add a parameter `model_type` to the function. This is not an ideal solution as it is error prone, and the model type should be carried by the Model classes somehow. This is a temporary fix that should be changed before merging.
-
Rémi Louf authored
Since I started my PR the `add_special_token_single_sequence` function has been deprecated for another; I replaced it with the new function.
-
- 09 Dec, 2019 14 commits
-
-
Pierric Cistac authored
-
Bilal Khan authored
-
Bilal Khan authored
-
Bilal Khan authored
-
Bilal Khan authored
-
Bilal Khan authored
-
Bilal Khan authored
-
thomwolf authored
-
thomwolf authored
-
thomwolf authored
-
LysandreJik authored
-
Lysandre Debut authored
-
thomwolf authored
-
Rémi Louf authored
We currently create encoder attention masks (when they're not provided) based on the shape of the inputs to the encoder. This is obviously wrong; sequences can be of different lengths. We now create the encoder attention mask based on the batch_size and sequence_length of the encoder hidden states.
-
- 07 Dec, 2019 1 commit
-
-
Aymeric Augustin authored
-
- 06 Dec, 2019 2 commits
-
-
Michael Watkins authored
-
Aymeric Augustin authored
* Switch to plain unittest for skipping slow tests. Add a RUN_SLOW environment variable for running them. * Switch to plain unittest for PyTorch dependency. * Switch to plain unittest for TensorFlow dependency. * Avoid leaking open files in the test suite. This prevents spurious warnings when running tests. * Fix unicode warning on Python 2 when running tests. The warning was: UnicodeWarning: Unicode equal comparison failed to convert both arguments to Unicode - interpreting them as being unequal * Support running PyTorch tests on a GPU. Reverts 27e015bd. * Tests no longer require pytest. * Make tests pass on cuda
-