- 30 Apr, 2021 6 commits
-
-
Manuel Romero authored
Add link to code
-
Patrick von Platen authored
-
Philip May authored
add test for pickle simplify test fix test code style add missing pickle import fix test fix test fix test
-
Frederik Bode authored
Co-authored-by:Frederik Bode <frederik@paperbox.ai>
-
Lysandre Debut authored
-
Lysandre authored
-
- 29 Apr, 2021 4 commits
-
-
Sylvain Gugger authored
* Split checkpoint from model_name_or_path in examples * Address review comments * Address review comments
-
Michael Benayoun authored
Co-authored-by:Michael Benayoun <michael@huggingface.co>
-
Sylvain Gugger authored
* Reformat to make code clearer * Reformat to make code clearer
-
Patrick von Platen authored
* add attentions & hidden states * add model outputs + docs * finish docs * finish tests * finish impl * del @ * finish * finish * correct test * apply sylvains suggestions * Update src/transformers/models/bert/modeling_flax_bert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * simplify more Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 28 Apr, 2021 3 commits
-
-
Hamel Husain authored
-
Hamel Husain authored
* Update tokenization_utils_base.py * add assertion * check batch len * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * add error message Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
* Update min versions in README and add Flax * Adapt index
-
- 27 Apr, 2021 3 commits
-
-
Suraj Patil authored
* fix docs for decoder_input_ids * revert the changes for bart and mbart
-
Hamel Husain authored
* finish quicktour * fix import * fix print * explain config default better * Update docs/source/quicktour.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Hamel Husain authored
* update docs to reflect model output object * run make style`
-
- 26 Apr, 2021 20 commits
-
-
Ashwin Geet D'Sa authored
* removed max_len * removed max_length from BeamSearchScorer * correct max length * finish * del vim * finish & add test Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
Stas Bekman authored
* adding Z-inf * revamp config process * up version requirement * wip * massive rewrite * cleanup * cleanup * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * consistent json commas * act on suggestions * leave this feature for 0.3.16 * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Jaimeen Ahn authored
As the error comes from the inconsistency of variable meaning number of gpus in parser and its actual usage in the train.py script, 'gpus' and 'n_gpu' respectively, the correction makes the example work
-
Bhadresh Savani authored
* added changes for uniformity * modified files * corrected typo * fixed qa scripts * fix typos * fixed predict typo in qa no trainer * fixed test file * reverted trainer changes * reverted trainer changes in custom exmaples * updated readme * added changes in deepspeed test * added changes for predict and eval
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Stas Bekman authored
* fix invalid class name * proper ref * proper ref
-
Kostas Stathoulopoulos authored
* Improve documentation for is_split_into_words argument * Change description wording
-
Sylvain Gugger authored
* Pass along seed to DistributedSampler * Add seed to DistributedLengthGroupedSampler
-
LSinev authored
-
Amine Abdaoui authored
-
Sylvain Gugger authored
* Add FP16 support for SageMaker MP * Add print debugs * Squeeze * Remove debug statements * Add defensive check * Typo
-
Daniel Stancl authored
TF BART models - Add `cross_attentions` to model output and fix cross-attention head masking (#10699) * Add cross_attn_head_mask to BART * Fix cross_attentions in TFBart-like models * This commit enables returning of `cross_attentions` for TFBart-like models * It also fixes attention head masking in cross-attenion module * Update TF model templates * Fix missing , in TF model templates * Fix typo: congig -> config
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Patrick von Platen authored
-
Vasudev Gupta authored
-
abiolaTresor authored
-
- 25 Apr, 2021 2 commits
-
-
cronoik authored
* removes the creation of separate config objects and uses the existing ones instead+overwrite resize_token_embeddings from parent class because it is not working for the EncoderDecoderModel * rollback to current version of the huggingface master branch * reworked version that ties the encoder and decoder config of the parent encoderdecoder instance * overwrite of resize_token_embeddings throws an error now * review comment suggestion Co-authored-by:
Suraj Patil <surajp815@gmail.com> * implemented warning in case encoderdecoder is created with differing configs of encoderdecoderconfig and decoderconfig or encoderconfig * added test to avoid diverging configs of wrapper class and wrapped classes * Update src/transformers/models/encoder_decoder/modeling_encoder_decoder.py * make style Co-authored-by:
Suraj Patil <surajp815@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Daniel Stancl authored
* Add head_mask & decoder_head_mask + some corrections * Fix head masking for N-grams * Enable test_headmasking for encoder and decod * Fix one typo regarding in modeling_propgetnet.py * Enable test_headmasking for ProphetNetStandaloneDecoderModelTest and ProphetNetStandaloneEncoderModelTest in test_modeling_prophetnet.py * make style * Fix cross_head_mask * Fix attention head mask naming * `cross_head_mask` -> `cross_attn_head_mask` * `cross_layer_head_mask` -> `cross_attn_layer_head_mask` * Still need to merge #10605 to master to pass the tests
-
- 24 Apr, 2021 2 commits
-
-
Sylvain Gugger authored
-
cronoik authored
documentation linked to the parent class PreTrainedTokenizerFast but it should be the slow tokenizer (#11410)
-