"utils/vscode:/vscode.git/clone" did not exist on "45f56580a7e11b5b894374f8e1c7bdd54d982682"
- 03 Feb, 2021 1 commit
-
-
abhishek thakur authored
-
- 02 Feb, 2021 9 commits
-
-
Daniel Stancl authored
* Add {decoder_,}head_mask to LED * Fix create_custom_forward signatue in encoder * Add head_mask to longformer * Add head_mask to longformer to fix dependencies of LED on Longformer. * Not working yet * Add mising one input in longofrmer_modeling.py * make fix-copies -
Patrick von Platen authored
* add raw scaffold * implement feat extract layers * make style * remove + * correctly convert weights * make feat extractor work * make feature extraction proj work * run forward pass * finish forward pass * Succesful decoding example * remove unused files * more changes * add wav2vec tokenizer * add new structure * fix run forward * add other layer norm architecture * finish 2nd structure * add model tests * finish tests for tok and model * clean-up * make style * finish docstring for model and config * make style * correct docstring * correct tests * change checkpoints to fairseq * fix examples * finish wav2vec2 * make style * apply sylvains suggestions * apply lysandres suggestions * change print to log.info * re-add assert statement * add input_values as required input name * finish wav2vec2 tokenizer * Update tests/test_tokenization_wav2vec2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * apply sylvains suggestions Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Sylvain Gugger authored
-
Stefan Schweter authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Initial work * Fix doc styler and other models
-
Lysandre Debut authored
* ALBERT Tokenizer integration test * Batching * Style
-
Patrick von Platen authored
-
Patrick von Platen authored
* change tokenizer requirement * split line * Correct typo from list to str * improve style * make other function pretty as well * add comment * correct typo * add new test * pass tests for tok without padding token * Apply suggestions from code review
-
- 01 Feb, 2021 11 commits
-
-
Jan Jitse Venselaar authored
* Change documentation to correctly specify loss tensor size * Change documentation to correct input format for labels * Corrected output size of loss tensor for sequence classifier, multiple choice model and question answering
-
Suraj Patil authored
* fix conversion script * typo * import nn
-
Patrick von Platen authored
* add new model logic * fix docs * change structure * improve add_new_model * push new changes * up * up * correct spelling * improve docstring * correct line length * update readme * correct links * correct typos * only add rst file for now * Apply suggestions from code review 1 Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Bram Vanroy <Bram.Vanroy@UGent.be> * Apply suggestions from code review Co-authored-by:
Bram Vanroy <Bram.Vanroy@UGent.be> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stefan Schweter <stefan@schweter.it> Co-authored-by:
Bram Vanroy <Bram.Vanroy@UGent.be> * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Pierric Cistac <Pierrci@users.noreply.github.com> * finish adding all suggestions * make style * apply Niels feedback * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * apply sylvains suggestions Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Bram Vanroy <Bram.Vanroy@UGent.be> Co-authored-by:
Stefan Schweter <stefan@schweter.it> Co-authored-by:
Pierric Cistac <Pierrci@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Suraj Patil authored
-
CeShine Lee authored
This affects Adafactor with relative_step=False and scale_parameter=True. Updating group["lr"] makes the result of ._get_lr() depends on the previous call, i.e., on the scale of other parameters. This isn't supposed to happen.
-
Sylvain Gugger authored
* Remove subclass for sortish sampler * Use old Seq2SeqTrainer in script * Styling
-
wlhgtc authored
* MOD: fit chinese wwm to new datasets * MOD: move wwm to new folder * MOD: formate code * Styling * MOD add param and recover trainer Co-authored-by:Sylvain Gugger <sylvain.gugger@gmail.com>
-
Stas Bekman authored
* [t5 doc] typos a few run away backticks @sgugger * style * [trainer] put fp16 args together this PR proposes a purely cosmetic change that puts all the fp16 args together - so they are easier to manager/read @sgugger * style * [wandb] make WANDB_DISABLED disable wandb with any value This PR solves part of https://github.com/huggingface/transformers/issues/9623 It tries to actually do what https://github.com/huggingface/transformers/issues/9699 requested/discussed and that is any value of `WANDB_DISABLED` should disable wandb. The current behavior is that it has to be one of `ENV_VARS_TRUE_VALUES = {"1", "ON", "YES"}` I have been using `WANDB_DISABLED=true` everywhere in scripts as it was originally advertised. I have no idea why this was changed to a sub-set of possible values. And it's not documented anywhere. @sgugger * WANDB_DISABLED=true to disable; make tf trainer consistent * style
-
Stas Bekman authored
-
Sylvain Gugger authored
-
Daniel Stancl authored
* Add {decoder_,}head_mask to fsmt_modeling.py * Enable test_headmasking and some changes to docs * Remove test_head_masking flag from fsmt test file Remove test_head_masking flag from test_modeling_fsmt.py since test_head_masking is set to be True by default (thus it is redundant to store). * Merge master and remove test_head_masking = True * Rebase necessary due to an update of jaxlib * Remove test_head_masking=True in tests/test_modeling_fsmt.py as it is redundant.
-
- 31 Jan, 2021 2 commits
-
-
Kiyoung Kim authored
* TFBart lables consider both pad token and -100 * make style * fix for all other models Co-authored-by: kykim <kykim> Co-authored-by:patrickvonplaten <patrick.v.platen@gmail.com>
-
lewtun authored
* Clarify definition of seed argument in Trainer * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args_tf.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Fix style * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 30 Jan, 2021 1 commit
-
-
Stas Bekman authored
Apparently nested markup in RST is invalid: https://docutils.sourceforge.io/FAQ.html#is-nested-inline-markup-possible So currently this line doesn't get rendered properly, leaving inner markdown unrendered, resulting in: ``` https://docutils.sourceforge.io/FAQ.html#is-nested-inline-markup-possible ``` This PR removes the bold which fixes the link.
-
- 29 Jan, 2021 6 commits
-
-
Stas Bekman authored
-
Stas Bekman authored
-
Sylvain Gugger authored
* When on sagemaker use their env variables for saves * Address review comments * Quality
-
Julien Plu authored
-
Ethan Chau authored
-
Nicolas Patry authored
* Adding a new `return_full_text` parameter to TextGenerationPipeline. For text-generation, it's sometimes used as prompting text. In that context, prefixing `generated_text` with the actual input forces the caller to take an extra step to remove it. The proposed change adds a new parameter (for backward compatibility). `return_full_text` that enables the caller to prevent adding the prefix. * Doc quality.
-
- 28 Jan, 2021 10 commits
-
-
abhishek thakur authored
-
abhishek thakur authored
-
Stas Bekman authored
* expand install instructions * fix * white space * rewrite as discussed in the PR * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * change the wording to encourage issue report Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Daniel Stancl authored
* Remove redundant test_head_masking = True flags * Remove all redundant test_head_masking flags in PyTorch test_modeling_* files * Make test_head_masking = True as a default choice in test_modeling_tf_commong.py * Remove all redundant test_head_masking flags in TensorFlow test_modeling_tf_* files * Put back test_head_masking=False fot TFT5 models
-
Joe Davison authored
-
Sylvain Gugger authored
-
Funtowicz Morgan authored
* Fix computation of attention_probs when head_mask is provided. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Apply changes to the template Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr>
-
Nicolas Patry authored
-
Lysandre Debut authored
-
Lysandre Debut authored
* Allow partial loading of a cached tokenizer * Warning > Info * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Raise error if not local_files_only Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-