- 26 May, 2020 3 commits
-
-
Patrick von Platen authored
* add new longformer for question answering model * add new config as well * fix links * fix links part 2
-
ZhuBaohe authored
* fix * fix1
-
ZhuBaohe authored
-
- 25 May, 2020 6 commits
-
-
Sam Shleifer authored
-
Patrick von Platen authored
* fix reformer num buckets * fix * adapt docs * set num buckets in config
-
Elman Mansimov authored
-
Suraj Patil authored
-
Sho Arora authored
-
Suraj Patil authored
* added LongformerForQuestionAnswering * add LongformerForQuestionAnswering * fix import for LongformerForMaskedLM * add LongformerForQuestionAnswering * hardcoded sep_token_id * compute attention_mask if not provided * combine global_attention_mask with attention_mask when provided * update example in docstring * add assert error messages, better attention combine * add test for longformerForQuestionAnswering * typo * cast gloabl_attention_mask to long * make style * Update src/transformers/configuration_longformer.py * Update src/transformers/configuration_longformer.py * fix the code quality * Merge branch 'longformer-for-question-answering' of https://github.com/patil-suraj/transformers into longformer-for-question-answering Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 23 May, 2020 1 commit
-
-
Bharat Raghunathan authored
-
- 22 May, 2020 9 commits
-
-
Bijay Gurung authored
* Add Type Hints to modeling_utils.py Closes #3911 Add Type Hints to methods in `modeling_utils.py` Note: The coverage isn't 100%. Mostly skipped internal methods. * Reformat according to `black` and `isort` * Use typing.Iterable instead of Sequence * Parameterize Iterable by its generic type * Use typing.Optional when None is the default value * Adhere to style guideline * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Funtowicz Morgan authored
* Warn the user about max_len being on the path to be deprecated. * Ensure better backward compatibility when max_len is provided to a tokenizer. * Make sure to override the parameter and not the actual instance value. * Format & quality
-
Sam Shleifer authored
* Fix pipelines defaults bug * one liner * style
-
Julien Chaumond authored
As discussed w/ @lysandrejik packaging is maintained by PyPA (the Python Packaging Authority), and should be lightweight and stable
-
Lysandre authored
-
Anthony MOI authored
-
Lysandre authored
-
Lysandre authored
-
Frankie Liuzzi authored
* added functionality for electra classification head * unneeded dropout * Test ELECTRA for sequence classification * Style Co-authored-by:
Frankie <frankie@frase.io> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr>
-
- 21 May, 2020 3 commits
-
-
Lysandre authored
-
Lysandre Debut authored
* TPU hangs when saving optimizer/scheduler * Style * ParallelLoader is not a DataLoader * Style * Addressing @julien-c's comments
-
Zhangyx authored
Adds predict stage for glue tasks, and generate result files which can be submitted to gluebenchmark.com (#4463) * Adds predict stage for glue tasks, and generate result files which could be submitted to gluebenchmark.com website. * Use Split enum + always output the label name Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 20 May, 2020 4 commits
-
-
Cola authored
Remove warning of deprecated overload of addcdiv_ Fix #4451
-
Julien Plu authored
* Better None gradients handling * Apply Style * Apply Style
-
Oliver 脜strand authored
* Exclude LayerNorms from weight decay * Include both formats of layer norm
-
Rens authored
-
- 19 May, 2020 10 commits
-
-
Sam Shleifer authored
-
Patrick von Platen authored
* add longformer docs * improve docs
-
Patrick von Platen authored
* fix gpu slow tests in pytorch * change model to device syntax
-
Sam Shleifer authored
-
Iz Beltagy authored
* first commit * bug fixes * better examples * undo padding * remove wrong VOCAB_FILES_NAMES * License * make style * make isort happy * unit tests * integration test * make `black` happy by undoing `isort` changes!! * lint * no need for the padding value * batch_size not bsz * remove unused type casting * seqlen not seq_len * staticmethod * `bert` selfattention instead of `n2` * uint8 instead of bool + lints * pad inputs_embeds using embeddings not a constant * black * unit test with padding * fix unit tests * remove redundant unit test * upload model weights * resolve todo * simpler _mask_invalid_locations without lru_cache + backward compatible masked_fill_ * increase unittest coverage
-
Shaoyen authored
* Map optimizer to correct device after loading from checkpoint. * Make style test pass Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Julien Chaumond authored
-
Julien Chaumond authored
* Distributed eval: SequentialDistributedSampler + gather all results * For consistency only write to disk from world_master Close https://github.com/huggingface/transformers/issues/4272 * Working distributed eval * Hook into scripts * Fix #3721 again * TPU.mesh_reduce: stay in tensor space Thanks @jysohn23 * Just a small comment * whitespace * torch.hub: pip install packaging * Add test scenarii
-
Julien Chaumond authored
* Test case for #3936 * multigpu tests pass on pytorch 1.4.0 * Fixup * multigpu tests pass on pytorch 1.5.0 * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py * rename multigpu to require_multigpu * mode doc
-
Rakesh Chada authored
* makes fetching last learning late in trainer backward compatible * split comment to multiple lines * fixes black styling issue * uses version to create a more explicit logic
-
- 18 May, 2020 4 commits
-
-
Funtowicz Morgan authored
* Adding optimizations block from ONNXRuntime. * Turn off external data format by default for PyTorch export. * Correct the way use_external_format is passed through the cmdline args.
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
* fix fp16 in t5 * make style * refactor invert_attention_mask fn * fix typo
-