- 16 Mar, 2021 11 commits
-
-
Lysandre Debut authored
* Patches the full import failure and adds a test * Add comment
-
Lysandre authored
-
Lysandre authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Add DistributedSamplerWithLoop * Fix typo * Test and small fix
-
Lysandre Debut authored
* Fix DeBERTa-v2 variable assignment * Fix conversational pipeline test
-
Suraj Patil authored
-
Sylvain Gugger authored
-
Lysandre Debut authored
-
Sylvain Gugger authored
* Examples version update * Refactor a bit * All version updates * Fixes * README cleanup * Post-release/patch * Fixes * More fixes * Tests * More fixes * Moar fixes * Make commands and update setup * Replace spaces with weird tabs * Fix test * Style
-
Patrick von Platen authored
* make flax tests pytorch independent * fix typo * finish * improve circle ci * fix return tensors * correct flax test * re-add sentencepiece * last tokenizer fixes * finish maybe now
-
- 15 Mar, 2021 16 commits
-
-
Russell Klopfer authored
* independent training / eval with local files * remove redundant assert
-
Sylvain Gugger authored
* Add minimum version check in examples * Style * No need for new line maybe? * Add helpful comment
-
Joe Davison authored
-
Lysandre Debut authored
* Tests run on Docker Co-authored-by:
Morgan <funtowiczmo@gmail.com> * Comments from code review * Reply to itself * Dependencies Co-authored-by:
Morgan <funtowiczmo@gmail.com>
-
MikeG112 authored
* Update super class reference * Update default value reference * Update src/transformers/models/wav2vec2/feature_extraction_wav2vec2.py * Fix format style Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
-
Patrick von Platen authored
-
Sylvain Gugger authored
* Handle save differently * Missing imports * Fix typo * Adapt to recent changes in save_pretrained * Forgotten brackets * Optimizer load * Fix world size * Deal wth None * Remove needless self
-
Adam Pocock authored
* Adding required flags to non-default arguments. Signed-off-by:
Adam Pocock <adam.pocock@oracle.com> * make style fix. * Update src/transformers/hf_argparser.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Th茅o Matussi猫re authored
* split seq2seq script, update docs * needless diff * fix readme * remove test diff * s/summarization/translation Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * cr * fix arguments & better mbart/t5 refs * copyright Co-authored-by:
Suraj Patil <surajp815@gmail.com> * reword readme Co-authored-by:
Suraj Patil <surajp815@gmail.com> * s/summarization/translation * short script names * fix tests * fix isort, include mbart doc * delete old script, update tests * automate source prefix * automate source prefix for translation * s/translation/trans Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * fix script name (short version) * typos Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * exact parameter Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * remove superfluous source_prefix calls in docs * rename scripts & warn for source prefix * black * flake8 Co-authored-by:
theo <theo@matussie.re> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Suraj Patil <surajp815@gmail.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Igor Shalyminov authored
* GPT2DoubleHeadsModel made parallelizeable * GPT2DoubleHeadsModel added as parallelizeable onto the GPT2 test suite
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
cronoik authored
Documentation was referring to slow tokenizer class while it should be the fast tokenizer.
-
Suraj Patil authored
* enable auto tokenizer for mbart50 tokenizers * fix imports
-
Patrick von Platen authored
-
- 12 Mar, 2021 7 commits
-
-
Stas Bekman authored
-
ymfa authored
* Pass encoder_outputs into generate() * Remove an if-statement * Reformat * Minimize changes to generate() * Comment on input_ids
-
PaulLerner authored
* fix: #10628 expanduser path in TrainingArguments * docs: explain why we expand paths in TrainingArguments * Style Co-authored-by:Sylvain Gugger <sylvain.gugger@gmail.com>
-
Sylvain Gugger authored
* Add auto_wrap option in fairscale integration * Style
-
Lysandre Debut authored
* TF model exists for Blenderbot 400M * Marian * RAG
-
Nicolas Patry authored
* [WIP] Adding new parameter to `generate`: `max_time`. Generation by tokens number is sometimes a bit clunky because we don't know how many tokens are good enough or even how many tokens are in the payload (for pipelines users for instance). This leads to hard to understand behavior. This PR proposes a new argument `max_time` which is a float of seconds for the allowed time for `generate` to run on. Ideally combinations of `max_tokens=None`, `max_time=2` could be used to generate as many tokens as possible within time budget. NB: Another possible approach consists of passing a callback to `generate` putting the caller in charge of the actual decision of when to stop generating tokens. It opens the door to 'which args should we pass' to this callback. It's hard to imagine other use-cases for this early stopping behavior than time (that are not already covered by parameters of generate) * Revamp with StoppingCriteria * Removing deprecated mentions. * Forgot arguments to stopping criteria. * Readding max_length it's not just used as a stopping criteria. * Default value for `stopping_criteria`. * Address @patrickvonplaten comments. - More docstrings - Actual doc - Include in global namespace - Remove TF work. * Put back `max_length` (deprecation different PR). * Doc quality. * Fixing old behavior without `stopping_criteria` but with `max_length`. Making sure we don't break that in the future. * Adding more tests for possible inconsistencies between `max_length` and `stopping_criteria`. * Fixing the torch imports.
-
Lysandre Debut authored
-
- 11 Mar, 2021 6 commits
-
-
Benjamin Fineran authored
* fix typing error for TrainingArguments Optional[bool] * updating equality check for Optional[bool]
-
Sylvain Gugger authored
-
WybeKoper authored
* Fixed broken link * fixed max length violation Co-authored-by:WybeKoper <WybeKoper@users.noreply.github.com>
-
jeswan authored
* add deberta to pretraining mapping * add deberta_v2 to PRETRAINING_MAPPING
-
Lysandre Debut authored
-
Sylvain Gugger authored
* PoC * Fix slow tests for the PT1.8 Embedding problem
-