- 11 May, 2020 4 commits
-
-
Julien Plu authored
-
theblackcat102 authored
-
Patrick von Platen authored
* adapt convert script * update convert script * finish * fix marian pretrained docs
-
Patrick von Platen authored
-
- 10 May, 2020 3 commits
-
-
flozi00 authored
-
Sam Shleifer authored
- MarianSentencepieceTokenizer - > MarianTokenizer - Start using unk token. - add docs page - add better generation params to MarianConfig - more conversion utilities
-
Girishkumar authored
-
- 08 May, 2020 9 commits
-
-
Julien Chaumond authored
* [TPU] Doc, fix xla_spawn.py, only preprocess dataset once * Update examples/README.md * [xla_spawn] Add `_mp_fn` to other Trainer scripts * [TPU] Fix: eval dataloader was None
-
Julien Chaumond authored
-
Lorenzo De Mattei authored
* example updated to use generation pipeline * Update model_cards/LorenzoDeMattei/GePpeTto/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
rmroczkowski authored
-
rmroczkowski authored
-
Manuel Romero authored
-
Sava艧 Y谋ld谋r谋m authored
* Create README.md * Adding code fence around code block
-
Manuel Romero authored
model card for my De Novo Drug discovery model using MLM
-
Patrick von Platen authored
* fix PR * move tests to correct place
-
- 07 May, 2020 17 commits
-
-
Jared T Nielsen authored
* Add AlbertForPreTraining and TFAlbertForPreTraining models. * PyTorch conversion * TensorFlow conversion * style Co-authored-by:Lysandre <lysandre.debut@reseau.eseo.fr>
-
Julien Chaumond authored
-
Sava艧 Y谋ld谋r谋m authored
-
Julien Chaumond authored
-
Julien Chaumond authored
* README * Update README.md
-
Lysandre authored
-
Lysandre authored
-
Julien Chaumond authored
* Created using Colaboratory * [examples] reorganize files * remove run_tpu_glue.py as superseded by TPU support in Trainer * Bugfix: int, not tuple * move files around
-
Julien Chaumond authored
cc @patrickvonplaten @thomwolf
-
Lysandre Debut authored
* wip * wip * a last wip * Better logging when using TPUs * Correct argument name * Tests * fix * Metrics in evaluation * Update src/transformers/training_args.py * [tpu] Use launcher script instead * [tpu] lots of tweaks * Fix formatting Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Funtowicz Morgan authored
Ensure fast tokenizer can construct tensor without pad token if only one sample is provided. (#4201)
-
Funtowicz Morgan authored
* Rewritten batch support in pipelines. Signed-off-by:
Morgan Funtowicz <morgan@huggingface.co> * Fix imports sorting
馃敡 Signed-off-by:Morgan Funtowicz <morgan@huggingface.co> * Set pad_to_max_length=True by default on Pipeline. * Set pad_to_max_length=False for generation pipelines. Most of generation models doesn't have padding token. * Address @joeddav review comment: Uniformized *args. Signed-off-by:
Morgan Funtowicz <morgan@huggingface.co> * Address @joeddav review comment: Uniformized *args (second). Signed-off-by:
Morgan Funtowicz <morgan@huggingface.co>
-
Patrick von Platen authored
-
Patrick von Platen authored
* fix example reformer * fix error message and example docstring * improved error message
-
Patrick von Platen authored
-
Patrick von Platen authored
* first copy & past commit from Bert and morgans LSH code * add easy way to compare to trax original code * translate most of function * make trax lsh self attention deterministic with numpy seed + copy paste code * add same config * add same config * make layer init work * implemented hash_vectors function for lsh attention * continue reformer translation * hf LSHSelfAttentionLayer gives same output as trax layer * refactor code * refactor code * refactor code * refactor * refactor + add reformer config * delete bogus file * split reformer attention layer into two layers * save intermediate step * save intermediate step * make test work * add complete reformer block layer * finish reformer layer * implement causal and self mask * clean reformer test and refactor code * fix merge conflicts * fix merge conflicts * update init * fix device for GPU * fix chunk length init for tests * include morgans optimi...
-
- 06 May, 2020 7 commits
-
-
Clement authored
-
Julien Plu authored
* First commit to add a TF version of the trainer. * Make the TF trainer closer to what looks the PT trainer * Refactoring common code between the PT and TF trainer into an util file. * Some bugfix + better similarity with the PT trainer * Add missing class in transformers init * Bugfix over prediction + use classification report instead of simple metrics * Fix name error * Fix optimization tests + style * Apply style * Several bugfix for multi-gpu training * Apply style * Apply style * Add glue example for the TF trainer * Several bugix + address the reviews * Fix on the TF training args file * Add a debug mode * Bugfix in utils_ner.py when segment_ids is None * Apply style * Apply style * Add TPU strategy * Fix selection strategy
-
Simone Primarosa authored
-
kumapo authored
-
martindh authored
Description for the model card describing the camembert-large-fquad model.
-
Julien Plu authored
-
Manuel Romero authored
-