- 22 Apr, 2020 2 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
* doc * [tests] Add sample files for a regression task * [HUGE] Trainer * Feedback from @sshleifer * Feedback from @thomwolf + logging tweak * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes * [glue] Use default max_seq_length of 128 like before * [glue] move DataTrainingArguments around * [ner] Change interface of InputExample, and align run_{tf,pl} * Re-align the pl scripts a little bit * ner * [ner] Add integration test * Fix language_modeling with API tweak * [ci] Tweak loss target * Don't break console output * amp.initialize: model must be on right device before * [multiple-choice] update for Trainer * Re-align to 827d6d6e
-
- 21 Apr, 2020 4 commits
-
-
Julien Chaumond authored
-
Spencer Adams authored
* create readme for spentaur/yelp model * update spentaur/yelp/README.md * remove typo
-
Julien Chaumond authored
-
Bharat Raghunathan authored
-
- 20 Apr, 2020 10 commits
-
-
Andrey Kulagin authored
-
husein zolkepli authored
(cherry picked from commit b5f2dc5d627d44b8cbb0ccf8ad2b46bea211a236)
-
Punyajoy Saha authored
The first model added to the repo
-
Manuel Romero authored
-
Funtowicz Morgan authored
Introduce tqdm_enabled parameter on squad_convert_examples_to_features() default to True and set to False in QA pipelines.
-
Jared T Nielsen authored
* Add qas_id * Fix incorrect name in squad.py * Make output files optional for squad eval
-
Patrick von Platen authored
* remove max_length = tokenizer.max_length when encoding * make style
-
Mohamed El-Geish authored
* exbert links for my albert model cards * Added exbert tag to the metadata block * Adding "how to cite"
-
Sam Shleifer authored
-
ahotrod authored
-
- 18 Apr, 2020 6 commits
-
-
Patrick von Platen authored
-
Thomas Wolf authored
* First pass on utility classes and python tokenizers * finishing cleanup pass * style and quality * Fix tests * Updating following @mfuntowicz comment * style and quality * Fix Roberta * fix batch_size/seq_length inBatchEncoding * add alignement methods + tests * Fix OpenAI and Transfo-XL tokenizers * adding trim_offsets=True default for GPT2 et RoBERTa * style and quality * fix tests * add_prefix_space in roberta * bump up tokenizers to rc7 * style * unfortunately tensorfow does like these - removing shape/seq_len for now * Update src/transformers/tokenization_utils.py Co-Authored-By:
Stefan Schweter <stefan@schweter.it> * Adding doc and docstrings * making flake8 happy Co-authored-by:
Stefan Schweter <stefan@schweter.it>
-
-
Julien Chaumond authored
-
Benjamin Muller authored
-
Patrick von Platen authored
* better config serialization * finish configuration utils
-
- 17 Apr, 2020 8 commits
-
-
Lysandre Debut authored
* XLM tokenizer should encode with bos token * Update tests
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Harutaka Kawamura authored
-
Santiago Castro authored
* Add support for the null answer in `QuestionAnsweringPipeline` * black * Fix min null score computation * Fix a PR comment
-
Simon B枚hm authored
token_type_id is converted into the segment embedding. For question answering, this needs to highlight whether a token belongs to sequence 0 or 1. encode_plus takes care of correctly setting this parameter automatically.
-
Pierric Cistac authored
* Add TFAlbertForQuestionAnswering * Add TFRobertaForQuestionAnswering * Update TFAutoModel with Roberta/Albert for QA * Clean `super` TF Albert calls
-
Patrick von Platen authored
-
- 16 Apr, 2020 10 commits
-
-
Sam Shleifer authored
renames `run_bart_sum.py` to `finetune.py`
-
Jonathan Sum authored
Changing from "fine-grained token-leven" to "fine-grained token-level"
-
Aryansh Omray authored
-
Sam Shleifer authored
-
Patrick von Platen authored
* Refactored use of newstest2013 to newstest2014. Fixed bug where argparse consumed first command line argument as model_size argument rather than using default model_size by forcing explicit --model_size flag inclusion * More pythonic file handling through 'with' context * COSMETIC - ran Black and isort * Fixed reference to number of lines in newstest2014 * Fixed failing test. More pythonic file handling * finish PR from tholiao * remove outcommented lines * make style * make isort happy Co-authored-by:Thomas Liao <tholiao@gmail.com>
-
Lysandre Debut authored
-
Davide Fiocco authored
-
Patrick von Platen authored
-
Patrick von Platen authored
* correct gpt2 test inputs * make style * delete modeling_gpt2 change in test file * translate from pytorch * correct tests * fix conflicts * fix conflicts * fix conflicts * fix conflicts * make tensorflow t5 caching work * make style * clean reorder cache * remove unnecessary spaces * fix test
-
Patrick von Platen authored
-