- 06 Jul, 2020 2 commits
-
-
Arnav Sharma authored
-
ELanning authored
-
- 03 Jul, 2020 10 commits
-
-
Sam Shleifer authored
-
Patrick von Platen authored
-
Thomas Wolf authored
-
Thomas Wolf authored
-
Lysandre Debut authored
* Exposing prepare_for_model for both slow & fast tokenizers * Update method signature * The traditional style commit * Hide the warnings behind the verbose flag * update default truncation strategy and prepare_for_model * fix tests and prepare_for_models methods Co-authored-by:Thomas Wolf <thomwolf@users.noreply.github.com>
-
Manuel Romero authored
Create model card for electicidad-small (Spanish Electra) fine-tuned on SQUAD-esv1
-
Moseli Motsoehli authored
- fixed grammar and spelling - added an intro - updated Training data references
-
chrisliu authored
-
Manuel Romero authored
Create model card for electra-small-discriminator fine-tuned on SQUAD v2.0
-
Funtowicz Morgan authored
* Make QA pipeline supports models with more than 2 outputs such as BART assuming start/end are the two first outputs. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * When using the new padding/truncation paradigm setting padding="max_length" + max_length=X actually pads the input up to max_length. This result in every sample going through QA pipelines to be of size 384 whatever the actual input size is making the overall pipeline very slow. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Mask padding & question before applying softmax. Softmax has been refactored to operate in log space for speed and stability. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Format. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Use PaddingStrategy.LONGEST instead of DO_NOT_PAD Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Revert "When using the new padding/truncation paradigm setting padding="max_length" + max_length=X actually pads the input up to max_length." This reverts commit 1b00a9a2 Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Trigger CI after unattended failure * Trigger CI
-
- 02 Jul, 2020 11 commits
-
-
Pierric Cistac authored
-
Sylvain Gugger authored
* Work on tokenizer summary * Finish tutorial * Link to it * Apply suggestions from code review Co-authored-by:
Anthony MOI <xn1t0x@gmail.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Add vocab definition Co-authored-by:
Anthony MOI <xn1t0x@gmail.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Shen authored
`ElectraDiscriminatorPredictions.forward` should not need `attention_mask`.
-
Manuel Romero authored
Create model card for electra-base-discriminator fine-tuned on SQUAD v1.1
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Lysandre Debut authored
-
George Ho authored
-
Teven authored
* overriding _parse_and_tokenize in `TextGenerationPipeine` to allow for TransfoXl tokenizer arguments
-
Teven authored
* Changed expected_output_ids in TransfoXL generation test to match #4826 generation PR. * making black happy * making isort happy
-
tommccoy authored
Fixed duplicated memory use in Transformer-XL generation leading to bad predictions and performance.
-
- 01 Jul, 2020 17 commits
-
-
Patrick von Platen authored
* Add Reformer MLM notebook * Update notebooks/README.md
-
Patrick von Platen authored
* fix conflicts * fix * happy rebasing
-
Funtowicz Morgan authored
Signed-off-by:Morgan Funtowicz <funtowiczmo@gmail.com>
-
Joe Davison authored
* allow tensor label inputs to default collator * replace try/except with type check
-
Patrick von Platen authored
-
Patrick von Platen authored
* refactor naming * add small slow test * refactor * refactor naming * rename selected to extra * big global attention refactor * make style * refactor naming * save intermed * refactor functions * finish function refactor * fix tests * fix longformer * fix longformer * fix longformer * fix all tests but one * finish longformer * address sams and izs comments * fix transpose
-
Sam Shleifer authored
-
Funtowicz Morgan authored
* Added PipelineException Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * fill-mask pipeline raises exception when more than one mask_token detected. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Put everything in a function. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Added tests on pipeline fill-mask when input has != 1 mask_token Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Fix numel() computation for TF Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Addressing PR comments. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Remove function typing to avoid import on specific framework. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Quality. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Retry typing with @julien-c tip. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Quality虏. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Simplify fill-mask mask_token checking. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * Trigger CI
-
Sylvain Gugger authored
* Trigger CI * Fix dropdown bug in searches
-
Sylvain Gugger authored
* Cleanup and unify Trainer/TFTrainer * Forgot to adapt TFTrainingArgs * In tf scripts n_gpu -> n_replicas * Update src/transformers/training_args.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Address review comments * Formatting * Fix typo Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Sam Shleifer authored
-
Sam Shleifer authored
-
sgugger authored
-
Sylvain Gugger authored
* Add support for past states * Style and forgotten self * You mean, documenting is not enough? I have to actually add it too? * Add memory support during evaluation * Fix tests in eval and add TF support * No need to change this line anymore
-
Sylvain Gugger authored
-
Moseli Motsoehli authored
* Create README.md * Update model_cards/MoseliMotsoehli/TswanaBert/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Jay authored
-