- 25 Feb, 2021 5 commits
-
-
Patrick von Platen authored
[PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor, Wav2Vec2Tokenizer (#10324) * push to show * small improvement * small improvement * Update src/transformers/feature_extraction_utils.py * Update src/transformers/feature_extraction_utils.py * implement base * add common tests * make all tests pass for wav2vec2 * make padding work & add more tests * finalize feature extractor utils * add call method to feature extraction * finalize feature processor * finish tokenizer * finish general processor design * finish tests * typo * remove bogus file * finish docstring * add docs * finish docs * small fix * correct docs * save intermediate * load changes * apply changes * apply changes to doc * change tests * apply surajs recommend * final changes * Apply suggestions from code review * fix typo * fix import * correct docstring
-
abhishek thakur authored
-
mingruimingrui authored
* Assumption of padding_idx <2 might not stand * Use offset instead of 2 * Fix with black * Change behavior to warning instead for backward compatibility. * Fix with black * Remove warning * Make padding_idx non-required * padding_idx fix for blenderbot * padding_idx fix for blenderbot_small * padding_idx fix for led * padding_idx fix for mbart * Remove extra whitespaces * padding_idx fix for template * Fix padding_idx passed to nn.Embedding mistake * Fixed padding_idx passed to positional embedding in template * Remove padding_idx from pytorch learned positional embeddings * Remove accidentally added quotes * Remove padding_idx from tf learned positional embeddings * Remove zeroing of weights in __init__ Co-authored-by:Wang Ming Rui <mingrui.wang@C02CJTUYMD6M.local>
-
Lysandre Debut authored
-
Lysandre Debut authored
-
- 24 Feb, 2021 6 commits
-
-
Lysandre authored
-
Stas Bekman authored
* move secondary methods into a separate file * cleanup * style
-
Poedator authored
This is to fix deprecated reference to `tokenizer.max_len` with `tokenizer.model_max_length` - similar to [issue 8739](https://github.com/huggingface/transformers/issues/8739) and [PR 8604](https://github.com/huggingface/transformers/pull/8604). Example [here](https://colab.research.google.com/gist/poedator/f8776349e5c625ce287fc6fcd312fa1e/tokenizer-max_len-error-in-transformers_glue.ipynb). The error happens when `glue_convert_examples_to_features` is called without `max_length` parameter specified. In that case line 119 with wrong reference gets called. This simple fix should do it.
-
Julien Plu authored
-
abhishek thakur authored
* convbert conversion test * fin * fin * fin * clean up tf<->pt conversion * remove from_pt Co-authored-by:patrickvonplaten <patrick.v.platen@gmail.com>
-
Stas Bekman authored
* handle get_last_lr() before first step() * abstract away the lr getting logic * cleanup * add test * move to utils
-
- 23 Feb, 2021 3 commits
-
-
Julien Chaumond authored
-
Akmal authored
-
Lysandre authored
-
- 22 Feb, 2021 9 commits
-
-
Sylvain Gugger authored
-
Stas Bekman authored
* make logging and saving trainer built-in * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Tanmay Garg authored
Enhance resume_from_checkpoint argument of Trainer.train to accept bool type. If True given, last saved checkpoint in self.args.output_dir will be loaded. (#10280)
-
Stas Bekman authored
* implement gradient_accumulation_steps support in DeepSpeed integration * typo * cleanup * cleanup
-
Stas Bekman authored
-
Sylvain Gugger authored
* Deprecate prepare_seq2seq_batch * Fix last tests * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Suraj Patil <surajp815@gmail.com> * More review comments Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Suraj Patil <surajp815@gmail.com>
-
Lysandre Debut authored
-
Julien Plu authored
* AMP * Add LED * Apply style * Fix longformer
-
Lysandre Debut authored
Co-authored-by:
Pengcheng He <penhe@microsoft.com> Co-authored-by:
Pengcheng He <penhe@microsoft.com>
-
- 21 Feb, 2021 1 commit
-
-
tagucci authored
* fix typo in conversion script * style Co-authored-by:Stas Bekman <stas@stason.org>
-
- 20 Feb, 2021 3 commits
-
-
Stas Bekman authored
-
Sylvain Gugger authored
-
cronoik authored
-
- 19 Feb, 2021 13 commits
-
-
Pengcheng He authored
* Integrate DeBERTa v2(the 1.5B model surpassed human performance on SuperGLUE); Add DeBERTa v2 900M,1.5B models; * DeBERTa-v2 * Fix v2 model loading issue (#10129) * Doc members * Update src/transformers/models/deberta/modeling_deberta.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Address Sylvain's comments * Address Patrick's comments Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Style Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
-
Julien Plu authored
-
Joe Davison authored
-
Stas Bekman authored
-
Tanmay Garg authored
Introduce logging_strategy training argument in TrainingArguments and TFTrainingArguments. (#9838)
-
Julien Plu authored
* Fix AMP and XLA * Remove useless var
-
Julien Plu authored
* Fix AMP * Apply style * Remove unused import
-
Julien Plu authored
-
Julien Plu authored
* Fix XLA * Rework cast * Apply style
-
Julien Plu authored
* Fix AMP * Trigger CI * Rework cast
-
Julien Plu authored
* Fix AMP * Rework cast * Apply style
-
Stas Bekman authored
* propose using google colab to reproduce problems * Update ISSUES.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-