"vscode:/vscode.git/clone" did not exist on "d6b6ab11f04eec306a7386d2b52bd31018902088"
- 04 Mar, 2021 3 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
This reverts commit f3660613.
-
Sylvain Gugger authored
-
- 03 Mar, 2021 9 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Fix gradient accumulation for SM Model Parallelism * Style and divide loss by grad accum steps
-
felixgwu authored
* fix all_hidden_states * use output_states instead of next_kv
-
Stas Bekman authored
* remap classes to strings * missing new util * style * doc * move the autogenerated file * Trigger CI
-
Sylvain Gugger authored
* Refactor checkpoint name in BERT and MobileBERT * Add option to check copies * Add QuestionAnswering * Add last models * Make black happy
-
Jeff Yang authored
* feat(docs): navigate with left/right arrow keys * fix: add missing comma
-
Patrick von Platen authored
* fix speed degradation bug t5 * fix for all models * fix code quality
-
WybeKoper authored
Co-authored-by:WybeKoper <WybeKoper@users.noreply.github.com>
-
Mehrad Moradshahi authored
-
- 02 Mar, 2021 1 commit
-
-
Martin Schmitt authored
Changed `num_beams` to `num_beams // num_beam_groups` when initialising `PrefixConstrainedLogitsProcessor` in `_get_logits_processor` to fix compatibility issue when constrained decoding is used together with grouped beam search (#10475)
-
- 01 Mar, 2021 5 commits
-
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Suraj Patil authored
* smal fixes * don't check for None
-
Patrick von Platen authored
-
Patrick von Platen authored
* add encode labels function to tokenizer * start adding finetuning * init dropout * upload * correct convert script * apply changes * fix second typo * make first dummy training run * adapt convert script * push confg for comparison * remove conf * finish training * adapt data collator * add research folder * update according to fairseq feedback * some minor corrections * refactor masking indices a bit * some minor changes * clean tokenizer * finish clean-up * remove previous logic * update run script * correct training * finish changes * finish model * correct bug * fix training a bit more * add some tests * finish gradient checkpointing * finish example * correct gradient checkpointing * improve tokenization method * revert changes in tokenizer * revert general change * adapt fine-tuning * update * save intermediate test * Update README.md * finish finetuning * delete conversion script * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py * Update src/transformers/models/wav2vec2/processing_wav2vec2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * finish wav2vec2 script * finish wav2vec2 fine-tuning * finalize test * correct test * adapt tests * finish * remove test file Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 28 Feb, 2021 3 commits
-
-
Patrick von Platen authored
-
Darigov Research authored
* feat: Adds three definitions to glossary from @cronoik Needed a definition for transformer which in turn needed 2 more definitions To do with issue https://github.com/huggingface/transformers/issues/9078 * fix: Adjusts definition of neural network to make it easier to read
-
Tanmay Garg authored
* Introduce save_strategy training argument * deprecate EvaluationStrategy * collapse EvaluationStrategy and LoggingStrategy into a single IntervalStrategy enum * modify tests to use modified enum
-
- 27 Feb, 2021 5 commits
-
-
Bhadresh Savani authored
* updated logging and saving metrics * space removal
-
Stas Bekman authored
This PR restores the original functionality that for some reason was modified. Fixes: https://github.com/huggingface/transformers/issues/10381 @sgugger
-
Lysandre Debut authored
-
Stas Bekman authored
* refactors * typo
-
Amog Kamsetty authored
* fixes * update resources * formatting * remove import * add log statement * use fstring * add period * Update src/transformers/integrations.py
-
- 26 Feb, 2021 4 commits
-
-
Kai Fricke authored
-
Patrick von Platen authored
* correct docs * correct tf model docs as well
-
Mansi Mane authored
* Added tb fix * Removed local rank condition * Updated reference to args
-
Julien Chaumond authored
馃槀
-
- 25 Feb, 2021 10 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Make Barthez tokenizer tests a bit faster * Quality
-
Andrea Bacciu authored
* Fix None in add_token_positions - issue #10210 Fix None in add_token_positions related to the issue #10210 * add_token_positions fix None values in end_positions vector add_token_positions fix None in end_positions vector as proposed by @joeddav
-
Sylvain Gugger authored
* Ass support for ZeRO-2/3 and ZeRO-offload in fairscale * Quality * Rework from review comments * Add doc * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Address review comments Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Lysandre Debut authored
-
Sehoon Kim authored
* IBertConfig, IBertTokentizer added * IBert Model names moified * tokenizer bugfix * embedding -> QuantEmbedding * quant utils added * quant_mode added to configuration * QuantAct added, Embedding layer + QuantAct addition * QuantAct added * unused path removed, QKV quantized * self attention layer all quantized, except softmax * temporarl commit * all liner layers quantized * quant_utils bugfix * bugfix: requantization missing * IntGELU added * IntSoftmax added * LayerNorm implemented * LayerNorm implemented all * names changed: roberta->ibert * config not inherit from ROberta * No support for CausalLM * static quantization added, quantize_model.py removed * import modules uncommented * copyrights fixed * minor bugfix * quant_modules, quant_utils merged as one file * import * fixed * unused runfile removed * make style run * configutration.py docstring fixed * refactoring: comments removed, function name fixed * unused dependency removed * typo fixed * comments(Copied from), assertion string added * refactoring: super(..) -> super(), etc. * refactoring * refarctoring * make style * refactoring * cuda -> to(x.device) * weight initialization removed * QuantLinear set_param removed * QuantEmbedding set_param removed * IntLayerNorm set_param removed * assert string added * assertion error message fixed * is_decoder removed * enc-dec arguments/functions removed * Converter removed * quant_modules docstring fixed * conver_slow_tokenizer rolled back * quant_utils docstring fixed * unused aruments e.g. use_cache removed from config * weight initialization condition fixed * x_min, x_max initialized with small values to avoid div-zero exceptions * testing code for ibert * test emb, linear, gelu, softmax added * test ln and act added * style reformatted * force_dequant added * error tests overrided * make style * Style + Docs * force dequant tests added * Fix fast tokenizer in init * Fix doc * Remove space * docstring, IBertConfig, chunk_size * test_modeling_ibert refactoring * quant_modules.py refactoring * e2e integration test added * tokenizers removed * IBertConfig added to tokenizer_auto.py * bugfix * fix docs & test * fix style num 2 * final fixes Co-authored-by:
Sehoon Kim <sehoonkim@berkeley.edu> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Sylvain Gugger <sylvain.gugger@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Patrick von Platen authored
[PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor, Wav2Vec2Tokenizer (#10324) * push to show * small improvement * small improvement * Update src/transformers/feature_extraction_utils.py * Update src/transformers/feature_extraction_utils.py * implement base * add common tests * make all tests pass for wav2vec2 * make padding work & add more tests * finalize feature extractor utils * add call method to feature extraction * finalize feature processor * finish tokenizer * finish general processor design * finish tests * typo * remove bogus file * finish docstring * add docs * finish docs * small fix * correct docs * save intermediate * load changes * apply changes * apply changes to doc * change tests * apply surajs recommend * final changes * Apply suggestions from code review * fix typo * fix import * correct docstring
-
abhishek thakur authored
-
mingruimingrui authored
* Assumption of padding_idx <2 might not stand * Use offset instead of 2 * Fix with black * Change behavior to warning instead for backward compatibility. * Fix with black * Remove warning * Make padding_idx non-required * padding_idx fix for blenderbot * padding_idx fix for blenderbot_small * padding_idx fix for led * padding_idx fix for mbart * Remove extra whitespaces * padding_idx fix for template * Fix padding_idx passed to nn.Embedding mistake * Fixed padding_idx passed to positional embedding in template * Remove padding_idx from pytorch learned positional embeddings * Remove accidentally added quotes * Remove padding_idx from tf learned positional embeddings * Remove zeroing of weights in __init__ Co-authored-by:Wang Ming Rui <mingrui.wang@C02CJTUYMD6M.local>
-
Lysandre Debut authored
-