1. 02 Mar, 2021 1 commit
  2. 01 Mar, 2021 5 commits
    • Lysandre Debut's avatar
      Add I-BERT to README (#10462) · 0c232519
      Lysandre Debut authored
      0c232519
    • Lysandre Debut's avatar
      9248e270
    • Suraj Patil's avatar
      [Wav2Vec2FeatureExtractor] smal fixes (#10455) · a106bde5
      Suraj Patil authored
      * smal fixes
      
      * don't check for None
      a106bde5
    • Patrick von Platen's avatar
      11655faf
    • Patrick von Platen's avatar
      Add Fine-Tuning for Wav2Vec2 (#10145) · 0234de84
      Patrick von Platen authored
      
      
      * add encode labels function to tokenizer
      
      * start adding finetuning
      
      * init dropout
      
      * upload
      
      * correct convert script
      
      * apply changes
      
      * fix second typo
      
      * make first dummy training run
      
      * adapt convert script
      
      * push confg for comparison
      
      * remove conf
      
      * finish training
      
      * adapt data collator
      
      * add research folder
      
      * update according to fairseq feedback
      
      * some minor corrections
      
      * refactor masking indices a bit
      
      * some minor changes
      
      * clean tokenizer
      
      * finish clean-up
      
      * remove previous logic
      
      * update run script
      
      * correct training
      
      * finish changes
      
      * finish model
      
      * correct bug
      
      * fix training a bit more
      
      * add some tests
      
      * finish gradient checkpointing
      
      * finish example
      
      * correct gradient checkpointing
      
      * improve tokenization method
      
      * revert changes in tokenizer
      
      * revert general change
      
      * adapt fine-tuning
      
      * update
      
      * save intermediate test
      
      * Update README.md
      
      * finish finetuning
      
      * delete conversion script
      
      * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py
      
      * Update src/transformers/models/wav2vec2/processing_wav2vec2.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * finish wav2vec2 script
      
      * finish wav2vec2 fine-tuning
      
      * finalize test
      
      * correct test
      
      * adapt tests
      
      * finish
      
      * remove test file
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      0234de84
  3. 28 Feb, 2021 3 commits
  4. 27 Feb, 2021 5 commits
  5. 26 Feb, 2021 4 commits
  6. 25 Feb, 2021 11 commits
    • Sylvain Gugger's avatar
    • Sylvain Gugger's avatar
      Make Barthez tokenizer tests a bit faster (#10399) · 26f8b2cb
      Sylvain Gugger authored
      * Make Barthez tokenizer tests a bit faster
      
      * Quality
      26f8b2cb
    • Andrea Bacciu's avatar
      Fix None in add_token_positions - issue #10210 (#10374) · b040e6ef
      Andrea Bacciu authored
      * Fix None in add_token_positions - issue #10210
      
      Fix None in add_token_positions related to the issue #10210
      
      * add_token_positions fix None values in end_positions vector
      
      add_token_positions fix None in end_positions vector as proposed by @joeddav
      b040e6ef
    • Sylvain Gugger's avatar
      Add support for ZeRO-2/3 and ZeRO-offload in fairscale (#10354) · 9d14be5c
      Sylvain Gugger authored
      
      
      * Ass support for ZeRO-2/3 and ZeRO-offload in fairscale
      
      * Quality
      
      * Rework from review comments
      
      * Add doc
      
      * Apply suggestions from code review
      Co-authored-by: default avatarStas Bekman <stas00@users.noreply.github.com>
      
      * Address review comments
      Co-authored-by: default avatarStas Bekman <stas00@users.noreply.github.com>
      9d14be5c
    • Lysandre Debut's avatar
      88cc26dc
    • Sehoon Kim's avatar
      I-BERT model support (#10153) · 63645b3b
      Sehoon Kim authored
      
      
      * IBertConfig, IBertTokentizer added
      
      * IBert Model names moified
      
      * tokenizer bugfix
      
      * embedding -> QuantEmbedding
      
      * quant utils added
      
      * quant_mode added to configuration
      
      * QuantAct added, Embedding layer + QuantAct addition
      
      * QuantAct added
      
      * unused path removed, QKV quantized
      
      * self attention layer all quantized, except softmax
      
      * temporarl commit
      
      * all liner layers quantized
      
      * quant_utils bugfix
      
      * bugfix: requantization missing
      
      * IntGELU added
      
      * IntSoftmax added
      
      * LayerNorm implemented
      
      * LayerNorm implemented all
      
      * names changed: roberta->ibert
      
      * config not inherit from ROberta
      
      * No support for CausalLM
      
      * static quantization added, quantize_model.py removed
      
      * import modules uncommented
      
      * copyrights fixed
      
      * minor bugfix
      
      * quant_modules, quant_utils merged as one file
      
      * import * fixed
      
      * unused runfile removed
      
      * make style run
      
      * configutration.py docstring fixed
      
      * refactoring: comments removed, function name fixed
      
      * unused dependency removed
      
      * typo fixed
      
      * comments(Copied from), assertion string added
      
      * refactoring: super(..) -> super(), etc.
      
      * refactoring
      
      * refarctoring
      
      * make style
      
      * refactoring
      
      * cuda -> to(x.device)
      
      * weight initialization removed
      
      * QuantLinear set_param removed
      
      * QuantEmbedding set_param removed
      
      * IntLayerNorm set_param removed
      
      * assert string added
      
      * assertion error message fixed
      
      * is_decoder removed
      
      * enc-dec arguments/functions removed
      
      * Converter removed
      
      * quant_modules docstring fixed
      
      * conver_slow_tokenizer rolled back
      
      * quant_utils docstring fixed
      
      * unused aruments e.g. use_cache removed from config
      
      * weight initialization condition fixed
      
      * x_min, x_max initialized with small values to avoid div-zero exceptions
      
      * testing code for ibert
      
      * test emb, linear, gelu, softmax added
      
      * test ln and act added
      
      * style reformatted
      
      * force_dequant added
      
      * error tests overrided
      
      * make style
      
      * Style + Docs
      
      * force dequant tests added
      
      * Fix fast tokenizer in init
      
      * Fix doc
      
      * Remove space
      
      * docstring, IBertConfig, chunk_size
      
      * test_modeling_ibert refactoring
      
      * quant_modules.py refactoring
      
      * e2e integration test added
      
      * tokenizers removed
      
      * IBertConfig added to tokenizer_auto.py
      
      * bugfix
      
      * fix docs & test
      
      * fix style num 2
      
      * final fixes
      Co-authored-by: default avatarSehoon Kim <sehoonkim@berkeley.edu>
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      63645b3b
    • Patrick von Platen's avatar
      [PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor,... · cb38ffcc
      Patrick von Platen authored
      [PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor, Wav2Vec2Tokenizer (#10324)
      
      * push to show
      
      * small improvement
      
      * small improvement
      
      * Update src/transformers/feature_extraction_utils.py
      
      * Update src/transformers/feature_extraction_utils.py
      
      * implement base
      
      * add common tests
      
      * make all tests pass for wav2vec2
      
      * make padding work & add more tests
      
      * finalize feature extractor utils
      
      * add call method to feature extraction
      
      * finalize feature processor
      
      * finish tokenizer
      
      * finish general processor design
      
      * finish tests
      
      * typo
      
      * remove bogus file
      
      * finish docstring
      
      * add docs
      
      * finish docs
      
      * small fix
      
      * correct docs
      
      * save intermediate
      
      * load changes
      
      * apply changes
      
      * apply changes to doc
      
      * change tests
      
      * apply surajs recommend
      
      * final changes
      
      * Apply suggestions from code review
      
      * fix typo
      
      * fix import
      
      * correct docstring
      cb38ffcc
    • abhishek thakur's avatar
      9dc78257
    • mingruimingrui's avatar
      Bugfix: Removal of padding_idx in BartLearnedPositionalEmbedding (#10200) · 894db670
      mingruimingrui authored
      
      
      * Assumption of padding_idx <2 might not stand
      
      * Use offset instead of 2
      
      * Fix with black
      
      * Change behavior to warning instead for backward compatibility.
      
      * Fix with black
      
      * Remove warning
      
      * Make padding_idx non-required
      
      * padding_idx fix for blenderbot
      
      * padding_idx fix for blenderbot_small
      
      * padding_idx fix for led
      
      * padding_idx fix for mbart
      
      * Remove extra whitespaces
      
      * padding_idx fix for template
      
      * Fix padding_idx passed to nn.Embedding mistake
      
      * Fixed padding_idx passed to positional embedding in template
      
      * Remove padding_idx from pytorch learned positional embeddings
      
      * Remove accidentally added quotes
      
      * Remove padding_idx from tf learned positional embeddings
      
      * Remove zeroing of weights in __init__
      Co-authored-by: default avatarWang Ming Rui <mingrui.wang@C02CJTUYMD6M.local>
      894db670
    • Lysandre Debut's avatar
      55fe80d0
    • Lysandre Debut's avatar
      22bd047e
  7. 24 Feb, 2021 6 commits
  8. 23 Feb, 2021 3 commits
  9. 22 Feb, 2021 2 commits