1. 01 Mar, 2021 1 commit
    • Patrick von Platen's avatar
      Add Fine-Tuning for Wav2Vec2 (#10145) · 0234de84
      Patrick von Platen authored
      
      
      * add encode labels function to tokenizer
      
      * start adding finetuning
      
      * init dropout
      
      * upload
      
      * correct convert script
      
      * apply changes
      
      * fix second typo
      
      * make first dummy training run
      
      * adapt convert script
      
      * push confg for comparison
      
      * remove conf
      
      * finish training
      
      * adapt data collator
      
      * add research folder
      
      * update according to fairseq feedback
      
      * some minor corrections
      
      * refactor masking indices a bit
      
      * some minor changes
      
      * clean tokenizer
      
      * finish clean-up
      
      * remove previous logic
      
      * update run script
      
      * correct training
      
      * finish changes
      
      * finish model
      
      * correct bug
      
      * fix training a bit more
      
      * add some tests
      
      * finish gradient checkpointing
      
      * finish example
      
      * correct gradient checkpointing
      
      * improve tokenization method
      
      * revert changes in tokenizer
      
      * revert general change
      
      * adapt fine-tuning
      
      * update
      
      * save intermediate test
      
      * Update README.md
      
      * finish finetuning
      
      * delete conversion script
      
      * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py
      
      * Update src/transformers/models/wav2vec2/processing_wav2vec2.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * finish wav2vec2 script
      
      * finish wav2vec2 fine-tuning
      
      * finalize test
      
      * correct test
      
      * adapt tests
      
      * finish
      
      * remove test file
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      0234de84
  2. 28 Feb, 2021 1 commit
  3. 26 Feb, 2021 2 commits
  4. 25 Feb, 2021 3 commits
    • Sylvain Gugger's avatar
      Make Barthez tokenizer tests a bit faster (#10399) · 26f8b2cb
      Sylvain Gugger authored
      * Make Barthez tokenizer tests a bit faster
      
      * Quality
      26f8b2cb
    • Sehoon Kim's avatar
      I-BERT model support (#10153) · 63645b3b
      Sehoon Kim authored
      
      
      * IBertConfig, IBertTokentizer added
      
      * IBert Model names moified
      
      * tokenizer bugfix
      
      * embedding -> QuantEmbedding
      
      * quant utils added
      
      * quant_mode added to configuration
      
      * QuantAct added, Embedding layer + QuantAct addition
      
      * QuantAct added
      
      * unused path removed, QKV quantized
      
      * self attention layer all quantized, except softmax
      
      * temporarl commit
      
      * all liner layers quantized
      
      * quant_utils bugfix
      
      * bugfix: requantization missing
      
      * IntGELU added
      
      * IntSoftmax added
      
      * LayerNorm implemented
      
      * LayerNorm implemented all
      
      * names changed: roberta->ibert
      
      * config not inherit from ROberta
      
      * No support for CausalLM
      
      * static quantization added, quantize_model.py removed
      
      * import modules uncommented
      
      * copyrights fixed
      
      * minor bugfix
      
      * quant_modules, quant_utils merged as one file
      
      * import * fixed
      
      * unused runfile removed
      
      * make style run
      
      * configutration.py docstring fixed
      
      * refactoring: comments removed, function name fixed
      
      * unused dependency removed
      
      * typo fixed
      
      * comments(Copied from), assertion string added
      
      * refactoring: super(..) -> super(), etc.
      
      * refactoring
      
      * refarctoring
      
      * make style
      
      * refactoring
      
      * cuda -> to(x.device)
      
      * weight initialization removed
      
      * QuantLinear set_param removed
      
      * QuantEmbedding set_param removed
      
      * IntLayerNorm set_param removed
      
      * assert string added
      
      * assertion error message fixed
      
      * is_decoder removed
      
      * enc-dec arguments/functions removed
      
      * Converter removed
      
      * quant_modules docstring fixed
      
      * conver_slow_tokenizer rolled back
      
      * quant_utils docstring fixed
      
      * unused aruments e.g. use_cache removed from config
      
      * weight initialization condition fixed
      
      * x_min, x_max initialized with small values to avoid div-zero exceptions
      
      * testing code for ibert
      
      * test emb, linear, gelu, softmax added
      
      * test ln and act added
      
      * style reformatted
      
      * force_dequant added
      
      * error tests overrided
      
      * make style
      
      * Style + Docs
      
      * force dequant tests added
      
      * Fix fast tokenizer in init
      
      * Fix doc
      
      * Remove space
      
      * docstring, IBertConfig, chunk_size
      
      * test_modeling_ibert refactoring
      
      * quant_modules.py refactoring
      
      * e2e integration test added
      
      * tokenizers removed
      
      * IBertConfig added to tokenizer_auto.py
      
      * bugfix
      
      * fix docs & test
      
      * fix style num 2
      
      * final fixes
      Co-authored-by: default avatarSehoon Kim <sehoonkim@berkeley.edu>
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      63645b3b
    • Patrick von Platen's avatar
      [PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor,... · cb38ffcc
      Patrick von Platen authored
      [PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor, Wav2Vec2Tokenizer (#10324)
      
      * push to show
      
      * small improvement
      
      * small improvement
      
      * Update src/transformers/feature_extraction_utils.py
      
      * Update src/transformers/feature_extraction_utils.py
      
      * implement base
      
      * add common tests
      
      * make all tests pass for wav2vec2
      
      * make padding work & add more tests
      
      * finalize feature extractor utils
      
      * add call method to feature extraction
      
      * finalize feature processor
      
      * finish tokenizer
      
      * finish general processor design
      
      * finish tests
      
      * typo
      
      * remove bogus file
      
      * finish docstring
      
      * add docs
      
      * finish docs
      
      * small fix
      
      * correct docs
      
      * save intermediate
      
      * load changes
      
      * apply changes
      
      * apply changes to doc
      
      * change tests
      
      * apply surajs recommend
      
      * final changes
      
      * Apply suggestions from code review
      
      * fix typo
      
      * fix import
      
      * correct docstring
      cb38ffcc
  5. 24 Feb, 2021 1 commit
  6. 22 Feb, 2021 2 commits
  7. 19 Feb, 2021 9 commits
  8. 18 Feb, 2021 4 commits
  9. 17 Feb, 2021 4 commits
  10. 16 Feb, 2021 2 commits
  11. 15 Feb, 2021 5 commits
  12. 13 Feb, 2021 1 commit
    • Nicolas Patry's avatar
      Conversion from slow to fast for BPE spm vocabs contained an error. (#10120) · c9837a0d
      Nicolas Patry authored
      * Conversion from slow to fast for BPE spm vocabs contained an error.
      
      - There is only 1 test currently (tokenizers + slow) that used the modified path
      and it's reformer, which does not contain any ids modification so the
      bug was silent for now.
      - The real issue is that vocab variable was overloaded by
      SentencePieceExtractor, leading to Slow specific vocab oddities to be
      completely ignored
      - The bug was reported here https://github.com/huggingface/transformers/issues/9518
      - Ran the complete tokenization test suite with slow without error
      (`RUN_SLOW=1 pytest -sv tests/test_tokenization_*`)
      
      * Remove rebase error.
      
      * Adding the fixture.
      c9837a0d
  13. 12 Feb, 2021 2 commits
  14. 11 Feb, 2021 1 commit
  15. 10 Feb, 2021 2 commits
    • Suraj Patil's avatar
      remove adjust_logits_during_generation method (#10087) · c130e67d
      Suraj Patil authored
      * add forced logits processors
      
      * delete adjust_logits method
      
      * add forced_eos_token_id argument in config
      
      * add tests for forced logits processors
      
      * update gen utils tests
      
      * add forced option to tf generate
      
      * remove adjust_logits method from tf models
      
      * update adjust_logits for marian
      
      * delete _force_token_id_to_be_generated method
      
      * style
      
      * import warnings
      
      * pass max_length to _get_logits_processor
      
      * set forced_eos_token_id to None
      
      * set forced attributes in conf utils
      
      * typo
      
      * fix rag generate
      
      * add forced_eos_token_id in rag config
      
      * remove force_bos_token_to_be_generated from BartConfig
      
      * remove _force_token_ids_generation from FSMT
      
      * nit
      
      * fix negative constant
      
      * apply suggestions from code review
      c130e67d
    • Julien Plu's avatar
      Fix TF LED/Longformer attentions computation (#10007) · 22a32cf4
      Julien Plu authored
      * Fix test
      
      * Remove commented test
      
      * Fix name
      
      * Apply style
      
      * Fix check copies
      
      * Remove prints
      
      * Restore boolean
      
      * Fix reshape
      22a32cf4