1. 15 Mar, 2021 2 commits
  2. 12 Mar, 2021 1 commit
  3. 11 Mar, 2021 2 commits
  4. 10 Mar, 2021 2 commits
  5. 09 Mar, 2021 1 commit
  6. 08 Mar, 2021 4 commits
  7. 06 Mar, 2021 1 commit
  8. 05 Mar, 2021 1 commit
  9. 04 Mar, 2021 3 commits
  10. 01 Mar, 2021 1 commit
    • Patrick von Platen's avatar
      Add Fine-Tuning for Wav2Vec2 (#10145) · 0234de84
      Patrick von Platen authored
      
      
      * add encode labels function to tokenizer
      
      * start adding finetuning
      
      * init dropout
      
      * upload
      
      * correct convert script
      
      * apply changes
      
      * fix second typo
      
      * make first dummy training run
      
      * adapt convert script
      
      * push confg for comparison
      
      * remove conf
      
      * finish training
      
      * adapt data collator
      
      * add research folder
      
      * update according to fairseq feedback
      
      * some minor corrections
      
      * refactor masking indices a bit
      
      * some minor changes
      
      * clean tokenizer
      
      * finish clean-up
      
      * remove previous logic
      
      * update run script
      
      * correct training
      
      * finish changes
      
      * finish model
      
      * correct bug
      
      * fix training a bit more
      
      * add some tests
      
      * finish gradient checkpointing
      
      * finish example
      
      * correct gradient checkpointing
      
      * improve tokenization method
      
      * revert changes in tokenizer
      
      * revert general change
      
      * adapt fine-tuning
      
      * update
      
      * save intermediate test
      
      * Update README.md
      
      * finish finetuning
      
      * delete conversion script
      
      * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py
      
      * Update src/transformers/models/wav2vec2/processing_wav2vec2.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * finish wav2vec2 script
      
      * finish wav2vec2 fine-tuning
      
      * finalize test
      
      * correct test
      
      * adapt tests
      
      * finish
      
      * remove test file
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      0234de84
  11. 27 Feb, 2021 3 commits
  12. 25 Feb, 2021 3 commits
  13. 24 Feb, 2021 1 commit
  14. 23 Feb, 2021 1 commit
  15. 22 Feb, 2021 3 commits
  16. 19 Feb, 2021 2 commits
  17. 18 Feb, 2021 2 commits
  18. 17 Feb, 2021 1 commit
  19. 16 Feb, 2021 1 commit
  20. 15 Feb, 2021 2 commits
  21. 12 Feb, 2021 1 commit
  22. 11 Feb, 2021 2 commits
    • Stas Bekman's avatar
      [DeepSpeed in notebooks] Jupyter + Colab (#10130) · b54cb0bd
      Stas Bekman authored
      * init devices/setup explicitly
      
      * docs + test
      
      * simplify
      
      * cleanup
      
      * cleanup
      
      * cleanup
      
      * correct the required dist setup
      
      * derive local_rank from env LOCAL_RANK
      b54cb0bd
    • Qbiwan's avatar
      Update run_xnli.py to use Datasets library (#9829) · 8dcfaea0
      Qbiwan authored
      * remove xnli_compute_metrics, add load_dataset, load_metric, set_seed,metric.compute,load_metric
      
      * fix
      
      * fix
      
      * fix
      
      * push
      
      * fix
      
      * everything works
      
      * fix init
      
      * fix
      
      * special treatment for sepconv1d
      
      * style
      
      * 馃檹馃徑
      
      * add doc and cleanup
      
      
      * fix doc
      
      * fix doc again
      
      * fix doc again
      
      * Apply suggestions from code review
      
      * make style
      
      * Proposal that should work
      
      * Remove needless code
      
      * Fix test
      
      * Apply suggestions from code review
      
      * remove xnli_compute_metrics, add load_dataset, load_metric, set_seed,metric.compute,load_metric
      
      * amend README
      
      * removed data_args.task_name and replaced with task_name = "xnli"; use split function to load train and validation dataset separately; remove __post_init__; remove flag --task_name from README.
      
      * removed dict task_to_keys, use str "xnli" instead of variable task_name, change preprocess_function to use examples["premise"], examples["hypothesis"] directly, remove sentence1_key and sentence2_key, change compute_metrics function to cater only to accuracy metric, add condition for train_langauge is None when using dataset.load_dataset()
      
      * removed `torch.distributed.barrier()` and `import torch` as `from_pretrained` is able to do the work; amend README
      8dcfaea0