1. 20 Sep, 2020 1 commit
  2. 17 Sep, 2020 1 commit
    • Stas Bekman's avatar
      [ported model] FSMT (FairSeq MachineTranslation) (#6940) · 1eeb206b
      Stas Bekman authored
      * ready for PR
      
      * cleanup
      
      * correct FSMT_PRETRAINED_MODEL_ARCHIVE_LIST
      
      * fix
      
      * perfectionism
      
      * revert change from another PR
      
      * odd, already committed this one
      
      * non-interactive upload workaround
      
      * backup the failed experiment
      
      * store langs in config
      
      * workaround for localizing model path
      
      * doc clean up as in https://github.com/huggingface/transformers/pull/6956
      
      
      
      * style
      
      * back out debug mode
      
      * document: run_eval.py --num_beams 10
      
      * remove unneeded constant
      
      * typo
      
      * re-use bart's Attention
      
      * re-use EncoderLayer, DecoderLayer from bart
      
      * refactor
      
      * send to cuda and fp16
      
      * cleanup
      
      * revert (moved to another PR)
      
      * better error message
      
      * document run_eval --num_beams
      
      * solve the problem of tokenizer finding the right files when model is local
      
      * polish, remove hardcoded config
      
      * add a note that the file is autogenerated to avoid losing changes
      
      * prep for org change, remove unneeded code
      
      * switch to model4.pt, update scores
      
      * s/python/bash/
      
      * missing init (but doesn't impact the finetuned model)
      
      * cleanup
      
      * major refactor (reuse-bart)
      
      * new model, new expected weights
      
      * cleanup
      
      * cleanup
      
      * full link
      
      * fix model type
      
      * merge porting notes
      
      * style
      
      * cleanup
      
      * have to create a DecoderConfig object to handle vocab_size properly
      
      * doc fix
      
      * add note (not a public class)
      
      * parametrize
      
      * - add bleu scores integration tests
      
      * skip test if sacrebleu is not installed
      
      * cache heavy models/tokenizers
      
      * some tweaks
      
      * remove tokens that aren't used
      
      * more purging
      
      * simplify code
      
      * switch to using decoder_start_token_id
      
      * add doc
      
      * Revert "major refactor (reuse-bart)"
      
      This reverts commit 226dad15ca6a9ef4e26178526e878e8fc5c85874.
      
      * decouple from bart
      
      * remove unused code #1
      
      * remove unused code #2
      
      * remove unused code #3
      
      * update instructions
      
      * clean up
      
      * move bleu eval to examples
      
      * check import only once
      
      * move data+gen script into files
      
      * reuse via import
      
      * take less space
      
      * add prepare_seq2seq_batch (auto-tested)
      
      * cleanup
      
      * recode test to use json instead of yaml
      
      * ignore keys not needed
      
      * use the new -y in transformers-cli upload -y
      
      * [xlm tok] config dict: fix str into int to match definition (#7034)
      
      * [s2s] --eval_max_generate_length (#7018)
      
      * Fix CI with change of name of nlp (#7054)
      
      * nlp -> datasets
      
      * More nlp -> datasets
      
      * Woopsie
      
      * More nlp -> datasets
      
      * One last
      
      * extending to support allen_nlp wmt models
      
      - allow a specific checkpoint file to be passed
      - more arg settings
      - scripts for allen_nlp models
      
      * sync with changes
      
      * s/fsmt-wmt/wmt/ in model names
      
      * s/fsmt-wmt/wmt/ in model names (p2)
      
      * s/fsmt-wmt/wmt/ in model names (p3)
      
      * switch to a better checkpoint
      
      * typo
      
      * make non-optional args such - adjust tests where possible or skip when there is no other choice
      
      * consistency
      
      * style
      
      * adjust header
      
      * cards moved (model rename)
      
      * use best custom hparams
      
      * update info
      
      * remove old cards
      
      * cleanup
      
      * s/stas/facebook/
      
      * update scores
      
      * s/allen_nlp/allenai/
      
      * url maps aren't needed
      
      * typo
      
      * move all the doc / build /eval generators to their own scripts
      
      * cleanup
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * fix indent
      
      * duplicated line
      
      * style
      
      * use the correct add_start_docstrings
      
      * oops
      
      * resizing can't be done with the core approach, due to 2 dicts
      
      * check that the arg is a list
      
      * style
      
      * style
      Co-authored-by: default avatarSam Shleifer <sshleifer@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      1eeb206b
  3. 16 Sep, 2020 1 commit
  4. 15 Sep, 2020 1 commit
  5. 14 Sep, 2020 3 commits
  6. 11 Sep, 2020 2 commits
    • Sylvain Gugger's avatar
      Compute loss method (#7074) · 4cbd50e6
      Sylvain Gugger authored
      4cbd50e6
    • Sylvain Gugger's avatar
      Automate the lists in auto-xxx docs (#7061) · e841b75d
      Sylvain Gugger authored
      * More readable dict
      
      * More nlp -> datasets
      
      * Revert "More nlp -> datasets"
      
      This reverts commit 3cd1883d226c63c4a686fc1fed35f2cd586ebe45.
      
      * Automate the lists in auto-xxx docs
      
      * More readable dict
      
      * Revert "More nlp -> datasets"
      
      This reverts commit 3cd1883d226c63c4a686fc1fed35f2cd586ebe45.
      
      * Automate the lists in auto-xxx docs
      
      * nlp -> datasets
      
      * Fix new key
      e841b75d
  7. 10 Sep, 2020 5 commits
  8. 09 Sep, 2020 1 commit
  9. 08 Sep, 2020 2 commits
  10. 03 Sep, 2020 1 commit
  11. 02 Sep, 2020 2 commits
  12. 01 Sep, 2020 6 commits
  13. 27 Aug, 2020 1 commit
  14. 26 Aug, 2020 1 commit
  15. 25 Aug, 2020 1 commit
  16. 24 Aug, 2020 2 commits
  17. 21 Aug, 2020 4 commits
  18. 20 Aug, 2020 2 commits
    • Joe Davison's avatar
      add intro to nlp lib & dataset links to custom datasets tutorial (#6583) · 039d8d65
      Joe Davison authored
      * add intro to nlp lib + links
      
      * unique links...
      039d8d65
    • Romain Rigaux's avatar
      Docs copy button misses ... prefixed code (#6518) · cabfdfaf
      Romain Rigaux authored
      Tested in a local build of the docs.
      
      e.g. Just above https://huggingface.co/transformers/task_summary.html#causal-language-modeling
      
      Copy will copy the full code, e.g.
      
      for token in top_5_tokens:
           print(sequence.replace(tokenizer.mask_token, tokenizer.decode([token])))
      
      Instead of currently only:
      
      for token in top_5_tokens:
      
      
      >>> for token in top_5_tokens:
      ...     print(sequence.replace(tokenizer.mask_token, tokenizer.decode([token])))
      Distilled models are smaller than the models they mimic. Using them instead of the large versions would help reduce our carbon footprint.
      Distilled models are smaller than the models they mimic. Using them instead of the large versions would help increase our carbon footprint.
      Distilled models are smaller than the models they mimic. Using them instead of the large versions would help decrease our carbon footprint.
      Distilled models are smaller than the models they mimic. Using them instead of the large versions would help offset our carbon footprint.
      Distilled models are smaller than the models they mimic. Using them instead of the large versions would help improve our carbon footprint.
      
      Docs for the option fix:
      https://sphinx-copybutton.readthedocs.io/en/latest/
      cabfdfaf
  19. 19 Aug, 2020 1 commit
  20. 18 Aug, 2020 2 commits