1. 22 Sep, 2020 3 commits
  2. 21 Sep, 2020 1 commit
  3. 18 Sep, 2020 1 commit
    • Dat Quoc Nguyen's avatar
      Add new pre-trained models BERTweet and PhoBERT (#6129) · af2322c7
      Dat Quoc Nguyen authored
      * Add BERTweet and PhoBERT models
      
      * Update modeling_auto.py
      
      Re-add `bart` to LM_MAPPING
      
      * Update tokenization_auto.py
      
      Re-add `from .configuration_mobilebert import MobileBertConfig`
      not sure why it's replaced by `from transformers.configuration_mobilebert import MobileBertConfig`
      
      * Add BERTweet and PhoBERT to pretrained_models.rst
      
      * Update tokenization_auto.py
      
      Remove BertweetTokenizer and PhobertTokenizer out of tokenization_auto.py (they are currently not supported by AutoTokenizer.
      
      * Update BertweetTokenizer - without nltk
      
      * Update model card for BERTweet
      
      * PhoBERT - with Auto mode - without import fastBPE
      
      * PhoBERT - with Auto mode - without import fastBPE
      
      * BERTweet - with Auto mode - without import fastBPE
      
      * Add PhoBERT and BERTweet to TF modeling auto
      
      * Improve Docstrings for PhobertTokenizer and BertweetTokenizer
      
      * Update PhoBERT and BERTweet model cards
      
      * Fixed a merge conflict in tokenization_auto
      
      * Used black to reformat BERTweet- and PhoBERT-related files
      
      * Used isort to reformat BERTweet- and PhoBERT-related files
      
      * Reformatted BERTweet- and PhoBERT-related files based on flake8
      
      * Updated test files
      
      * Updated test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Update commits from huggingface
      
      * Delete unnecessary files
      
      * Add tokenizers to auto and init files
      
      * Add test files for tokenizers
      
      * Revised model cards
      
      * Update save_vocabulary function in BertweetTokenizer and PhobertTokenizer and test files
      
      * Revised test files
      
      * Update orders of Phobert and Bertweet tokenizers in auto tokenization file
      af2322c7
  4. 17 Sep, 2020 3 commits
    • Stas Bekman's avatar
      999a1c95
    • Stas Bekman's avatar
      [ported model] FSMT (FairSeq MachineTranslation) (#6940) · 1eeb206b
      Stas Bekman authored
      * ready for PR
      
      * cleanup
      
      * correct FSMT_PRETRAINED_MODEL_ARCHIVE_LIST
      
      * fix
      
      * perfectionism
      
      * revert change from another PR
      
      * odd, already committed this one
      
      * non-interactive upload workaround
      
      * backup the failed experiment
      
      * store langs in config
      
      * workaround for localizing model path
      
      * doc clean up as in https://github.com/huggingface/transformers/pull/6956
      
      
      
      * style
      
      * back out debug mode
      
      * document: run_eval.py --num_beams 10
      
      * remove unneeded constant
      
      * typo
      
      * re-use bart's Attention
      
      * re-use EncoderLayer, DecoderLayer from bart
      
      * refactor
      
      * send to cuda and fp16
      
      * cleanup
      
      * revert (moved to another PR)
      
      * better error message
      
      * document run_eval --num_beams
      
      * solve the problem of tokenizer finding the right files when model is local
      
      * polish, remove hardcoded config
      
      * add a note that the file is autogenerated to avoid losing changes
      
      * prep for org change, remove unneeded code
      
      * switch to model4.pt, update scores
      
      * s/python/bash/
      
      * missing init (but doesn't impact the finetuned model)
      
      * cleanup
      
      * major refactor (reuse-bart)
      
      * new model, new expected weights
      
      * cleanup
      
      * cleanup
      
      * full link
      
      * fix model type
      
      * merge porting notes
      
      * style
      
      * cleanup
      
      * have to create a DecoderConfig object to handle vocab_size properly
      
      * doc fix
      
      * add note (not a public class)
      
      * parametrize
      
      * - add bleu scores integration tests
      
      * skip test if sacrebleu is not installed
      
      * cache heavy models/tokenizers
      
      * some tweaks
      
      * remove tokens that aren't used
      
      * more purging
      
      * simplify code
      
      * switch to using decoder_start_token_id
      
      * add doc
      
      * Revert "major refactor (reuse-bart)"
      
      This reverts commit 226dad15ca6a9ef4e26178526e878e8fc5c85874.
      
      * decouple from bart
      
      * remove unused code #1
      
      * remove unused code #2
      
      * remove unused code #3
      
      * update instructions
      
      * clean up
      
      * move bleu eval to examples
      
      * check import only once
      
      * move data+gen script into files
      
      * reuse via import
      
      * take less space
      
      * add prepare_seq2seq_batch (auto-tested)
      
      * cleanup
      
      * recode test to use json instead of yaml
      
      * ignore keys not needed
      
      * use the new -y in transformers-cli upload -y
      
      * [xlm tok] config dict: fix str into int to match definition (#7034)
      
      * [s2s] --eval_max_generate_length (#7018)
      
      * Fix CI with change of name of nlp (#7054)
      
      * nlp -> datasets
      
      * More nlp -> datasets
      
      * Woopsie
      
      * More nlp -> datasets
      
      * One last
      
      * extending to support allen_nlp wmt models
      
      - allow a specific checkpoint file to be passed
      - more arg settings
      - scripts for allen_nlp models
      
      * sync with changes
      
      * s/fsmt-wmt/wmt/ in model names
      
      * s/fsmt-wmt/wmt/ in model names (p2)
      
      * s/fsmt-wmt/wmt/ in model names (p3)
      
      * switch to a better checkpoint
      
      * typo
      
      * make non-optional args such - adjust tests where possible or skip when there is no other choice
      
      * consistency
      
      * style
      
      * adjust header
      
      * cards moved (model rename)
      
      * use best custom hparams
      
      * update info
      
      * remove old cards
      
      * cleanup
      
      * s/stas/facebook/
      
      * update scores
      
      * s/allen_nlp/allenai/
      
      * url maps aren't needed
      
      * typo
      
      * move all the doc / build /eval generators to their own scripts
      
      * cleanup
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * fix indent
      
      * duplicated line
      
      * style
      
      * use the correct add_start_docstrings
      
      * oops
      
      * resizing can't be done with the core approach, due to 2 dicts
      
      * check that the arg is a list
      
      * style
      
      * style
      Co-authored-by: default avatarSam Shleifer <sshleifer@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      1eeb206b
    • Sylvain Gugger's avatar
      Trainer multi label (#7191) · 492bb6aa
      Sylvain Gugger authored
      * Trainer accep multiple labels
      
      * Missing import
      
      * Fix dosctrings
      492bb6aa
  5. 16 Sep, 2020 1 commit
  6. 15 Sep, 2020 4 commits
  7. 14 Sep, 2020 3 commits
  8. 11 Sep, 2020 3 commits
  9. 10 Sep, 2020 5 commits
  10. 09 Sep, 2020 3 commits
  11. 08 Sep, 2020 1 commit
  12. 07 Sep, 2020 1 commit
  13. 04 Sep, 2020 1 commit
  14. 03 Sep, 2020 1 commit
  15. 02 Sep, 2020 3 commits
  16. 01 Sep, 2020 1 commit
    • Patrick von Platen's avatar
      [Generate] Facilitate PyTorch generate using `ModelOutputs` (#6735) · afc4ece4
      Patrick von Platen authored
      * fix generate for GPT2 Double Head
      
      * fix gpt2 double head model
      
      * fix  bart / t5
      
      * also add for no beam search
      
      * fix no beam search
      
      * fix encoder decoder
      
      * simplify t5
      
      * simplify t5
      
      * fix t5 tests
      
      * fix BART
      
      * fix transfo-xl
      
      * fix conflict
      
      * integrating sylvains and sams comments
      
      * fix tf past_decoder_key_values
      
      * fix enc dec test
      afc4ece4
  17. 31 Aug, 2020 2 commits
  18. 30 Aug, 2020 1 commit
  19. 29 Aug, 2020 1 commit
  20. 28 Aug, 2020 1 commit