1. 06 Sep, 2021 1 commit
    • Nils Reimers's avatar
      Update model configs - Allow setters for common properties (#13026) · c8be8a9a
      Nils Reimers authored
      * refactor GPT Config to allow dyn. properties
      
      * make attribute_map a class attribute
      
      * remove old code
      
      * update unit test to test config: Add test for common properties setter
      
      * update unit test to test config: Add test for common properties passed as parameters to __init__
      
      * update to black code format
      
      * Allow that setters are not defined for certain config classes
      
      * update config classes to implement attribute_map
      
      * bugfix lxmert config - id2labels was not defined when num_labels was set
      
      * update broken configs - add attribute_maps
      
      * update bart config
      
      * update black codestyle
      
      * update documentation on common config attributes
      
      * update GPTJ config to new attribute map
      
      * update docs on common attributes
      
      * gptj config: add max_position_embeddings
      
      * gptj config: format with black
      
      * update speech to text 2 config
      
      * format doc file to max_len 119
      
      * update config template
      c8be8a9a
  2. 10 Feb, 2021 1 commit
    • Suraj Patil's avatar
      remove adjust_logits_during_generation method (#10087) · c130e67d
      Suraj Patil authored
      * add forced logits processors
      
      * delete adjust_logits method
      
      * add forced_eos_token_id argument in config
      
      * add tests for forced logits processors
      
      * update gen utils tests
      
      * add forced option to tf generate
      
      * remove adjust_logits method from tf models
      
      * update adjust_logits for marian
      
      * delete _force_token_id_to_be_generated method
      
      * style
      
      * import warnings
      
      * pass max_length to _get_logits_processor
      
      * set forced_eos_token_id to None
      
      * set forced attributes in conf utils
      
      * typo
      
      * fix rag generate
      
      * add forced_eos_token_id in rag config
      
      * remove force_bos_token_to_be_generated from BartConfig
      
      * remove _force_token_ids_generation from FSMT
      
      * nit
      
      * fix negative constant
      
      * apply suggestions from code review
      c130e67d
  3. 25 Nov, 2020 1 commit
    • Patrick von Platen's avatar
      [XLNet] Fix mems behavior (#8567) · 2a6fbe6a
      Patrick von Platen authored
      * fix mems in xlnet
      
      * fix use_mems
      
      * fix use_mem_len
      
      * fix use mems
      
      * clean docs
      
      * fix tf typo
      
      * make xlnet tf for generation work
      
      * fix tf test
      
      * refactor use cache
      
      * add use cache for missing models
      
      * correct use_cache in generate
      
      * correct use cache in tf generate
      
      * fix tf
      
      * correct getattr typo
      
      * make sylvain happy
      
      * change in docs as well
      
      * do not apply to cookie cutter statements
      
      * fix tf test
      
      * make pytorch model fully backward compatible
      2a6fbe6a
  4. 17 Nov, 2020 1 commit
    • Sylvain Gugger's avatar
      Reorganize repo (#8580) · c89bdfbe
      Sylvain Gugger authored
      * Put models in subfolders
      
      * Styling
      
      * Fix imports in tests
      
      * More fixes in test imports
      
      * Sneaky hidden imports
      
      * Fix imports in doc files
      
      * More sneaky imports
      
      * Finish fixing tests
      
      * Fix examples
      
      * Fix path for copies
      
      * More fixes for examples
      
      * Fix dummy files
      
      * More fixes for example
      
      * More model import fixes
      
      * Is this why you're unhappy GitHub?
      
      * Fix imports in conver command
      c89bdfbe
  5. 30 Oct, 2020 1 commit
    • TFUsers's avatar
      Replace swish with silu (#8166) · 00112c35
      TFUsers authored
      
      
      * Replace swish with silu
      
      * revert nn.silu to nn.swish due to older version
      
      * simplify optimized silu conditional and fix format
      
      * Update activations.py
      
      * Update activations_tf.py
      
      * Update modeling_flax_utils.py
      
      * Update modeling_openai.py
      
      * add swish testcase
      
      * add pytorch swish testcase
      
      * Add more robust python version check
      
      * more formatting fixes
      Co-authored-by: default avatarTFUsers <TFUsers@gmail.com>
      00112c35
  6. 29 Oct, 2020 1 commit
  7. 26 Oct, 2020 1 commit
    • Sylvain Gugger's avatar
      Doc styling (#8067) · 08f534d2
      Sylvain Gugger authored
      * Important files
      
      * Styling them all
      
      * Revert "Styling them all"
      
      This reverts commit 7d029395fdae8513b8281cbc2a6c239f8093503e.
      
      * Syling them for realsies
      
      * Fix syntax error
      
      * Fix benchmark_utils
      
      * More fixes
      
      * Fix modeling auto and script
      
      * Remove new line
      
      * Fixes
      
      * More fixes
      
      * Fix more files
      
      * Style
      
      * Add FSMT
      
      * More fixes
      
      * More fixes
      
      * More fixes
      
      * More fixes
      
      * Fixes
      
      * More fixes
      
      * More fixes
      
      * Last fixes
      
      * Make sphinx happy
      08f534d2
  8. 22 Oct, 2020 1 commit
  9. 23 Sep, 2020 1 commit
  10. 17 Sep, 2020 1 commit
    • Stas Bekman's avatar
      [ported model] FSMT (FairSeq MachineTranslation) (#6940) · 1eeb206b
      Stas Bekman authored
      * ready for PR
      
      * cleanup
      
      * correct FSMT_PRETRAINED_MODEL_ARCHIVE_LIST
      
      * fix
      
      * perfectionism
      
      * revert change from another PR
      
      * odd, already committed this one
      
      * non-interactive upload workaround
      
      * backup the failed experiment
      
      * store langs in config
      
      * workaround for localizing model path
      
      * doc clean up as in https://github.com/huggingface/transformers/pull/6956
      
      
      
      * style
      
      * back out debug mode
      
      * document: run_eval.py --num_beams 10
      
      * remove unneeded constant
      
      * typo
      
      * re-use bart's Attention
      
      * re-use EncoderLayer, DecoderLayer from bart
      
      * refactor
      
      * send to cuda and fp16
      
      * cleanup
      
      * revert (moved to another PR)
      
      * better error message
      
      * document run_eval --num_beams
      
      * solve the problem of tokenizer finding the right files when model is local
      
      * polish, remove hardcoded config
      
      * add a note that the file is autogenerated to avoid losing changes
      
      * prep for org change, remove unneeded code
      
      * switch to model4.pt, update scores
      
      * s/python/bash/
      
      * missing init (but doesn't impact the finetuned model)
      
      * cleanup
      
      * major refactor (reuse-bart)
      
      * new model, new expected weights
      
      * cleanup
      
      * cleanup
      
      * full link
      
      * fix model type
      
      * merge porting notes
      
      * style
      
      * cleanup
      
      * have to create a DecoderConfig object to handle vocab_size properly
      
      * doc fix
      
      * add note (not a public class)
      
      * parametrize
      
      * - add bleu scores integration tests
      
      * skip test if sacrebleu is not installed
      
      * cache heavy models/tokenizers
      
      * some tweaks
      
      * remove tokens that aren't used
      
      * more purging
      
      * simplify code
      
      * switch to using decoder_start_token_id
      
      * add doc
      
      * Revert "major refactor (reuse-bart)"
      
      This reverts commit 226dad15ca6a9ef4e26178526e878e8fc5c85874.
      
      * decouple from bart
      
      * remove unused code #1
      
      * remove unused code #2
      
      * remove unused code #3
      
      * update instructions
      
      * clean up
      
      * move bleu eval to examples
      
      * check import only once
      
      * move data+gen script into files
      
      * reuse via import
      
      * take less space
      
      * add prepare_seq2seq_batch (auto-tested)
      
      * cleanup
      
      * recode test to use json instead of yaml
      
      * ignore keys not needed
      
      * use the new -y in transformers-cli upload -y
      
      * [xlm tok] config dict: fix str into int to match definition (#7034)
      
      * [s2s] --eval_max_generate_length (#7018)
      
      * Fix CI with change of name of nlp (#7054)
      
      * nlp -> datasets
      
      * More nlp -> datasets
      
      * Woopsie
      
      * More nlp -> datasets
      
      * One last
      
      * extending to support allen_nlp wmt models
      
      - allow a specific checkpoint file to be passed
      - more arg settings
      - scripts for allen_nlp models
      
      * sync with changes
      
      * s/fsmt-wmt/wmt/ in model names
      
      * s/fsmt-wmt/wmt/ in model names (p2)
      
      * s/fsmt-wmt/wmt/ in model names (p3)
      
      * switch to a better checkpoint
      
      * typo
      
      * make non-optional args such - adjust tests where possible or skip when there is no other choice
      
      * consistency
      
      * style
      
      * adjust header
      
      * cards moved (model rename)
      
      * use best custom hparams
      
      * update info
      
      * remove old cards
      
      * cleanup
      
      * s/stas/facebook/
      
      * update scores
      
      * s/allen_nlp/allenai/
      
      * url maps aren't needed
      
      * typo
      
      * move all the doc / build /eval generators to their own scripts
      
      * cleanup
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * fix indent
      
      * duplicated line
      
      * style
      
      * use the correct add_start_docstrings
      
      * oops
      
      * resizing can't be done with the core approach, due to 2 dicts
      
      * check that the arg is a list
      
      * style
      
      * style
      Co-authored-by: default avatarSam Shleifer <sshleifer@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      1eeb206b