1. 25 Feb, 2020 3 commits
  2. 24 Feb, 2020 13 commits
  3. 23 Feb, 2020 3 commits
  4. 22 Feb, 2020 6 commits
  5. 21 Feb, 2020 7 commits
    • Lysandre Debut's avatar
      Remove double bias (#2958) · 94ff2d6e
      Lysandre Debut authored
      94ff2d6e
    • Sam Shleifer's avatar
      Only use F.gelu for torch >=1.4.0 (#2955) · b5b3445c
      Sam Shleifer authored
      * Only use F.gelu for torch >=1.4.0
      
      * Use F.gelu for newer torch
      b5b3445c
    • Patrick von Platen's avatar
      Improve special_token_id logic in run_generation.py and add tests (#2885) · fc38d4c8
      Patrick von Platen authored
      
      
      * improving generation
      
      * finalized special token behaviour for no_beam_search generation
      
      * solved modeling_utils merge conflict
      
      * solve merge conflicts in modeling_utils.py
      
      * add run_generation improvements from PR #2749
      
      * adapted language generation to not use hardcoded -1 if no padding token is available
      
      * remove the -1 removal as hard coded -1`s are not necessary anymore
      
      * add lightweight language generation testing for randomely initialized models - just checking whether no errors are thrown
      
      * add slow language generation tests for pretrained models using hardcoded output with pytorch seed
      
      * delete ipdb
      
      * check that all generated tokens are valid
      
      * renaming
      
      * renaming Generation -> Generate
      
      * make style
      
      * updated so that generate_beam_search has same token behavior than generate_no_beam_search
      
      * consistent return format for run_generation.py
      
      * deleted pretrain lm generate tests -> will be added in another PR
      
      * cleaning of unused if statements and renaming
      
      * run_generate will always return an iterable
      
      * make style
      
      * consistent renaming
      
      * improve naming, make sure generate function always returns the same tensor, add docstring
      
      * add slow tests for all lmhead models
      
      * make style and improve example comments modeling_utils
      
      * better naming and refactoring in modeling_utils
      
      * improving generation
      
      * finalized special token behaviour for no_beam_search generation
      
      * solved modeling_utils merge conflict
      
      * solve merge conflicts in modeling_utils.py
      
      * add run_generation improvements from PR #2749
      
      * adapted language generation to not use hardcoded -1 if no padding token is available
      
      * remove the -1 removal as hard coded -1`s are not necessary anymore
      
      * add lightweight language generation testing for randomely initialized models - just checking whether no errors are thrown
      
      * add slow language generation tests for pretrained models using hardcoded output with pytorch seed
      
      * delete ipdb
      
      * check that all generated tokens are valid
      
      * renaming
      
      * renaming Generation -> Generate
      
      * make style
      
      * updated so that generate_beam_search has same token behavior than generate_no_beam_search
      
      * consistent return format for run_generation.py
      
      * deleted pretrain lm generate tests -> will be added in another PR
      
      * cleaning of unused if statements and renaming
      
      * run_generate will always return an iterable
      
      * make style
      
      * consistent renaming
      
      * improve naming, make sure generate function always returns the same tensor, add docstring
      
      * add slow tests for all lmhead models
      
      * make style and improve example comments modeling_utils
      
      * better naming and refactoring in modeling_utils
      
      * changed fast random lm generation testing design to more general one
      
      * delete in old testing design in gpt2
      
      * correct old variable name
      
      * temporary fix for encoder_decoder lm generation tests - has to be updated when t5 is fixed
      
      * adapted all fast random generate tests to new design
      
      * better warning description in modeling_utils
      
      * better comment
      
      * better comment and error message
      Co-authored-by: default avatarThomas Wolf <thomwolf@users.noreply.github.com>
      fc38d4c8
    • maximeilluin's avatar
      Added CamembertForQuestionAnswering (#2746) · c749a543
      maximeilluin authored
      * Added CamembertForQuestionAnswering
      
      * fixed camembert tokenizer case
      c749a543
    • Bram Vanroy's avatar
      Update modeling_tf_utils.py (#2924) · 5211d333
      Bram Vanroy authored
      Tensorflow does not use .eval() vs .train().
      
      closes https://github.com/huggingface/transformers/issues/2906
      5211d333
    • ahotrod's avatar
      Create README.md for xlnet_large_squad (#2942) · 3e98f27e
      ahotrod authored
      3e98f27e
    • Martin Malmsten's avatar
  6. 20 Feb, 2020 8 commits