1. 06 Jan, 2021 2 commits
  2. 05 Jan, 2021 13 commits
    • Patrick von Platen's avatar
      [PyTorch Bart] Split Bart into different models (#9343) · eef66035
      Patrick von Platen authored
      * first try
      
      * remove old template
      
      * finish bart
      
      * finish mbart
      
      * delete unnecessary line
      
      * init pegasus
      
      * save intermediate
      
      * correct pegasus
      
      * finish pegasus
      
      * remove cookie cutter leftover
      
      * add marian
      
      * finish blenderbot
      
      * replace in file
      
      * correctly split blenderbot
      
      * delete "old" folder
      
      * correct "add statement"
      
      * adapt config for tf comp
      
      * correct configs for tf
      
      * remove ipdb
      
      * fix more stuff
      
      * fix mbart
      
      * push pegasus fix
      
      * fix mbart
      
      * more fixes
      
      * fix research projects code
      
      * finish docs for bart, mbart, and marian
      
      * delete unnecessary file
      
      * correct attn typo
      
      * correct configs
      
      * remove pegasus for seq class
      
      * correct peg docs
      
      * correct peg docs
      
      * finish configs
      
      * further improve docs
      
      * add copied from statements to mbart
      
      * fix copied from in mbart
      
      * add copy statements to marian
      
      * add copied from to marian
      
      * add pegasus copied from
      
      * finish pegasus
      
      * finish copied from
      
      * Apply suggestions from code review
      
      * make style
      
      * backward comp blenderbot
      
      * apply lysandres and sylvains suggestions
      
      * apply suggestions
      
      * push last fixes
      
      * fix docs
      
      * fix tok tests
      
      * fix imports code style
      
      * fix doc
      eef66035
    • Clement's avatar
      4eec5d0c
    • Stas Bekman's avatar
      add experimental warning (#9412) · d9e848c1
      Stas Bekman authored
      d9e848c1
    • Stas Bekman's avatar
      [trainer] group fp16 args together (#9409) · 29acabd8
      Stas Bekman authored
      * [t5 doc] typos
      
      a few run away backticks
      
      @sgugger
      
      * style
      
      * [trainer] put fp16 args together
      
      this PR proposes a purely cosmetic change that puts all the fp16 args together - so they are easier to manager/read
      
      @sgugger
      
      * style
      29acabd8
    • Yusuke Mori's avatar
    • Patrick von Platen's avatar
      LED (#9278) · 189387e9
      Patrick von Platen authored
      * create model
      
      * add integration
      
      * save current state
      
      * make integration tests pass
      
      * add one more test
      
      * add explanation to tests
      
      * remove from bart
      
      * add padding
      
      * remove unnecessary test
      
      * make all tests pass
      
      * re-add cookie cutter tests
      
      * finish PyTorch
      
      * fix attention test
      
      * Update tests/test_modeling_common.py
      
      * revert change
      
      * remove unused file
      
      * add string to doc
      
      * save intermediate
      
      * make tf integration tests pass
      
      * finish tf
      
      * fix doc
      
      * fix docs again
      
      * add led to doctree
      
      * add to auto tokenizer
      
      * added tips for led
      
      * make style
      
      * apply jplus statements
      
      * correct tf longformer
      
      * apply lysandres suggestions
      
      * apply sylvains suggestions
      
      * Apply suggestions from code review
      189387e9
    • Sugeeth's avatar
      Fix documentation links always pointing to master. (#9217) · 314cca28
      Sugeeth authored
      
      
      * Use extlinks to point hyperlink with the version of code
      
      * Point to version on release and master until then
      
      * Apply style
      
      * Correct links
      
      * Add missing backtick
      
      * Simple missing backtick after all.
      Co-authored-by: default avatarRaghavendra Sugeeth P S <raghav-5305@raghav-5305.csez.zohocorpin.com>
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      314cca28
    • Julien Plu's avatar
      Fix TF Funnel (#9300) · 52d62e68
      Julien Plu authored
      * Fix Funnel
      
      * Apply Patrick's comment
      
      * Remove comment
      
      * Fix dummy value
      
      * Apply style
      52d62e68
    • Stas Bekman's avatar
      [trainer] --model_parallel hasn't been implemented for most models (#9347) · 748006c0
      Stas Bekman authored
      * --model_parallel hasn't been implemented for most models
      
      * make the help clear as well
      
      * implement is_parallelizable; use it
      
      * oops
      
      * remove property
      748006c0
    • Julien Plu's avatar
      Use stable functions (#9369) · 4225740a
      Julien Plu authored
      4225740a
    • Stas Bekman's avatar
      [logging] autoflush (#9385) · 4aa8f6ad
      Stas Bekman authored
      This PR proposes to:
      
      * auto-flush `transformers` logging 
      
      When using logging for tracing signals from different parts of the code and which could be mixed with print debug this aids to get all the logging events synchronized. 
      
      I don't think this change will introduce any performance impacts.
      
      If it helps someone here is the code I used to sync `transformers` logging with various other debug prints.
      
      I was porting bart to MP and I needed to trace that the device switching happens correctly and I added a bunch of logger.info calls inside `modeling_bart.py` and also had some other helpers `print` debug messages which weren't logger based:
      
      ```
      
      # auto flush std streams
      from sys import stdout, stderr
      def stdout_write_flush(args, w=stderr.write): w(args); stderr.flush()
      def stderr_write_flush(args, w=stderr.write): w(args); stderr.flush()
      stdout.write = stdout_write_flush
      stderr.write = stderr_write_flush
      
      from transformers import BartTokenizer, BartForConditionalGeneration, BartConfig
      
      import logging
      import transformers.utils.logging
      import transformers.models.bart.modeling_bart
      
      # I wanted a shorter simpler format
      handlers = transformers.utils.logging._get_library_root_logger().handlers
      for handler in handlers:
          formatter = logging.Formatter("[%(funcName)s] %(message)s")
          handler.setFormatter(formatter)
      
      transformers.models.bart.modeling_bart.logger.setLevel(transformers.logging.INFO)
      ```
      
      @LysandreJik, @sgugger, @patrickvonplaten
      4aa8f6ad
    • Julien Plu's avatar
      Fix TF Longformer (#9348) · 83eec97e
      Julien Plu authored
      * Fix longformer
      
      * Apply style
      
      * Remove serving content
      
      * Forgot a condition
      
      * Apply style
      
      * Address Patrick's comments
      
      * Fix dtype
      83eec97e
    • Boris Dayma's avatar
      feat(wandb): save model as artifact (#8119) · 30fa0b78
      Boris Dayma authored
      * feat(wandb): log artifacts
      
      * fix: typo
      
      * feat(wandb): ensure name is allowed
      
      * feat(wandb): log artifact
      
      * feat(wandb): saving logic
      
      * style: improve formatting
      
      * fix: unrelated typo
      
      * feat:聽use a fake trainer
      
      * fix:聽simplify
      
      * feat(wandb): log model files as artifact
      
      * style: fix style
      
      * docs(wandb): correct description
      
      * feat: unpack model + allow env Truethy values
      
      * feat: TrainerCallback can access tokenizer
      
      * style:聽fix style
      
      * feat(wandb): log more interesting metadata
      
      * feat: unpack tokenizer
      
      * feat(wandb): metadata with load_best_model_at_end
      
      * feat(wandb): more robust metadata
      
      * style(wandb): fix formatting
      30fa0b78
  3. 04 Jan, 2021 13 commits
  4. 03 Jan, 2021 1 commit
  5. 02 Jan, 2021 3 commits
  6. 30 Dec, 2020 1 commit
  7. 29 Dec, 2020 3 commits
  8. 28 Dec, 2020 3 commits
  9. 27 Dec, 2020 1 commit