1. 26 Apr, 2021 6 commits
  2. 25 Apr, 2021 2 commits
    • cronoik's avatar
      EncoderDecoderConfigs should not create new objects (#11300) · 35cd8eed
      cronoik authored
      
      
      * removes the creation of separate config objects and uses the existing ones instead+overwrite resize_token_embeddings from parent class because it is not working for the EncoderDecoderModel
      
      * rollback to current version of the huggingface master branch
      
      * reworked version that ties the encoder and decoder config of the parent encoderdecoder instance
      
      * overwrite of resize_token_embeddings throws an error now
      
      * review comment suggestion
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      * implemented warning in case encoderdecoder is created with differing configs of encoderdecoderconfig and decoderconfig or encoderconfig
      
      * added test to avoid diverging configs of wrapper class and wrapped classes
      
      * Update src/transformers/models/encoder_decoder/modeling_encoder_decoder.py
      
      * make style
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      35cd8eed
    • Daniel Stancl's avatar
      Add head_mask, decoder_head_mask, cross_head_mask to ProphetNet (#9964) · f45cb66b
      Daniel Stancl authored
      * Add head_mask & decoder_head_mask + some corrections
      
      * Fix head masking for N-grams
      
      * Enable test_headmasking for encoder and decod
      
      * Fix one typo regarding in modeling_propgetnet.py
      
      * Enable test_headmasking for ProphetNetStandaloneDecoderModelTest
      and ProphetNetStandaloneEncoderModelTest in test_modeling_prophetnet.py
      
      * make style
      
      * Fix cross_head_mask
      
      * Fix attention head mask naming
      
      * `cross_head_mask` -> `cross_attn_head_mask`
      
      * `cross_layer_head_mask` -> `cross_attn_layer_head_mask`
      
      * Still need to merge #10605 to master to pass the tests
      f45cb66b
  3. 24 Apr, 2021 2 commits
  4. 23 Apr, 2021 10 commits
  5. 22 Apr, 2021 5 commits
  6. 21 Apr, 2021 7 commits
    • Sylvain Gugger's avatar
      Add huggingface_hub dep for #11328 · ff26f8ee
      Sylvain Gugger authored
      ff26f8ee
    • wlhgtc's avatar
      Fix token_type_ids error for big_bird model. (#11355) · 5e04d708
      wlhgtc authored
      
      
      * MOD: fit chinese wwm to new datasets
      
      * MOD: move wwm to new folder
      
      * MOD: formate code
      
      * Styling
      
      * MOD add param and recover trainer
      
      * MOD: add token_type_ids method for big bird
      
      * MOD: format code
      
      * MOD: format code
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      5e04d708
    • Patrick von Platen's avatar
      Remove boiler plate code (#11340) · 50595a33
      Patrick von Platen authored
      * remove boiler plate code
      
      * adapt roberta
      
      * correct docs
      
      * finish refactor
      50595a33
    • lewtun's avatar
      Extract metric_key_prefix during NotebookProgressCallback.on_evaluate (#11347) · 41f3133a
      lewtun authored
      * Pass metric_key_prefix as kwarg to on_evaluate
      
      * Replace eval_loss with metric_key_prefix_loss
      
      * Default to "eval" if metric_key_prefix not in kwargs
      
      * Add kwargs to CallbackHandler.on_evaluate signature
      
      * Revert "Add kwargs to CallbackHandler.on_evaluate signature"
      
      This reverts commit 8d4c85ed512f558f7579d36771e907b3379947b7.
      
      * Revert "Pass metric_key_prefix as kwarg to on_evaluate"
      
      This reverts commit 7766bfe2718601230ae593d37b1317bd53cfc075.
      
      * Extract metric_key_prefix from metrics
      41f3133a
    • Sylvain Gugger's avatar
      Examples reorg (#11350) · dabeb152
      Sylvain Gugger authored
      
      
      * Base move
      
      * Examples reorganization
      
      * Update references
      
      * Put back test data
      
      * Move conftest
      
      * More fixes
      
      * Move test data to test fixtures
      
      * Update path
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Address review comments and clean
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      dabeb152
    • Stas Bekman's avatar
      [deepspeed] fix resume from checkpoint (#11352) · ca7ff64f
      Stas Bekman authored
      This PR fixes a bug that most likely somehow got exposed (not caused) by https://github.com/huggingface/transformers/pull/11318 - surprisingly the same test worked just fine before that other PR.
      ca7ff64f
    • Nicolas Patry's avatar
      Removed `max_length` from being mandatory within `generate`. (#11314) · aad95c7c
      Nicolas Patry authored
      * Removed `max_length` from being mandatory within `generate`.
      
      - Moving on to fully using `StoppingCriteria` for `greedy` and `sample`
      modes.
      - `max_length` still used for `beam_search` and `group_beam_search`
      (Follow up PR)
      - Fixes a bug with MaxLengthStoppingCriteria (we should stop as soon a
      we hit the max_length, the comparison needs to be or equal, that affects
      the tests).
      - Added options to use `logits_processor` and `stopping_criteria`
      directly within `generate` function (so some users can define their own
      `logits_processor` and `stopping_criteria`).
      - Modified the backward compat tests to make sure we issue a warning.
      
      * Fix `max_length` argument in `generate`.
      
      * Moving validate to being functional.
      
      - Renamed `smax_length` to `stoppping_max_length`.
      
      * Removing `logits_processor` and `stopping_criteria` from `generate`
      arguments.
      
      * Deepcopy.
      
      * Fix global variable name.
      aad95c7c
  7. 20 Apr, 2021 5 commits
  8. 19 Apr, 2021 3 commits