1. 18 May, 2021 4 commits
    • Patrick von Platen's avatar
      Add more subsections to main doc (#11758) · cebb96f5
      Patrick von Platen authored
      * add headers to main doc
      
      * Apply suggestions from code review
      
      * update
      
      * upload
      cebb96f5
    • Tommy Chiang's avatar
      Fix incorrect newline in #11650 (#11757) · da7e73b7
      Tommy Chiang authored
      da7e73b7
    • Sylvain Gugger's avatar
      Fix checkpoint deletion (#11748) · a515caa3
      Sylvain Gugger authored
      a515caa3
    • Nicolas Patry's avatar
      [TokenClassification] Label realignment for subword aggregation (#11680) · b88e0e01
      Nicolas Patry authored
      * [TokenClassification] Label realignment for subword aggregation
      
      Tentative to replace https://github.com/huggingface/transformers/pull/11622/files
      
      
      
      - Added `AggregationStrategy`
      - `ignore_subwords` and `grouped_entities` arguments are now fused
        into `aggregation_strategy`. It makes more sense anyway because
        `ignore_subwords=True` with `grouped_entities=False` did not have a
        meaning anyway.
      - Added 2 new ways to aggregate which are MAX, and AVERAGE
      - AVERAGE requires a bit more information than the others, for now this
      case is slightly specific, we should keep that in mind for future
      changes.
      - Testing has been modified to reflect new argument, and to check the
      correct deprecation and the new aggregation_strategy.
      - Put the testing argument and testing results for aggregation_strategy,
      close together, so that readers can understand what is supposed to
      happen.
      - `aggregate` is now only tested on a small model as it does not mean
      anything to test it globally for all models.
      - Previous tests are unchanged in desired output.
      - Added a new test case that showcases better the difference between the
        FIRST, MAX and AVERAGE strategies.
      
      * Wrong framework.
      
      * Addressing three issues.
      
      1- Tags might not follow B-, I- convention, so any tag should work now
      (assumed as B-TAG)
      2- Fixed an issue with average that leads to a substantial code change.
      3- The testing suite was not checking for the "index" key for "none"
      strategy. This is now fixed.
      
      The issue is that "O" could not be chosen by AVERAGE strategy because
      those tokens were filtered out beforehand, so their relative scores were
      not counted in the average. Now filtering on
      ignore_labels will happen at the very end of the pipeline fixing
      that issue.
      It's a bit hard to make sure this stays like that because we do
      not have a end-to-end test for that behavior
      
      * Formatting.
      
      * Adding formatting to code + cleaner handling of B-, I- tags.
      Co-authored-by: default avatarFrancesco Rubbo <rubbo.francesco@gmail.com>
      Co-authored-by: default avatarelk-cloner <rezakakhki.rk@gmail.com>
      
      * Typo.
      Co-authored-by: default avatarFrancesco Rubbo <rubbo.francesco@gmail.com>
      Co-authored-by: default avatarelk-cloner <rezakakhki.rk@gmail.com>
      b88e0e01
  2. 17 May, 2021 7 commits
  3. 14 May, 2021 4 commits
  4. 13 May, 2021 8 commits
    • Volodymyr Byno's avatar
    • Sylvain Gugger's avatar
      Fix v4.6.0 doc · 25208200
      Sylvain Gugger authored
      25208200
    • Sylvain Gugger's avatar
      Fix doc deployment · cbbf49f6
      Sylvain Gugger authored
      cbbf49f6
    • lexhuismans's avatar
      [T5] Add 3D attention mask to T5 model (2) (#9643) (#11197) · 91cf2915
      lexhuismans authored
      * Add 3D attention mask to T5 model (#9643)
      
      Added code for 3D attention mask in T5 model. Similar to BERT model.
      
      * Add test for 3D attention mask
      
      Added test for 3D attention mask: test_decoder_model_past_with_3d_attn_mask()
      3D attention mask of the shape [Batch_size, Seq_length, Seq_length] both for
      attention mask and decoder attention mask. Test is passing.
      91cf2915
    • Vasudev Gupta's avatar
      add everything (#11651) · 6ee1a4fd
      Vasudev Gupta authored
      6ee1a4fd
    • Patrick von Platen's avatar
      [Flax] Fix BERT initialization & token_type_ids default (#11695) · 57b6a80d
      Patrick von Platen authored
      
      
      * fix some stuff
      
      * fix roberta & electra as well
      
      * del run bug
      Co-authored-by: default avatarPatrick von Platen <patrick@huggingface.co>
      57b6a80d
    • Lysandre Debut's avatar
      Fix gpt-2 warnings (#11709) · daf0d6a9
      Lysandre Debut authored
      daf0d6a9
    • Philip May's avatar
      Enable option for subword regularization in more tokenizers. (#11417) · 37ed3ab7
      Philip May authored
      * improve slow class tok usage at xlm rob
      
      * add subword regularization for barthez
      
      * improve barthez tok. test
      
      * fix tokenizer tests
      
      * add subword regularization for camembert
      
      * add subword regularization for deberta v2 tokenizer
      
      * add more doc to deberta v2 tokenizer
      
      * add subword regularization for speech to text tok.
      
      * fix sp_model_kwargs type in speech 2 text tok.
      
      * add subword regularization for M2M100 tok.
      
      * add more concrete type hints
      
      * fix tests for m2m100 and s2t tok.
      
      * add missing Any import
      
      * fix syntax error in m2m100 tok.
      
      * fix unpickle of m2m100 and s2t tok.
      
      * fix test of m2m100 and s2t tok.
      
      * improve unpickle of deberta v2 tok.
      
      * add test for pickle of barthez & camembert
      
      * fix pickle of barthez & camembert
      
      * add test for deberta v2 tok. pickle
      
      * fix m2m100 tok. pickle
      
      * fix s2t tok. pickle
      
      * add subword regularization to albert tok.
      
      * refactor subword reg. test into TokenizerTesterMixin
      
      improve albert tok. test
      
      remove sample argument form albert tok.
      
      check subword reg. using TokenizerTesterMixin
      
      improve tok. tests
      
      improve xlm roberta tok. tests
      
      improve xlm roberta tok. tests
      
      * add subword regularization for big bird t.
      
      * improve xlm roberta tok. test
      
      * add subword regularization for mbart50 tok.
      
      * add subword regularization for pegasus tok.
      
      * add subword regularization for reformer tok.
      
      * add subword regularization for T5 tok.
      
      * fix t5 tok. test formatting
      
      * add subword regularization for xlm_proph. tok.
      
      * add subword regularization for xlnet tok.
      
      * add subword regularization for gert_gen tok.
      
      * add typing to tokenizers
      
      * add typing to xlm rob. tok
      
      * add subword regularization for marian tok.
      
      * add reverse tok. test
      
      * fix marian tok test
      
      * fix marian tok test
      
      * fix casing in tok. tests
      
      * fix style of tok. common test
      
      * fix deberta v2 tok test
      
      * add type annotations to tok. tests
      
      * add type annotations to tok. __init__
      
      * add typing to kokenizer
      
      * add type annotations to tok. __init__
      
      * don't specify the default when it's None
      
      * fix barthez tok. doc
      
      * move sentencepiece tok. tests to TokenizerTesterMixin
      
      * fix unused imports
      
      * fix albert tok. test
      
      * add comment to sentencepiece test options
      
      * fix Any import at big bird tok.
      
      * fix Any import at xlm prophetnet tok.
      
      * empty commit to trigger CI
      37ed3ab7
  5. 12 May, 2021 9 commits
  6. 11 May, 2021 8 commits