1. 20 Feb, 2020 1 commit
    • Sam Shleifer's avatar
      New BartModel (#2745) · 53ce3854
      Sam Shleifer authored
      * Results same as fairseq
      * Wrote a ton of tests
      * Struggled with api signatures
      * added some docs
      
      53ce3854
  2. 19 Feb, 2020 3 commits
  3. 18 Feb, 2020 1 commit
  4. 13 Feb, 2020 2 commits
    • Joe Davison's avatar
      Preserve spaces in GPT-2 tokenizers (#2778) · f1e8a51f
      Joe Davison authored
      * Preserve spaces in GPT-2 tokenizers
      
      Preserves spaces after special tokens in GPT-2 and inhereted (RoBERTa)
      tokenizers, enabling correct BPE encoding. Automatically inserts a space
      in front of first token in encode function when adding special tokens.
      
      * Add tokenization preprocessing method
      
      * Add framework argument to pipeline factory
      
      Also fixes pipeline test issue. Each test input now treated as a
      distinct sequence.
      f1e8a51f
    • Sam Shleifer's avatar
      get_activation('relu') provides a simple mapping from strings i… (#2807) · ef74b0f0
      Sam Shleifer authored
      * activations.py contains a mapping from string to activation function
      * resolves some `gelu` vs `gelu_new` ambiguity
      ef74b0f0
  5. 11 Feb, 2020 1 commit
    • Oleksiy Syvokon's avatar
      BERT decoder: Fix causal mask dtype. · ee5de0ba
      Oleksiy Syvokon authored
      PyTorch < 1.3 requires multiplication operands to be of the same type.
      This was violated when using default attention mask (i.e.,
      attention_mask=None in arguments) given BERT in the decoder mode.
      
      In particular, this was breaking Model2Model and made tutorial
      from the quickstart failing.
      ee5de0ba
  6. 07 Feb, 2020 2 commits
  7. 04 Feb, 2020 8 commits
  8. 31 Jan, 2020 1 commit
  9. 30 Jan, 2020 2 commits
    • Julien Chaumond's avatar
      fill_mask helper (#2576) · 9fa836a7
      Julien Chaumond authored
      * fill_mask helper
      
      * [poc] FillMaskPipeline
      
      * Revert "[poc] FillMaskPipeline"
      
      This reverts commit 67eeea55b0f97b46c2b828de0f4ee97d87338335.
      
      * Revert "fill_mask helper"
      
      This reverts commit cacc17b884e14bb6b07989110ffe884ad9e36eaa.
      
      * README: clarify that Pipelines can also do text-classification
      
      cf. question at the AI&ML meetup last week, @mfuntowicz
      
      * Fix test: test feature-extraction pipeline
      
      * Test tweaks
      
      * Slight refactor of existing pipeline (in preparation of new FillMaskPipeline)
      
      * Extraneous doc
      
      * More robust way of doing this
      
      @mfuntowicz as we don't rely on the model name anymore (see AutoConfig)
      
      * Also add RobertaConfig as a quickfix for wrong token_type_ids
      
      * cs
      
      * [BIG] FillMaskPipeline
      9fa836a7
    • Lysandre's avatar
      Rename test_examples to test_doc_samples · df27648b
      Lysandre authored
      df27648b
  10. 29 Jan, 2020 2 commits
  11. 28 Jan, 2020 1 commit
  12. 27 Jan, 2020 2 commits
  13. 23 Jan, 2020 6 commits
  14. 17 Jan, 2020 1 commit
  15. 16 Jan, 2020 3 commits
  16. 15 Jan, 2020 3 commits
  17. 14 Jan, 2020 1 commit
    • Lysandre's avatar
      Bias should be resized with the weights · 100e3b6f
      Lysandre authored
      Created a link between the linear layer bias and the model attribute bias. This does not change anything for the user nor for the conversion scripts, but allows the `resize_token_embeddings` method to resize the bias as well as the weights of the decoder.
      
      Added a test.
      100e3b6f