"web/vscode:/vscode.git/clone" did not exist on "4c54c2ec0f642ce505980373aae71ed09cb1702c"
  1. 07 Dec, 2020 4 commits
  2. 06 Dec, 2020 1 commit
  3. 05 Dec, 2020 2 commits
  4. 04 Dec, 2020 5 commits
  5. 03 Dec, 2020 7 commits
  6. 02 Dec, 2020 7 commits
    • Patrick von Platen's avatar
      [PyTorch] Refactor Resize Token Embeddings (#8880) · 443f67e8
      Patrick von Platen authored
      * fix resize tokens
      
      * correct mobile_bert
      
      * move embedding fix into modeling_utils.py
      
      * refactor
      
      * fix lm head resize
      
      * refactor
      
      * break lines to make sylvain happy
      
      * add news tests
      
      * fix typo
      
      * improve test
      
      * skip bart-like for now
      
      * check if base_model = get(...) is necessary
      
      * clean files
      
      * improve test
      
      * fix tests
      
      * revert style templates
      
      * Update templates/adding_a_new_model/cookiecutter-template-{{cookiecutter.modelname}}/modeling_{{cookiecutter.lowercase_modelname}}.py
      443f67e8
    • Devangi Purkayastha's avatar
      Update README.md (#8906) · e52f9c0a
      Devangi Purkayastha authored
      e52f9c0a
    • ryota-mo's avatar
      Fix typo in docstring (#8905) · 801b2cb3
      ryota-mo authored
      801b2cb3
    • Stas Bekman's avatar
      [trainer] improve code readability (#8903) · 7e1cb00c
      Stas Bekman authored
      * [trainer] improve code
      
      This PR:
      - removes redundant code 
      ```
      self.model = model if model is not None else None
      ```
      and
      ```
      self.model = model
      ```
      are the same.
      
      * separate attribute assignment from code logic - which simplifies things further.
      
      * whitespace
      7e1cb00c
    • Nicolas Patry's avatar
      Warning about too long input for fast tokenizers too (#8799) · a8c3f9aa
      Nicolas Patry authored
      * Warning about too long input for fast tokenizers too
      
      If truncation is not set in tokenizers, but the tokenization is too long
      for the model (`model_max_length`), we used to trigger a warning that
      
      The input would probably fail (which it most likely will).
      
      This PR re-enables the warning for fast tokenizers too and uses common
      code for the trigger to make sure it's consistent across.
      
      * Checking for pair of inputs too.
      
      * Making the function private and adding it's doc.
      
      * Remove formatting ?? in odd place.
      
      * Missed uppercase.
      a8c3f9aa
    • sandip's avatar
      Transfoxl seq classification (#8868) · f6b44e61
      sandip authored
      * Transfoxl sequence classification
      
      * Transfoxl sequence classification
      f6b44e61
    • Stas Bekman's avatar
      [ci] skip doc jobs take #3 (#8885) · 24f0c2fe
      Stas Bekman authored
      * check that we get any match first
      
      * docs only
      
      * 2 docs only
      
      * add code
      
      * restore
      24f0c2fe
  7. 01 Dec, 2020 11 commits
  8. 30 Nov, 2020 3 commits
    • Stas Bekman's avatar
      [s2s trainer] fix DP mode (#8823) · 7f34d757
      Stas Bekman authored
      * fix DP case on multi-gpu
      
      * make executable
      
      * test all 3 modes
      
      * use the correct check for distributed
      
      * dp doesn't need a special case
      
      * restore original name
      
      * cleanup
      7f34d757
    • Nicolas Patry's avatar
      NerPipeline (TokenClassification) now outputs offsets of words (#8781) · d8fc26e9
      Nicolas Patry authored
      * NerPipeline (TokenClassification) now outputs offsets of words
      
      - It happens that the offsets are missing, it forces the user to pattern
      match the "word" from his input, which is not always feasible.
      For instance if a sentence contains the same word twice, then there
      is no way to know which is which.
      - This PR proposes to fix that by outputting 2 new keys for this
      pipelines outputs, "start" and "end", which correspond to the string
      offsets of the word. That means that we should always have the
      invariant:
      
      ```python
      input[entity["start"]: entity["end"]] == entity["entity_group"]
                                          # or entity["entity"] if not grouped
      ```
      
      * Fixing doc style
      d8fc26e9
    • LysandreJik's avatar
      fix pypi complaint on version naming · 5fd3d81e
      LysandreJik authored
      5fd3d81e