"git@developer.sourcefind.cn:wuxk1/megatron-lm.git" did not exist on "62632d39364e499bbfe9e143d2856dbe3f3a6802"
  1. 01 Mar, 2024 3 commits
  2. 29 Feb, 2024 3 commits
  3. 28 Feb, 2024 3 commits
  4. 27 Feb, 2024 1 commit
  5. 26 Feb, 2024 1 commit
  6. 23 Feb, 2024 1 commit
    • Matt's avatar
      Allow remote code repo names to contain "." (#29175) · 371b572e
      Matt authored
      * stash commit
      
      * stash commit
      
      * It works!
      
      * Remove unnecessary change
      
      * We don't actually need the cache_dir!
      
      * Update docstring
      
      * Add test
      
      * Add test with custom cache dir too
      
      * Update model repo path
      371b572e
  7. 22 Feb, 2024 1 commit
  8. 21 Feb, 2024 3 commits
  9. 20 Feb, 2024 3 commits
    • Joao Gante's avatar
    • amyeroberts's avatar
      Revert low cpu mem tie weights (#29135) · 0996a100
      amyeroberts authored
      * Revert "Add tie_weights() to LM heads and set bias in set_output_embeddings() (#28948)"
      
      This reverts commit 725f4ad1.
      
      * Revert "Patch to skip failing `test_save_load_low_cpu_mem_usage` tests (#29043)"
      
      This reverts commit 4156f517.
      0996a100
    • Arthur's avatar
      [`Core tokenization`] `add_dummy_prefix_space` option to help with latest issues (#28010) · 15cfe389
      Arthur authored
      * add add_dummy_prefix_space option to slow
      
      * checking kwargs might be better. Should be there for all spm tokenizer IMO
      
      * nits
      
      * fix copies
      
      * more copied
      
      * nits
      
      * add prefix space
      
      * nit
      
      * nits
      
      * Update src/transformers/convert_slow_tokenizer.py
      
      * fix inti
      
      * revert wrong styling
      
      * fix
      
      * nits
      
      * style
      
      * updates
      
      * make sure we use slow tokenizer for conversion instead of looking for the decoder
      
      * support llama ast well
      
      * update llama tokenizer fast
      
      * nits
      
      * nits nits nits
      
      * update the doc
      
      * update
      
      * update to fix tests
      
      * skip unrelated tailing test
      
      * Update src/transformers/convert_slow_tokenizer.py
      
      * add proper testing
      
      * test decode as well
      
      * more testing
      
      * format
      
      * fix llama test
      
      * Apply suggestions from code review
      15cfe389
  10. 19 Feb, 2024 1 commit
  11. 16 Feb, 2024 2 commits
  12. 15 Feb, 2024 2 commits
  13. 14 Feb, 2024 4 commits
    • JB (Don)'s avatar
      Add tie_weights() to LM heads and set bias in set_output_embeddings() (#28948) · 725f4ad1
      JB (Don) authored
      * Add tie_weights() to LM heads and set bias in set_output_embeddings()
      
      The bias were not tied correctly in some LM heads, and this change should fix that.
      
      * Moving test_save_and_load_low_cpu_mem_usage to ModelTesterMixin
      
      * Adding _tie_weights() to MPNet and Vilt
      
      * Skip test for low cpu mem usage for Deta/DeformableDetr since they cannot init on meta device
      
      * Rename to test name to save_load to match the convention
      725f4ad1
    • Raushan Turganbay's avatar
    • NielsRogge's avatar
      Add SiglipForImageClassification and CLIPForImageClassification (#28952) · 63ffd56d
      NielsRogge authored
      * First draft
      
      * Add CLIPForImageClassification
      
      * Remove scripts
      
      * Fix doctests
      63ffd56d
    • Jonathan Tow's avatar
      Add `StableLM` (#28810) · de6029a0
      Jonathan Tow authored
      * Add `StableLM`
      
      * fix(model): re-create from `huggingface-cli add-new-model-like persimmon`
      
      * fix: re-add changes to address comments
      
      * fix(readme): add links to paper
      
      * fix(tokenization_auto): remove `GPTNeoXTokenizerFastFast` ref
      
      * fix(tests): re-add `@slow` decorator to integration tests
      
      * fix(tests): import slow...
      
      * fix(readme_hd): remove whitespace edit
      
      * fix(tokenizer): auto tokenizer tuple
      
      * skip doctests for `modeling_stablelm`
      de6029a0
  14. 13 Feb, 2024 3 commits
  15. 08 Feb, 2024 1 commit
  16. 06 Feb, 2024 3 commits
  17. 05 Feb, 2024 1 commit
  18. 02 Feb, 2024 3 commits
  19. 01 Feb, 2024 1 commit
    • JB (Don)'s avatar
      Adding [T5/MT5/UMT5]ForTokenClassification (#28443) · 0d26abdd
      JB (Don) authored
      * Adding [T5/MT5/UMT5]ForTokenClassification
      
      * Add auto mappings for T5ForTokenClassification and variants
      
      * Adding ForTokenClassification to the list of models
      
      * Adding attention_mask param to the T5ForTokenClassification test
      
      * Remove outdated comment in test
      
      * Adding EncoderOnly and Token Classification tests for MT5 and UMT5
      
      * Fix typo in umt5 string
      
      * Add tests for all the existing MT5 models
      
      * Fix wrong comment in dependency_versions_table
      
      * Reverting change to common test for _keys_to_ignore_on_load_missing
      
      The test is correctly picking up redundant keys in _keys_to_ignore_on_load_missing.
      
      * Removing _keys_to_ignore_on_missing from MT5 since the key is not used in the model
      
      * Add fix-copies to MT5ModelTest
      0d26abdd