1. 06 Aug, 2024 4 commits
    • Ao Tang's avatar
      Add Nemotron HF Support (#31699) · 6a03942d
      Ao Tang authored
      * Add nemotron support
      
      * fix inference
      
      * add unit test
      
      * add layernorm1p as a class to avoid meta device mismatch
      
      * test fixed
      
      * Add copied_from statements
      
      * remove pretraining_tp args
      
      * remove nemotronlayernorm
      
      * force LN computation done in FP32
      
      * remove nemotrontokenizer and use llamatokenizer
      
      * license update
      
      * add option for kv_channels for minitron8b
      
      * remove assert
      
      * o_proj fixed
      
      * o_proj reshape
      
      * add gated_proj option
      
      * typo
      
      * remove todos
      
      * fix broken test after merging latest main
      
      * remove nezha/nat after meging main
      
      * chnage default config to 15b model
      
      * add nemo conversion script
      
      * rename conversion script
      
      * remove gate_proj option
      
      * pr comment resolved
      
      * fix unit test
      
      * rename kv_channels to head_dim
      
      * resolve PR issue
      
      * add nemotron md
      
      * fix broken tests
      
      * refactor rope for nemotron
      
      * test fix
      
      * remove linearscaling
      
      * whitespace and import
      
      * fix some copied-from
      
      * code style fix
      
      * reformatted
      
      * add position_embedding to nemotronattention
      
      * rope refactor to only use config, copied-from fix
      
      * format
      
      * Run make fix-copies
      
      * nemotron md with autodoc
      
      * doc  fix
      
      * fix order
      
      * pass check_config_docstrings.py
      
      * fix config_attributes
      
      * remove all llama BC related code
      
      * Use PreTrainedTokenizerFast
      
      * ruff check examples
      
      * conversion script update
      
      * add nemotron to toctree
      6a03942d
    • Francisco Kurucz's avatar
    • Pavel Iakubovskii's avatar
      Update kwargs validation for `preprocess` with decorator (#32024) · fb66ef81
      Pavel Iakubovskii authored
      * BLIP preprocess
      
      * BIT preprocess
      
      * BRIDGETOWER preprocess
      
      * CHAMELEON preprocess
      
      * CHINESE_CLIP preprocess
      
      * CONVNEXT preprocess
      
      * DEIT preprocess
      
      * DONUT preprocess
      
      * DPT preprocess
      
      * FLAVA preprocess
      
      * EFFICIENTNET preprocess
      
      * FUYU preprocess
      
      * GLPN preprocess
      
      * IMAGEGPT preprocess
      
      * INTRUCTBLIPVIDEO preprocess
      
      * VIVIT preprocess
      
      * ZOEDEPTH preprocess
      
      * VITMATTE preprocess
      
      * VIT preprocess
      
      * VILT preprocess
      
      * VIDEOMAE preprocess
      
      * VIDEOLLAVA
      
      * TVP processing
      
      * TVP fixup
      
      * SWIN2SR preprocess
      
      * SIGLIP preprocess
      
      * SAM preprocess
      
      * RT-DETR preprocess
      
      * PVT preprocess
      
      * POOLFORMER preprocess
      
      * PERCEIVER preprocess
      
      * OWLVIT preprocess
      
      * OWLV2 preprocess
      
      * NOUGAT preprocess
      
      * MOBILEVIT preprocess
      
      * MOBILENETV2 preprocess
      
      * MOBILENETV1 preprocess
      
      * LEVIT preprocess
      
      * LAYOUTLMV2 preprocess
      
      * LAYOUTLMV3 preprocess
      
      * Add test
      
      * Update tests
      fb66ef81
    • Fanli Lin's avatar
      add the missing flash attention test marker (#32419) · e85d8639
      Fanli Lin authored
      * add flash attention check
      
      * fix
      
      * fix
      
      * add the missing marker
      
      * bug fix
      
      * add one more
      
      * remove order
      
      * add one more
      e85d8639
  2. 05 Aug, 2024 4 commits
  3. 01 Aug, 2024 2 commits
  4. 31 Jul, 2024 4 commits
  5. 30 Jul, 2024 1 commit
    • Joshua Lochner's avatar
      Fix slow GemmaTokenizer and improve SPM slow -> fast conversion process (#32191) · 6e2d04e4
      Joshua Lochner authored
      * Remove user-defined tokens which can be obtained through merges
      
      * Remove debug line
      
      * formatting
      
      * Refactor spm slow -> fast converter
      
      * revert unnecessary refactor
      
      * set comprehension
      
      * remove test files
      
      * Use `vocab_scores`
      
      * Always replace spiece underline with space in decode
      
      * we no longer need token filtering
      
      * Add save fast load slow unit test
      
      * Remove tokenizers version check
      
      * Remove duplicate code
      
      * Make `<start_of_turn>` and `<end_of_turn>` special tokens
      
      * Bias merge priority with length if score is the same
      
      * Add unit test for merge priority
      
      * CI
      6e2d04e4
  6. 29 Jul, 2024 2 commits
    • Kamil Akesbi's avatar
      Whisper tokenizer word level timestamps (#32197) · 3fbaaaa6
      Kamil Akesbi authored
      * fix _fix_key in PreTrainedModel
      
      * fix _find_longest_common_sequence
      
      * add test
      
      * remove result.json
      
      * nit
      
      * update test
      3fbaaaa6
    • Joao Gante's avatar
      Generate: end-to-end compilation (#30788) · 7ffe25f2
      Joao Gante authored
      * mvp
      
      * added test (a few models need fixes)
      
      * fix a few test cases
      
      * test nits
      
      * harder test 馃槇
      
      * revert changes in stablelm
      
      * test with improved condition
      
      * add todo
      
      * tmp commit
      
      * merged with main
      
      * nits
      
      * add todo
      
      * final corrections
      
      * add docs for generation compilation
      
      * docs nits
      
      * add  tip
      
      * PR suggestions
      
      * add more details to the compilation docs
      
      * fix cache positions
      
      * cache is now init in generate; update docs
      
      * tag test as flaky
      
      * docs
      
      * post rebase make fixup and other nits
      
      * remove unintended changes
      
      * whisper (encoder-decoder) not supported
      
      * move token default updates to ; add tests for token defaults
      
      * push changes
      
      * manual rebase
      
      * chameleon doesn't support this
      
      * fix test_static_cache_mha_mqa_gqa (broken in another PR)
      
      * docs: dynamic is better with end-to-end compilation
      7ffe25f2
  7. 26 Jul, 2024 2 commits
  8. 25 Jul, 2024 3 commits
  9. 24 Jul, 2024 3 commits
  10. 23 Jul, 2024 9 commits
  11. 22 Jul, 2024 4 commits
  12. 19 Jul, 2024 2 commits