1. 22 Dec, 2023 1 commit
  2. 18 Dec, 2023 1 commit
  3. 27 Nov, 2023 1 commit
  4. 22 Nov, 2023 1 commit
    • Patrick von Platen's avatar
      [Whisper] Add sequential longform decoding (#27492) · 4151fbb4
      Patrick von Platen authored
      * [Whisper] Add seq gen
      
      * [Whisper] Add seq gen
      
      * more debug
      
      * Fix whisper logit processor
      
      * Improve whisper code further
      
      * Fix more
      
      * more debug
      
      * more debug
      
      * Improve further
      
      * Add tests
      
      * Prep for batch size > 1
      
      * Get batch_size>1 working
      
      * Correct more
      
      * Add extensive tests
      
      * more debug
      
      * more debug
      
      * more debug
      
      * add more tests
      
      * more debug
      
      * Apply suggestions from code review
      
      * more debug
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * Add more examples
      
      * add comments to explain the code better
      
      * fix more
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * correct
      
      * correct
      
      * finalize
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      4151fbb4
  5. 07 Nov, 2023 1 commit
    • Chi's avatar
      Remove a redundant variable. (#27288) · 9459d821
      Chi authored
      * Removed the redundant SiLUActivation class and now use nn.functional.silu directly.
      
      * I apologize for adding torch.functional.silu. I have replaced it with nn.SiLU.
      
      * Remove redundant variable in feature_extraction file
      9459d821
  6. 03 Nov, 2023 1 commit
  7. 02 Nov, 2023 1 commit
  8. 31 Oct, 2023 2 commits
  9. 27 Oct, 2023 1 commit
  10. 16 Oct, 2023 1 commit
    • Matt's avatar
      Conversation pipeline fixes (#26795) · 14b04b4b
      Matt authored
      * Adjust length limits and allow naked conversation list inputs
      
      * Adjust length limits and allow naked conversation list inputs
      
      * Maybe use a slightly more reasonable limit than 1024
      
      * Skip tests for old models that never supported this anyway
      
      * Cleanup input docstrings
      
      * More docstring cleanup + skip failing TF test
      
      * Make fixup
      14b04b4b
  11. 13 Oct, 2023 1 commit
  12. 12 Oct, 2023 2 commits
  13. 04 Oct, 2023 2 commits
  14. 03 Oct, 2023 1 commit
  15. 29 Sep, 2023 2 commits
  16. 22 Sep, 2023 2 commits
  17. 15 Sep, 2023 1 commit
    • Matt's avatar
      Tweaks to Chat Templates docs (#26168) · 2518e368
      Matt authored
      * Put tokenizer methods in the right alphabetical order in the docs
      
      * Quick tweak to ConversationalPipeline
      
      * Typo fixes in the developer doc
      
      * make fixup
      2518e368
  18. 14 Sep, 2023 2 commits
    • Joshua Lochner's avatar
      [Whisper] Fix word-level timestamps for audio < 30 seconds (#25607) · 95fe0f5d
      Joshua Lochner authored
      
      
      * Fix word-level timestamps for audio < 30 seconds
      
      * Fix code quality
      
      * fix unit tests
      
      * Fix unit tests
      
      * Fix unit test
      
      * temp: print out result
      
      * temp: set max diff to None
      
      * fix unit tests
      
      * fix typo
      
      * Fix typo
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Use generation config for `num_frames`
      
      * fix docs
      
      * Move `num_frames` to kwargs
      
      * compute stride/attn_mask once
      
      * mark test as slow
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarsanchit-gandhi <sanchit@huggingface.co>
      95fe0f5d
    • Matt's avatar
      Overhaul Conversation class and prompt templating (#25323) · 866df66f
      Matt authored
      
      
      * First commit while I figure this out
      
      * make fixup
      
      * Remove unused method
      
      * Store prompt attrib
      
      * Fix prompt argument for tests
      
      * Make same changes in fast tokenizer
      
      * Remove global prompts from fast tokenizer too
      
      * stash commit
      
      * stash commit
      
      * Migrate PromptConfig to its True Final Location
      
      * Replace Conversation entirely with the new class
      
      * Import/dependency fixes
      
      * Import/dependency fixes
      
      * Change format for lots of default prompts
      
      * More default prompt fixups
      
      * Revert llama old methods so we can compare
      
      * Fix some default configs
      
      * Fix some default configs
      
      * Fix misspelled kwarg
      
      * Fixes for Blenderbot
      
      * make fixup
      
      * little rebase cleanup
      
      * Add basic documentation
      
      * Quick doc fix
      
      * Truncate docstring for now
      
      * Add handling for the case when messages is a single string
      
      * Quick llama merges
      
      * Update conversational pipeline and tests
      
      * Add a couple of legacy properties for backward compatibility
      
      * More legacy handling
      
      * Add docstring for build_conversation_input_ids
      
      * Restructure PromptConfig
      
      * Let's start T E M P L A T I N G
      
      * Refactor all default configs to use templates instead
      
      * Revert changes to the special token properties since we don't need them anymore
      
      * More class templates
      
      * Make the sandbox even sandier
      
      * Everything replaced with pure templating
      
      * Remove docs for PromptConfig
      
      * Add testing and optional requirement boilerplate
      
      * Fix imports and make fixup
      
      * Fix LLaMA tests and add Conversation docstring
      
      * Finally get LLaMA working with the template system
      
      * Finally get LLaMA working with the template system
      
      * make fixup
      
      * make fixup
      
      * fmt-off for the long lists of test tokens
      
      * Rename method to apply_chat_template for now
      
      * Start on documentation
      
      * Make chat_template a property that reads through to the default if it's not set
      
      * Expand docs
      
      * Expand chat templating doc some more
      
      * trim/lstrip blocks by default and update doc
      
      * Few doc tweaks
      
      * rebase cleanup
      
      * Clarify docstring
      
      * rebase cleanup
      
      * rebase cleanup
      
      * make fixup
      
      * Quick doc edit
      
      * Reformat the standard template to match ChatML
      
      * Re-add PEFT check
      
      * Update docs/source/en/chat_templating.md
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Add apply_chat_template to the tokenizer doc
      
      * make fixup
      
      * Add doc links
      
      * Fix chat links
      
      * Fix chat links
      
      * Explain system messages in the doc
      
      * Add chat template test
      
      * Proper save-loading for chat template attribute
      
      * Add test skips for layout models
      
      * Remove _build_conversation_input_ids, add default_chat_template to code_llama
      
      * Make sure all LLaMA models are using the latest template
      
      * Remove default_system_prompt block in code_llama because it has no default prompt
      
      * Update ConversationPipeline preprocess
      
      * Add correct #Copied from links to the default_chat_templates
      
      * Remove unneeded type checking line
      
      * Add a dummy mark_processsed method
      
      * Reorganize Conversation to have **deprecated_kwargs
      
      * Update chat_templating.md
      
      * Quick fix to LLAMA tests
      
      * Small doc tweaks
      
      * Add proper docstrings and "copied from" statements to all default chat templates
      
      * Merge use_default_system_prompt support for code_llama too
      
      * Improve clarity around self.chat_template
      
      * Docstring fix
      
      * Fix blenderbot default template
      
      * More doctest fix
      
      * Break out some tokenizer kwargs
      
      * Update doc to explain default templates
      
      * Quick tweaks to tokenizer args
      
      * Cleanups for tokenizer args
      
      * Add note about cacheing
      
      * Quick tweak to the chat-templating doc
      
      * Update the LLaMA template with error checking and correct system message embedding
      
      * make fixup
      
      * make fixup
      
      * add requires_jinja
      
      * Cleanup to expected output formatting
      
      * Add cacheing
      
      * Fix typo in llama default template
      
      * Update LLaMA tests
      
      * Update documentation
      
      * Improved legacy handling in the Conversation class
      
      * Update Jinja template with proper error handling
      
      * Quick bugfix
      
      * Proper exception raising
      
      * Change cacheing behaviour so it doesn't try to pickle an entire Jinja env
      
      * make fixup
      
      * rebase cleanup
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      866df66f
  19. 12 Sep, 2023 1 commit
  20. 01 Sep, 2023 2 commits
  21. 31 Aug, 2023 1 commit
  22. 30 Aug, 2023 1 commit
    • Juan Pizarro's avatar
      Add Blip2 model in VQA pipeline (#25532) · 09dc9951
      Juan Pizarro authored
      * Add Blip2 model in VQA pipeline
      
      * use require_torch_gpu for test_large_model_pt_blip2
      
      * use can_generate in vqa pipeline
      
      * test Blip2ForConditionalGeneration using float16
      
      * remove custom can_generate from Blip2ForConditionalGeneration
      09dc9951
  23. 25 Aug, 2023 1 commit
    • Arthur's avatar
      [`CodeLlama`] Add support for `CodeLlama` (#25740) · 015f8e11
      Arthur authored
      
      
      * add all
      
      * Revert "Delete .github directory"
      
      This reverts commit 9b0ff7b052e2b20b629a26fb13606b78a42944d1.
      
      * make conversion script backward compatible
      
      * fixup
      
      * more styling
      
      * copy to llama changes
      
      * fix repo consistency
      
      * nits
      
      * document correct classes
      
      * updates
      
      * more fixes
      
      * nits
      
      * update auto mappings
      
      * add readmes
      
      * smallupdates
      
      * llama-code replace with llama_code
      
      * make fixup
      
      * updates to the testsing suite
      
      * fix fast nits
      
      * more small fixes
      
      * fix decode
      
      * fix template processing
      
      * properly reset the normalizer
      
      * nits processor
      
      * tokenization tests pass
      
      * styling
      
      * last tests
      
      * additional nits
      
      * one test is left
      
      * nits
      
      Co-authored-by faabian <faabian@users.noreply.github.com>
      
      * update failing test
      
      * fixup
      
      * remove decode infilling users should handle it on their onw after generation, padding can be a problem
      
      * update
      
      * make test slow and more meaningfull
      
      * fixup
      
      * doc update
      
      * fixup
      
      * Apply suggestions from code review
      
      * add kwargs doc
      
      * tokenizer requires `requires_backend`
      
      * type requires_backends
      
      * CodeLlama instead of LlamaCode
      
      * more name cahnges
      
      * nits
      
      * make doctests happy
      
      * small pipeline nits
      
      * last nit
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * update
      
      * add codellama to toctree
      
      ---------
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      015f8e11
  24. 24 Aug, 2023 1 commit
  25. 21 Aug, 2023 1 commit
  26. 18 Aug, 2023 1 commit
  27. 17 Aug, 2023 1 commit
    • Yoach Lacombe's avatar
      Add Text-To-Speech pipeline (#24952) · b8f69d0d
      Yoach Lacombe authored
      
      
      * add AutoModelForTextToSpeech class
      
      * add TTS pipeline and tessting
      
      * add docstrings to text_to_speech pipeline
      
      * fix torch dependency
      
      * corrector 'processor is None' case in Pipeline
      
      * correct repo id
      
      * modify text-to-speech -> text-to-audio
      
      * remove processor
      
      * rename text_to_speech pipelines files to text_audio
      
      * add textToWaveform and textToSpectrogram instead of textToAudio classes
      
      * update TTS pipeline to the bare minimum
      
      * update tests TTS pipeline
      
      * make style and erase useless import torch in TTS pipeline tests
      
      * modify how to check if generate or forward in TTS pipeline
      
      * remove unnecessary extra new lines
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * refactor input_texts -> text_inputs
      
      * correct docstrings of TTS.__call__
      
      * correct the shape of generated waveform
      
      * take care of Bark tokenizer special case
      
      * correct run_pipeline_test TTS
      
      * make style
      
      * update TTS docstrings
      
      * address Sylvain nit refactors
      
      * make style
      
      * refactor into one liners
      
      * correct squeeze
      
      * correct way to test if forward or generate
      
      * Update output audio waveform shape
      
      * make style
      
      * correct import
      
      * modify how the TTS pipeline test if a model can generate
      
      * align shape output of TTS pipeline with consistent shape
      
      ---------
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      b8f69d0d
  28. 16 Aug, 2023 1 commit
  29. 08 Aug, 2023 1 commit
  30. 03 Aug, 2023 1 commit
  31. 31 Jul, 2023 1 commit
  32. 28 Jul, 2023 1 commit
  33. 27 Jul, 2023 1 commit