1. 23 Nov, 2023 1 commit
  2. 22 Nov, 2023 1 commit
    • Patrick von Platen's avatar
      [Whisper] Add sequential longform decoding (#27492) · 4151fbb4
      Patrick von Platen authored
      * [Whisper] Add seq gen
      
      * [Whisper] Add seq gen
      
      * more debug
      
      * Fix whisper logit processor
      
      * Improve whisper code further
      
      * Fix more
      
      * more debug
      
      * more debug
      
      * Improve further
      
      * Add tests
      
      * Prep for batch size > 1
      
      * Get batch_size>1 working
      
      * Correct more
      
      * Add extensive tests
      
      * more debug
      
      * more debug
      
      * more debug
      
      * add more tests
      
      * more debug
      
      * Apply suggestions from code review
      
      * more debug
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * Add more examples
      
      * add comments to explain the code better
      
      * fix more
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * correct
      
      * correct
      
      * finalize
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      4151fbb4
  3. 16 Nov, 2023 2 commits
    • Arthur's avatar
      [`Styling`] stylify using ruff (#27144) · 651408a0
      Arthur authored
      
      
      * try to stylify using ruff
      
      * might need to remove these changes?
      
      * use ruf format andruff check
      
      * use isinstance instead of type comparision
      
      * use # fmt: skip
      
      * use # fmt: skip
      
      * nits
      
      * soem styling changes
      
      * update ci job
      
      * nits isinstance
      
      * more files update
      
      * nits
      
      * more nits
      
      * small nits
      
      * check and format
      
      * revert wrong changes
      
      * actually use formatter instead of checker
      
      * nits
      
      * well docbuilder is overwriting this commit
      
      * revert notebook changes
      
      * try to nuke docbuilder
      
      * style
      
      * fix feature exrtaction test
      
      * remve `indent-width = 4`
      
      * fixup
      
      * more nits
      
      * update the ruff version that we use
      
      * style
      
      * nuke docbuilder styling
      
      * leve the print for detected changes
      
      * nits
      
      * Remove file I/O
      Co-authored-by: default avatarcharliermarsh <charlie.r.marsh@gmail.com>
      
      * style
      
      * nits
      
      * revert notebook changes
      
      * Add # fmt skip when possible
      
      * Add # fmt skip when possible
      
      * Fix
      
      * More `  # fmt: skip` usage
      
      * More `  # fmt: skip` usage
      
      * More `  # fmt: skip` usage
      
      * NIts
      
      * more fixes
      
      * fix tapas
      
      * Another way to skip
      
      * Recommended way
      
      * Fix two more fiels
      
      * Remove asynch
      Remove asynch
      
      ---------
      Co-authored-by: default avatarcharliermarsh <charlie.r.marsh@gmail.com>
      651408a0
    • Patrick von Platen's avatar
      Revert "add attention_mask and position_ids in assisted model" (#27523) · 5603fad2
      Patrick von Platen authored
      * Revert "add attention_mask and position_ids in assisted model (#26892)"
      
      This reverts commit 184f60dc.
      
      * more debug
      5603fad2
  4. 09 Nov, 2023 1 commit
  5. 01 Nov, 2023 3 commits
  6. 31 Oct, 2023 1 commit
    • Hz, Ji's avatar
      device agnostic models testing (#27146) · 50378cbf
      Hz, Ji authored
      * device agnostic models testing
      
      * add decorator `require_torch_fp16`
      
      * make style
      
      * apply review suggestion
      
      * Oops, the fp16 decorator was misused
      50378cbf
  7. 30 Oct, 2023 1 commit
  8. 11 Oct, 2023 2 commits
  9. 03 Oct, 2023 1 commit
  10. 15 Sep, 2023 1 commit
  11. 14 Sep, 2023 2 commits
    • Matt's avatar
      Overhaul Conversation class and prompt templating (#25323) · 866df66f
      Matt authored
      
      
      * First commit while I figure this out
      
      * make fixup
      
      * Remove unused method
      
      * Store prompt attrib
      
      * Fix prompt argument for tests
      
      * Make same changes in fast tokenizer
      
      * Remove global prompts from fast tokenizer too
      
      * stash commit
      
      * stash commit
      
      * Migrate PromptConfig to its True Final Location
      
      * Replace Conversation entirely with the new class
      
      * Import/dependency fixes
      
      * Import/dependency fixes
      
      * Change format for lots of default prompts
      
      * More default prompt fixups
      
      * Revert llama old methods so we can compare
      
      * Fix some default configs
      
      * Fix some default configs
      
      * Fix misspelled kwarg
      
      * Fixes for Blenderbot
      
      * make fixup
      
      * little rebase cleanup
      
      * Add basic documentation
      
      * Quick doc fix
      
      * Truncate docstring for now
      
      * Add handling for the case when messages is a single string
      
      * Quick llama merges
      
      * Update conversational pipeline and tests
      
      * Add a couple of legacy properties for backward compatibility
      
      * More legacy handling
      
      * Add docstring for build_conversation_input_ids
      
      * Restructure PromptConfig
      
      * Let's start T E M P L A T I N G
      
      * Refactor all default configs to use templates instead
      
      * Revert changes to the special token properties since we don't need them anymore
      
      * More class templates
      
      * Make the sandbox even sandier
      
      * Everything replaced with pure templating
      
      * Remove docs for PromptConfig
      
      * Add testing and optional requirement boilerplate
      
      * Fix imports and make fixup
      
      * Fix LLaMA tests and add Conversation docstring
      
      * Finally get LLaMA working with the template system
      
      * Finally get LLaMA working with the template system
      
      * make fixup
      
      * make fixup
      
      * fmt-off for the long lists of test tokens
      
      * Rename method to apply_chat_template for now
      
      * Start on documentation
      
      * Make chat_template a property that reads through to the default if it's not set
      
      * Expand docs
      
      * Expand chat templating doc some more
      
      * trim/lstrip blocks by default and update doc
      
      * Few doc tweaks
      
      * rebase cleanup
      
      * Clarify docstring
      
      * rebase cleanup
      
      * rebase cleanup
      
      * make fixup
      
      * Quick doc edit
      
      * Reformat the standard template to match ChatML
      
      * Re-add PEFT check
      
      * Update docs/source/en/chat_templating.md
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Add apply_chat_template to the tokenizer doc
      
      * make fixup
      
      * Add doc links
      
      * Fix chat links
      
      * Fix chat links
      
      * Explain system messages in the doc
      
      * Add chat template test
      
      * Proper save-loading for chat template attribute
      
      * Add test skips for layout models
      
      * Remove _build_conversation_input_ids, add default_chat_template to code_llama
      
      * Make sure all LLaMA models are using the latest template
      
      * Remove default_system_prompt block in code_llama because it has no default prompt
      
      * Update ConversationPipeline preprocess
      
      * Add correct #Copied from links to the default_chat_templates
      
      * Remove unneeded type checking line
      
      * Add a dummy mark_processsed method
      
      * Reorganize Conversation to have **deprecated_kwargs
      
      * Update chat_templating.md
      
      * Quick fix to LLAMA tests
      
      * Small doc tweaks
      
      * Add proper docstrings and "copied from" statements to all default chat templates
      
      * Merge use_default_system_prompt support for code_llama too
      
      * Improve clarity around self.chat_template
      
      * Docstring fix
      
      * Fix blenderbot default template
      
      * More doctest fix
      
      * Break out some tokenizer kwargs
      
      * Update doc to explain default templates
      
      * Quick tweaks to tokenizer args
      
      * Cleanups for tokenizer args
      
      * Add note about cacheing
      
      * Quick tweak to the chat-templating doc
      
      * Update the LLaMA template with error checking and correct system message embedding
      
      * make fixup
      
      * make fixup
      
      * add requires_jinja
      
      * Cleanup to expected output formatting
      
      * Add cacheing
      
      * Fix typo in llama default template
      
      * Update LLaMA tests
      
      * Update documentation
      
      * Improved legacy handling in the Conversation class
      
      * Update Jinja template with proper error handling
      
      * Quick bugfix
      
      * Proper exception raising
      
      * Change cacheing behaviour so it doesn't try to pickle an entire Jinja env
      
      * make fixup
      
      * rebase cleanup
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      866df66f
    • Sanchit Gandhi's avatar
      [Whisper Tokenizer] Encode timestamps (#26054) · ac957f69
      Sanchit Gandhi authored
      * [Whisper Tokenizer] Fix tests after adding timestamps
      
      * fix s2t tokenizer tests
      
      * fix vocab test
      
      * backwards comp
      
      * fix tests
      
      * comment
      
      * style
      
      * fix last test
      
      * fix fast
      
      * make faster
      
      * move logic to decode
      
      * remove skip test
      
      * fix decode with offsets
      
      * fix special tokens
      
      * empty commit to re-trigger ci
      
      * use lru cache
      ac957f69
  12. 09 Sep, 2023 1 commit
  13. 17 Aug, 2023 1 commit
  14. 29 Jun, 2023 1 commit
  15. 28 Jun, 2023 1 commit
  16. 27 Jun, 2023 1 commit
  17. 26 Jun, 2023 1 commit
  18. 23 Jun, 2023 1 commit
  19. 21 Jun, 2023 1 commit
    • Matthijs Hollemans's avatar
      add word-level timestamps to Whisper (#23205) · cd927a47
      Matthijs Hollemans authored
      * let's go!
      
      * initial implementation of token-level timestamps
      
      * only return a single timestamp per token
      
      * remove token probabilities
      
      * fix return type
      
      * fix doc comment
      
      * strip special tokens
      
      * rename
      
      * revert to not stripping special tokens
      
      * only support models that have alignment_heads
      
      * add integration test
      
      * consistently name it token-level timestamps
      
      * small DTW tweak
      
      * initial support for ASR pipeline
      
      * fix pipeline doc comments
      
      * resolve token timestamps in pipeline with chunking
      
      * change warning when no final timestamp is found
      
      * return word-level timestamps
      
      * fixup
      
      * fix bug that skipped final word in each chunk
      
      * fix failing unit tests
      
      * merge punctuations into the words
      
      * also return word tokens
      
      * also return token indices
      
      * add (failing) unit test for combine_tokens_into_words
      
      * make combine_tokens_into_words private
      
      * restore OpenAI's punctuation rules
      
      * add pipeline tests
      
      * make requested changes
      
      * PR review changes
      
      * fix failing pipeline test
      
      * small stuff from PR
      
      * only return words and their timestamps, not segments
      
      * move alignment_heads into generation config
      
      * forgot to set alignment_heads in pipeline tests
      
      * tiny comment fix
      
      * grr
      cd927a47
  20. 20 Jun, 2023 1 commit
  21. 06 Jun, 2023 1 commit
    • Matt's avatar
      Move TF building to an actual build() method (#23760) · 4a55e478
      Matt authored
      * A fun new PR where I break the entire codebase again
      
      * A fun new PR where I break the entire codebase again
      
      * Handle cross-attention
      
      * Move calls to model(model.dummy_inputs) to the new build() method
      
      * Seeing what fails with the build context thing
      
      * make fix-copies
      
      * Let's see what fails with new build methods
      
      * Fix the pytorch crossload build calls
      
      * Fix the overridden build methods in vision_text_dual_encoder
      
      * Make sure all our build methods set self.built or call super().build(), which also sets it
      
      * make fix-copies
      
      * Remove finished TODO
      
      * Tentatively remove unneeded (?) line
      
      * Transpose b in deberta correctly and remove unused threading local
      
      * Get rid of build_with_dummies and all it stands for
      
      * Rollback some changes to TF-PT crossloading
      
      * Correctly call super().build()
      4a55e478
  22. 02 Jun, 2023 1 commit
  23. 31 May, 2023 1 commit
  24. 30 May, 2023 1 commit
  25. 24 May, 2023 2 commits
    • Sanchit Gandhi's avatar
      d8222be5
    • Matt's avatar
      Better TF docstring types (#23477) · f8b25744
      Matt authored
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Don't forget the imports
      
      * Add the imports to tests too
      
      * make fixup
      
      * Refactor tests that depended on get_type_hints
      
      * Better test refactor
      
      * Fix an old hidden bug in the test_keras_fit input creation code
      
      * Fix for the Deit tests
      f8b25744
  26. 23 May, 2023 1 commit
  27. 19 May, 2023 1 commit
    • Connor Henderson's avatar
      feat: Whisper prompting (#22496) · 2acedf47
      Connor Henderson authored
      * initial working additions
      
      * clean and rename, add cond stripping initial prompt to decode
      
      * cleanup, edit create_initial_prompt_ids, add tests
      
      * repo consistency, flip order of conditional
      
      * fix error, move the processor fn to the tokenizer
      
      * repo consistency, update test ids to corresponding tokenizer
      
      * use convert_tokens_to_ids not get_vocab...
      
      * use actual conditional in generate
      
      * make sytle
      
      * initial address comments
      
      * initial working add new params to pipeline
      
      * first draft of sequential generation for condition_on_previous_text
      
      * add/update tests, make compatible with timestamps
      
      * make compatible with diff. input kwargs and max length
      
      * add None check
      
      * add temperature check
      
      * flip temp check operand
      
      * refocusing to prev pr scope
      
      * remove the params too
      
      * make style
      
      * edits, move max length incorporating prompt to whisper
      
      * address comments
      
      * remove asr pipeline prompt decoding, fix indexing
      
      * address comments (more tests, validate prompt)
      
      * un-comment out tests (from debug)
      
      * remove old comment
      
      * address comments
      
      * fix typo
      
      * remove timestamp token from test
      
      * make style
      
      * cleanup
      
      * copy method to fast tokenizer, set max_new_tokens for test
      
      * prompt_ids type just pt
      
      * address Amy's comments
      
      * make style
      2acedf47
  28. 11 May, 2023 2 commits
  29. 09 May, 2023 1 commit
    • Matthijs Hollemans's avatar
      audio_utils improvements (#21998) · 7f919509
      Matthijs Hollemans authored
      * silly change to allow making a PR
      
      * clean up doc comments
      
      * simplify hertz_to_mel and mel_to_hertz
      
      * fixup
      
      * clean up power_to_db
      
      * also add amplitude_to_db
      
      * move functions
      
      * clean up mel_filter_bank
      
      * fixup
      
      * credit librosa & torchaudio authors
      
      * add unit tests
      
      * tests for power_to_db and amplitude_to_db
      
      * add mel_filter_bank tests
      
      * rewrite STFT
      
      * add convenience spectrogram function
      
      * missing transpose
      
      * fewer transposes
      
      * add integration test to M-CTC-T
      
      * frame length can be either window or FFT length
      
      * rewrite stft API
      
      * add preemphasis coefficient
      
      * move argument
      
      * add log option to spectrogram
      
      * replace M-CTC-T feature extractor
      
      * fix api thing
      
      * replace whisper STFT
      
      * replace whisper mel filters
      
      * replace tvlt's stft
      
      * allow alternate window names
      
      * replace speecht5 stft
      
      * fixup
      
      * fix integration tests
      
      * fix doc comments
      
      * remove manual FFT length calculation
      
      * fix docs
      
      * go away, deprecation warnings
      
      * combine everything into spectrogram function
      
      * add deprecated functions back
      
      * fixup
      7f919509
  30. 05 May, 2023 2 commits
  31. 04 May, 2023 2 commits