- 04 Aug, 2023 1 commit
-
-
Sanchit Gandhi authored
-
- 13 Jul, 2023 1 commit
-
-
Liyang90 authored
* Update modeling_llama.py Removing unnecessary `device=device` * fix in all occurrences of _make_causal_mask
-
- 11 Jul, 2023 1 commit
-
-
Joao Gante authored
-
- 07 Jul, 2023 1 commit
-
-
Joao Gante authored
-
- 29 Jun, 2023 1 commit
-
-
MS Kim(tony9402) authored
* fix annotations * fix copies
-
- 27 Jun, 2023 1 commit
-
-
Sylvain Gugger authored
* Preliminary work on some models * Fix test load missing and make sure nonpersistent buffers are tested * Always ignore nonpersistent buffers if in state_dict * Treat models * More models * Treat remaining models * Fix quality * Fix tests * Remove draft * This test is not needed anymore * Fix copies * Fix last test * Newly added models * Fix last tests * Address review comments
-
- 26 Jun, 2023 1 commit
-
-
Yih-Dar authored
* fix * fix * fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 23 Jun, 2023 1 commit
-
-
Bowen Bao authored
* Replace python random with torch.rand to enable dynamo.export * revert changes to flax model code * Remove unused random import * Fix torch template * Move torch.manual_seed(0) to right location
-
- 22 Jun, 2023 1 commit
-
-
Younes Belkada authored
Revert "Fix gradient checkpointing + fp16 autocast for most models (#24247)" This reverts commit 285a4801.
-
- 21 Jun, 2023 2 commits
-
-
Matthijs Hollemans authored
* let's go! * initial implementation of token-level timestamps * only return a single timestamp per token * remove token probabilities * fix return type * fix doc comment * strip special tokens * rename * revert to not stripping special tokens * only support models that have alignment_heads * add integration test * consistently name it token-level timestamps * small DTW tweak * initial support for ASR pipeline * fix pipeline doc comments * resolve token timestamps in pipeline with chunking * change warning when no final timestamp is found * return word-level timestamps * fixup * fix bug that skipped final word in each chunk * fix failing unit tests * merge punctuations into the words * also return word tokens * also return token indices * add (failing) unit test for combine_tokens_into_words * make combine_tokens_into_words private * restore OpenAI's punctuation rules * add pipeline tests * make requested changes * PR review changes * fix failing pipeline test * small stuff from PR * only return words and their timestamps, not segments * move alignment_heads into generation config * forgot to set alignment_heads in pipeline tests * tiny comment fix * grr
-
Younes Belkada authored
* fix gc bug * continue PoC on OPT * fixes * :exploding_head: * fix tests * remove pytest.mark * fixup * forward contrib credits from discussions * forward contrib credits from discussions * reverting changes on untouched files. --------- Co-authored-by:
zhaoqf123 <zhaoqf123@users.noreply.github.com> Co-authored-by:
7eu7d7 <7eu7d7@users.noreply.github.com>
-
- 13 Jun, 2023 1 commit
-
-
Sylvain Gugger authored
* First test * Add info for all models * style * Repo consistency * Fix last model and cleanup prints * Repo consistency * Use consistent function for detecting tied weights
-
- 12 Jun, 2023 1 commit
-
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 08 Jun, 2023 1 commit
-
-
Sadra Barikbin authored
-
- 24 May, 2023 1 commit
-
-
Connor Henderson authored
move text_prompt_ids trimming to top
-
- 19 May, 2023 1 commit
-
-
Connor Henderson authored
* initial working additions * clean and rename, add cond stripping initial prompt to decode * cleanup, edit create_initial_prompt_ids, add tests * repo consistency, flip order of conditional * fix error, move the processor fn to the tokenizer * repo consistency, update test ids to corresponding tokenizer * use convert_tokens_to_ids not get_vocab... * use actual conditional in generate * make sytle * initial address comments * initial working add new params to pipeline * first draft of sequential generation for condition_on_previous_text * add/update tests, make compatible with timestamps * make compatible with diff. input kwargs and max length * add None check * add temperature check * flip temp check operand * refocusing to prev pr scope * remove the params too * make style * edits, move max length incorporating prompt to whisper * address comments * remove asr pipeline prompt decoding, fix indexing * address comments (more tests, validate prompt) * un-comment out tests (from debug) * remove old comment * address comments * fix typo * remove timestamp token from test * make style * cleanup * copy method to fast tokenizer, set max_new_tokens for test * prompt_ids type just pt * address Amy's comments * make style
-
- 05 May, 2023 2 commits
-
-
Connor Henderson authored
* add fix * address comments * remove error formatting
-
Andrei Filatov authored
-
- 14 Apr, 2023 1 commit
-
-
oscar-garzon authored
-
- 04 Apr, 2023 1 commit
-
-
Sourab Mangrulkar authored
-
- 28 Mar, 2023 1 commit
-
-
Jeff Rasley authored
* ensure causal_mask is created directly on device * add copy tag to opt, update bart implementation * add device to all _make_causal_mask copies * formatting fixes * more manual fixes due to unlinked versions of _prepare_decoder_attention_mask
-
- 14 Mar, 2023 1 commit
-
-
Arthur authored
* temp fix * temporary fix * update * fix tests * fixup * update based on reveiew Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * update to fix tests * update docstring --------- Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
-
- 13 Mar, 2023 1 commit
-
-
Younes Belkada authored
* add `get_input_embeddings` to `WhisperForAudioClassification` * add common tests * fix another common test * Update tests/models/whisper/test_modeling_whisper.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix style --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
- 11 Mar, 2023 1 commit
-
-
Sanchit Gandhi authored
* [Whisper] Remove embed_tokens from encoder docstring * new line to retrigger CI * remove new line
-
- 08 Mar, 2023 1 commit
-
-
Somasree Majumder authored
* fixing * Update modeling_whisper.py * Update modeling_whisper.py * Update src/transformers/models/whisper/modeling_whisper.py --------- Co-authored-by:Joao Gante <joaofranciscocardosogante@gmail.com>
-
- 07 Mar, 2023 1 commit
-
-
Sanchit Gandhi authored
* [Whisper] Add model for audio classification * make fix-copies * add to docs * add docstring * empty returns * add code example * switch to fleurs * stick everything on one line
-
- 02 Mar, 2023 1 commit
-
-
Kashif Rasul authored
-
- 01 Mar, 2023 1 commit
-
-
raghavanone authored
* Change the .view call to .reshape * Change the .view call to .reshape to all the copies from bart attention * Fix copies and style * Fix copies and style * Fix copies and style * Fix copies and style * Fix copies and style * Revert unneccessary changes * Revert unneccessary changes * Revert unneccessary changes * Revert unneccessary changes
-
- 24 Feb, 2023 1 commit
-
-
bofeng huang authored
* Return and rescale attention_mask * Add SpecAugment to Whisper modeling * Fix test * Update docstring * Add SpecAug related parameters to model config * Add the _mask_input_features function to doc * Fix quality * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Remove dev comments * Add test * Resolve conflict * feat: mask {feature, time} prob fast tests * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
sanchit-gandhi <sanchit@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 16 Feb, 2023 2 commits
-
-
Jonatas Grosman authored
fix bug in reshaping labels
-
Xiaoyang Chen authored
* Update document of WhisperDecoderLayer * Update modeling_mbart.py * Update doc with utils/check_copies.py --fix_and_overwrite * Update modeling_xlm_prophetnet.py
-
- 07 Feb, 2023 1 commit
-
-
Arthur authored
* fix past renamed to past_key_value * update more `past`that were ski^锚d * fixup * remove changes made to rag * refactor `_reorder_cache` to use `past_key_values` * fix git `prepare_inputs_for_generation` to pass tests when false is needed in use_cache
-
- 06 Feb, 2023 1 commit
-
-
Sylvain Gugger authored
* Result of black 23.1 * Update target to Python 3.7 * Switch flake8 to ruff * Configure isort * Configure isort * Apply isort with line limit * Put the right black version * adapt black in check copies * Fix copies
-
- 25 Jan, 2023 1 commit
-
-
Arthur authored
* update whisper logit processor * add generate for whisper * remove part of the whisper specific code from pipeline * update logit processes * major update * enforce first timestamp * update generate * add more tests * update new decoding strategy * Apply suggestions from code review * update docstring * fixup * default config will not have multilingual ar * update expected tokenizer size, see pull on the hub for whisper-tiny
-
- 23 Jan, 2023 1 commit
-
-
Sylvain Gugger authored
* Clean all models * Style * Last to remove * address review comments * Address review comments
-
- 08 Jan, 2023 1 commit
-
-
Arthur authored
* start cleanup * more updates * more models are affected * more updates * update generation utils * style * revert change that removed reorder cachce * update generation utils * style * style * remove reorder cache
-
- 20 Dec, 2022 1 commit
-
-
Sanchit Gandhi authored
* [S2T, Whisper] Add copied from statements * rebase and fix-copies
-
- 06 Dec, 2022 1 commit
-
-
Sourab Mangrulkar authored
* updating T5 and BART models to support Prefix Tuning * `make fix-copies` * address comments * address comments
-
- 05 Dec, 2022 1 commit
-
-
Arthur authored
* Fix whisper and speech to text doc # What does this PR do? Previously the documentation was badly indented for both models and indicated that > If `decoder_input_ids` and `decoder_inputs_embeds` are both unset, `decoder_inputs_embeds` takes the value of `inputs_embeds`.` Which is on valid for the forward pass of the `ForConditionnalGeneration` not for the model alone. * other fixes
-
- 30 Nov, 2022 1 commit
-
-
Yih-Dar authored
* remove truncation * For TFWhisper Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-