1. 27 Oct, 2023 3 commits
    • Patrick von Platen's avatar
      [Attention Mask] Refactor all encoder-decoder attention mask (#27086) · ac589375
      Patrick von Platen authored
      
      
      * [FA2 Bart] Add FA2 to all Bart-like
      
      * better
      
      * Refactor attention mask
      
      * remove all customized atteniton logic
      
      * format
      
      * mass rename
      
      * replace _expand_mask
      
      * replace _expand_mask
      
      * mass rename
      
      * add pt files
      
      * mass replace & rename
      
      * mass replace & rename
      
      * mass replace & rename
      
      * mass replace & rename
      
      * Update src/transformers/models/idefics/modeling_idefics.py
      
      * fix more
      
      * clean more
      
      * fix more
      
      * make style
      
      * fix again
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * small fix mistral
      
      * finish
      
      * finish
      
      * finish
      
      * finish
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      ac589375
    • Isaac Chung's avatar
      Add early stopping for Bark generation via logits processor (#26675) · e2bffcfa
      Isaac Chung authored
      * add early stopping logits processor
      
      * black formmated
      
      * indent
      
      * follow method signature
      
      * actual logic
      
      * check for None
      
      * address comments on docstrings and method signature
      
      * add unit test under `LogitsProcessorTest` wip
      
      * unit test passing
      
      * black formatted
      
      * condition per sample
      
      * add to BarkModelIntegrationTests
      
      * wip BarkSemanticModelTest
      
      * rename and add to kwargs handling
      
      * not add to BarkSemanticModelTest
      
      * correct logic and assert last outputs tokens different in test
      
      * doc-builder style
      
      * read from kwargs as well
      
      * assert len of with less than that of without
      
      * ruff
      
      * add back seed and test case
      
      * add original impl default suggestion
      
      * doc-builder
      
      * rename and use softmax
      
      * switch back to LogitsProcessor and update docs wording
      
      * camelCase and spelling and saving compute
      
      * assert strictly less than
      
      * assert less than
      
      * expand test_generate_semantic_early_stop instead
      e2bffcfa
    • Arthur's avatar
      Revert "add exllamav2 arg" (#27102) · 90ee9cea
      Arthur authored
      Revert "add exllamav2 arg (#26437)"
      
      This reverts commit 8214d6e7.
      90ee9cea
  2. 26 Oct, 2023 3 commits
    • Zach Mueller's avatar
      Save TB logs as part of push_to_hub (#27022) · 34a64064
      Zach Mueller authored
      * Support runs/
      
      * Upload runs folder as part of push to hub
      
      * Add a test
      
      * Add to test deps
      
      * Update with proposed solution from Slack
      
      * Ensure that repo gets deleted in tests
      34a64064
    • Marc Sun's avatar
      add exllamav2 arg (#26437) · 8214d6e7
      Marc Sun authored
      * add_ xllamav2 arg
      
      * add test
      
      * style
      
      * add check
      
      * add doc
      
      * replace by use_exllama_v2
      
      * fix tests
      
      * fix doc
      
      * style
      
      * better condition
      
      * fix logic
      
      * add deprecate msg
      8214d6e7
    • Arthur's avatar
      Add-support for commit description (#26704) · 4864d08d
      Arthur authored
      * fix
      
      * update
      
      * revert
      
      * add dosctring
      
      * good to go
      
      * update
      
      * add a test
      4864d08d
  3. 25 Oct, 2023 2 commits
  4. 24 Oct, 2023 3 commits
    • JB (Don)'s avatar
      Add a default decoder_attention_mask for EncoderDecoderModel during training (#26752) · a0fd3448
      JB (Don) authored
      * Add a default decoder_attention_mask for EncoderDecoderModel during training
      
      Since we are already creating the default decoder_input_ids from the labels, we should also
      create a default decoder_attention_mask to go with it.
      
      * Fix test constant that relied on manual_seed()
      
      The test was changed to use a decoder_attention_mask that ignores padding instead (which is
      the default one created by BERT when attention_mask is None).
      
      * Create the decoder_attention_mask using decoder_input_ids instead of labels
      
      * Fix formatting in test
      a0fd3448
    • Alex McKinney's avatar
      Device agnostic testing (#25870) · 9da45171
      Alex McKinney authored
      
      
      * adds agnostic decorators and availability fns
      
      * renaming decorators and fixing imports
      
      * updating some representative example tests
      bloom, opt, and reformer for now
      
      * wip device agnostic functions
      
      * lru cache to device checking functions
      
      * adds `TRANSFORMERS_TEST_DEVICE_SPEC`
      if present, imports the target file and updates device to function
      mappings
      
      * comments `TRANSFORMERS_TEST_DEVICE_SPEC` code
      
      * extra checks on device name
      
      * `make style; make quality`
      
      * updates default functions for agnostic calls
      
      * applies suggestions from review
      
      * adds `is_torch_available` guard
      
      * Add spec file to docs, rename function dispatch names to backend_*
      
      * add backend import to docs example for spec file
      
      * change instances of  to
      
      * Move register backend to before device check as per @statelesshz changes
      
      * make style
      
      * make opt test require fp16 to run
      
      ---------
      Co-authored-by: default avatararsalanu <arsalanu@graphcore.ai>
      Co-authored-by: default avatararsalanu <hzji210@gmail.com>
      9da45171
    • Xuehai Pan's avatar
      Register ModelOutput as supported torch pytree nodes (#26618) · cc7803c0
      Xuehai Pan authored
      * Register ModelOutput as supported torch pytree nodes
      
      * Test ModelOutput as supported torch pytree nodes
      
      * Update type hints for pytree unflatten functions
      cc7803c0
  5. 23 Oct, 2023 3 commits
    • Patrick von Platen's avatar
      Remove ambiguous `padding_mask` and instead use a 2D->4D Attn Mask Mapper (#26792) · 33f98cfd
      Patrick von Platen authored
      
      
      * [Attn Mask Converter] refactor attn mask
      
      * up
      
      * Apply suggestions from code review
      Co-authored-by: default avatarfxmarty <9808326+fxmarty@users.noreply.github.com>
      
      * improve
      
      * rename
      
      * better cache
      
      * renaming
      
      * improve more
      
      * improve
      
      * fix bug
      
      * finalize
      
      * make style & make fix-copies
      
      * correct more
      
      * start moving attention_mask
      
      * fix llama
      
      * improve falcon
      
      * up
      
      * improve more
      
      * improve more
      
      * Update src/transformers/models/owlv2/modeling_owlv2.py
      
      * make style
      
      * make style
      
      * rename to converter
      
      * Apply suggestions from code review
      
      ---------
      Co-authored-by: default avatarfxmarty <9808326+fxmarty@users.noreply.github.com>
      33f98cfd
    • Yoach Lacombe's avatar
      Add Seamless M4T model (#25693) · cb45f71c
      Yoach Lacombe authored
      
      
      * first raw commit
      
      * still POC
      
      * tentative convert script
      
      * almost working speech encoder conversion scripts
      
      * intermediate code for encoder/decoders
      
      * add modeling code
      
      * first version of speech encoder
      
      * make style
      
      * add new adapter layer architecture
      
      * add adapter block
      
      * add first tentative config
      
      * add working speech encoder conversion
      
      * base model convert works now
      
      * make style
      
      * remove unnecessary classes
      
      * remove unecessary functions
      
      * add modeling code speech encoder
      
      * rework logics
      
      * forward pass of sub components work
      
      * add modeling codes
      
      * some config modifs and modeling code modifs
      
      * save WIP
      
      * new edits
      
      * same output speech encoder
      
      * correct attention mask
      
      * correct attention mask
      
      * fix generation
      
      * new generation logics
      
      * erase comments
      
      * make style
      
      * fix typo
      
      * add some descriptions
      
      * new state
      
      * clean imports
      
      * add tests
      
      * make style
      
      * make beam search and num_return_sequences>1 works
      
      * correct edge case issue
      
      * correct SeamlessM4TConformerSamePadLayer copied from
      
      * replace ACT2FN relu by nn.relu
      
      * remove unecessary return variable
      
      * move back a class
      
      * change name conformer_attention_mask ->conv_attention_mask
      
      * better nit code
      
      * add some Copied from statements
      
      * small nits
      
      * small nit in dict.get
      
      * rename t2u model -> conditionalgeneration
      
      * ongoing refactoring of structure
      
      * update models architecture
      
      * remove SeamlessM4TMultiModal classes
      
      * add tests
      
      * adapt tests
      
      * some non-working code for vocoder
      
      * add seamlessM4T vocoder
      
      * remove buggy line
      
      * fix some hifigan related bugs
      
      * remove hifigan specifc config
      
      * change
      
      * add WIP tokenization
      
      * add seamlessM4T working tokenzier
      
      * update tokenization
      
      * add tentative feature extractor
      
      * Update converting script
      
      * update working FE
      
      * refactor input_values -> input_features
      
      * update FE
      
      * changes in generation, tokenizer and modeling
      
      * make style and add t2u_decoder_input_ids
      
      * add intermediate outputs for ToSpeech models
      
      * add vocoder to speech models
      
      * update valueerror
      
      * update FE with languages
      
      * add vocoder convert
      
      * update config docstrings and names
      
      * update generation code and configuration
      
      * remove todos and update config.pad_token_id to generation_config.pad_token_id
      
      * move block vocoder
      
      * remove unecessary code and uniformize tospeech code
      
      * add feature extractor import
      
      * make style and fix some copies from
      
      * correct consistency + make fix-copies
      
      * add processor code
      
      * remove comments
      
      * add fast tokenizer support
      
      * correct pad_token_id in M4TModel
      
      * correct config
      
      * update tests and codes  + make style
      
      * make some suggested correstion - correct comments and change naming
      
      * rename some attributes
      
      * rename some attributes
      
      * remove unecessary sequential
      
      * remove option to use dur predictor
      
      * nit
      
      * refactor hifigan
      
      * replace normalize_mean and normalize_var with do_normalize + save lang ids to generation config
      
      * add tests
      
      * change tgt_lang logic
      
      * update generation ToSpeech
      
      * add support import SeamlessM4TProcessor
      
      * fix generate
      
      * make tests
      
      * update integration tests, add option to only return text and update tokenizer fast
      
      * fix wrong function call
      
      * update import and convert script
      
      * update integration tests + update repo id
      
      * correct paths and add first test
      
      * update how new attention masks are computed
      
      * update tests
      
      * take first care of batching in vocoder code
      
      * add batching with the vocoder
      
      * add waveform lengths to model outputs
      
      * make style
      
      * add generate kwargs + forward kwargs of M4TModel
      
      * add docstrings forward methods
      
      * reformate docstrings
      
      * add docstrings t2u model
      
      * add another round of modeling docstrings + reformate speaker_id -> spkr_id
      
      * make style
      
      * fix check_repo
      
      * make style
      
      * add seamlessm4t to toctree
      
      * correct check_config_attributes
      
      * write config docstrings + some modifs
      
      * make style
      
      * add docstrings tokenizer
      
      * add docstrings to processor, fe and tokenizers
      
      * make style
      
      * write first version of model docs
      
      * fix FE + correct FE test
      
      * fix tokenizer + add correct integration tests
      
      * fix most tokenization tests
      
      * make style
      
      * correct most processor test
      
      * add generation tests and fix num_return_sequences > 1
      
      * correct integration tests -still one left
      
      * make style
      
      * correct position embedding
      
      * change numbeams to 1
      
      * refactor some modeling code and correct one test
      
      * make style
      
      * correct typo
      
      * refactor intermediate fnn
      
      * refactor feedforward conformer
      
      * make style
      
      * remove comments
      
      * make style
      
      * fix tokenizer tests
      
      * make style
      
      * correct processor tests
      
      * make style
      
      * correct S2TT integration
      
      * Apply suggestions from Sanchit code review
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * correct typo
      
      * replace torch.nn->nn + make style
      
      * change Output naming (waveforms -> waveform) and ordering
      
      * nit renaming and formating
      
      * remove return None when not necessary
      
      * refactor SeamlessM4TConformerFeedForward
      
      * nit typo
      
      * remove almost copied from comments
      
      * add a copied from comment and remove an unecessary dropout
      
      * remove inputs_embeds from speechencoder
      
      * remove backward compatibiliy function
      
      * reformate class docstrings for a few components
      
      * remove unecessary methods
      
      * split over 2 lines smthg hard to read
      
      * make style
      
      * replace two steps offset by one step as suggested
      
      * nice typo
      
      * move warnings
      
      * remove useless lines from processor
      
      * make generation non-standard test more robusts
      
      * remove torch.inference_mode from tests
      
      * split integration tests
      
      * enrich md
      
      * rename control_symbol_vocoder_offset->vocoder_offset
      
      * clean convert file
      
      * remove tgt_lang and src_lang from FE
      
      * change generate docstring of ToText models
      
      * update generate docstring of tospeech models
      
      * unify how to deal withtext_decoder_input_ids
      
      * add default spkr_id
      
      * unify tgt_lang for t2u_model
      
      * simplify tgt_lang verification
      
      * remove a todo
      
      * change config docstring
      
      * make style
      
      * simplify t2u_tgt_lang_id
      
      * make style
      
      * enrich/correct comments
      
      * enrich .md
      
      * correct typo in docstrings
      
      * add torchaudio dependency
      
      * update tokenizer
      
      * make style and fix copies
      
      * modify SeamlessM4TConverter with new tokenizer behaviour
      
      * make style
      
      * correct small typo docs
      
      * fix import
      
      * update docs and add requirement to tests
      
      * add convert_fairseq2_to_hf in utils/not_doctested.txt
      
      * update FE
      
      * fix imports and make style
      
      * remove torchaudio in FE test
      
      * add seamless_m4t.md to utils/not_doctested.txt
      
      * nits and change the way docstring dataset is loaded
      
      * move checkpoints from ylacombe/ to facebook/ orga
      
      * refactor warning/error to be in the 119 line width limit
      
      * round overly precised floats
      
      * add stereo audio behaviour
      
      * refactor .md and make style
      
      * enrich docs with more precised architecture description
      
      * readd undocumented models
      
      * make fix-copies
      
      * apply some suggestions
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * correct bug from previous commit
      
      * refactor a parameter allowing to clean the code + some small nits
      
      * clean tokenizer
      
      * make style and fix
      
      * make style
      
      * clean tokenizers arguments
      
      * add precisions for some tests
      
      * move docs from not_tested to slow
      
      * modify tokenizer according to last comments
      
      * add copied from statements in tests
      
      * correct convert script
      
      * correct parameter docstring style
      
      * correct tokenization
      
      * correct multi gpus
      
      * make style
      
      * clean modeling code
      
      * make style
      
      * add copied from statements
      
      * add copied statements
      
      * add support with ASR pipeline
      
      * remove file added inadvertently
      
      * fix docstrings seamlessM4TModel
      
      * add seamlessM4TConfig to OBJECTS_TO_IGNORE due of unconventional markdown
      
      * add seamlessm4t to assisted generation ignored models
      
      ---------
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      cb45f71c
    • Arthur's avatar
      skip two tests (#27013) · ef978d0a
      Arthur authored
      * skip two tests
      
      * skip torch as well
      
      * fixup
      ef978d0a
  6. 20 Oct, 2023 1 commit
    • Pedro Cuenca's avatar
      Fix Fuyu image scaling bug (#26918) · c030fc89
      Pedro Cuenca authored
      * Fix Fuyu image scaling bug
      
      It could produce negative padding and hence inference errors for certain
      image sizes.
      
      * Fix aspect ratio scaling test
      c030fc89
  7. 19 Oct, 2023 1 commit
  8. 18 Oct, 2023 4 commits
    • Pablo Montalvo's avatar
      Add fuyu model (#26911) · caa0ff0b
      Pablo Montalvo authored
      
      
      * initial commit
      
      * add processor, add fuyu naming
      
      * add draft processor
      
      * fix processor
      
      * remove dropout to fix loading of weights
      
      * add image processing fixes from Pedro
      
      * fix
      
      * fix processor
      
      * add basic processing fuyu test
      
      * add documentation and TODO
      
      * address comments, add tests, add doc
      
      * replace assert with torch asserts
      
      * add Mixins and fix tests
      
      * clean imports
      
      * add model tester, clean imports
      
      * fix embedding test
      
      * add updated tests from pre-release model
      
      * Processor: return input_ids used for inference
      
      * separate processing and model tests
      
      * relax test tolerance for embeddings
      
      * add test for logit comparison
      
      * make sure fuyu image processor is imported in the init
      
      * fix formattingh
      
      * more formatting issues
      
      * and more
      
      * fixups
      
      * remove some stuff
      
      * nits
      
      * update init
      
      * remove the fuyu file
      
      * Update integration test with release model
      
      * Update conversion script.
      
      The projection is not used, as confirmed by the authors.
      
      * improve geenration
      
      * Remove duplicate function
      
      * Trickle down patches to model call
      
      * processing fuyu updates
      
      * remove things
      
      * fix prepare_inputs_for_generation to fix generate()
      
      * remove model_input
      
      * update
      
      * add generation tests
      
      * nits
      
      * draft leverage automodel and autoconfig
      
      * nits
      
      * fix dtype patch
      
      * address comments, update READMEs and doc, include tests
      
      * add working processing test, remove refs to subsequences
      
      * add tests, remove Sequence classification
      
      * processing
      
      * update
      
      * update the conversion script
      
      * more processing cleanup
      
      * safe import
      
      * take out ModelTesterMixin for early release
      
      * more cl;eanup
      
      * more cleanup
      
      * more cleanup
      
      * and more
      
      * register a buffer
      
      * nits
      
      * add postprocessing of generate output
      
      * nits
      
      * updates
      
      * add one working test
      
      * fix test
      
      * make fixup works
      
      * fixup
      
      * Arthur's updates
      
      * nits
      
      * update
      
      * update
      
      * fix processor
      
      * update tests
      
      * passe more fixups
      
      * fix
      
      * nits
      
      * don't import torch
      
      * skip fuyu config for now
      
      * fixup done
      
      * fixup
      
      * update
      
      * oups
      
      * nits
      
      * Use input embeddings
      
      * no buffer
      
      * update
      
      * styling processing fuyu
      
      * fix test
      
      * update licence
      
      * protect torch import
      
      * fixup and update not doctested
      
      * kwargs should be passed
      
      * udpates
      
      * update the impofixuprts in the test
      
      * protect import
      
      * protecting imports
      
      * protect imports in type checking
      
      * add testing decorators
      
      * protect top level import structure
      
      * fix typo
      
      * fix check init
      
      * move requires_backend to functions
      
      * Imports
      
      * Protect types
      
      ---------
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      Co-authored-by: default avatarArthurZucker <arthur.zucker@gmail.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarLysandre <lysandre@huggingface.co>
      caa0ff0b
    • Younes Belkada's avatar
      [`FA-2`] Final fix for FA2 dtype (#26846) · 5a73316b
      Younes Belkada authored
      
      
      * final fix for FA2 dtype
      
      * try
      
      * oops
      
      * Update src/transformers/models/falcon/modeling_falcon.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * apply fix everywhere
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      5a73316b
    • Matt's avatar
      de55ead1
    • Arthur's avatar
      [`Tokenizer`] Fix slow and fast serialization (#26570) · ef7e9369
      Arthur authored
      * fix
      
      * last attempt
      
      * current work
      
      * fix forward compatibility
      
      * save all special tokens
      
      * current state
      
      * revert additional changes
      
      * updates
      
      * remove tokenizer.model
      
      * add a test and the fix
      
      * nit
      
      * revert one more break
      
      * fix typefield issue
      
      * quality
      
      * more tests
      
      * fix fields for FC
      
      * more nits?
      
      * new additional changes
      
      * how
      
      * some updates
      
      * simplify all
      
      * more nits
      
      * revert some things to original
      
      * nice
      
      * nits
      
      * a small hack
      
      * more nits
      
      * ahhaha
      
      * fixup
      
      * update
      
      * make test run on ci
      
      * use subtesting
      
      * update
      
      * Update .circleci/create_circleci_config.py
      
      * updates
      
      * fixup
      
      * nits
      
      * replace typo
      
      * fix the test
      
      * nits
      
      * update
      
      * None max dif pls
      
      * a partial fix
      
      * had to revert one thing
      
      * test the fast
      
      * updates
      
      * fixup
      
      * and more nits
      
      * more fixes
      
      * update
      
      * Oupsy 馃憗
      
      
      
      * nits
      
      * fix marian
      
      * on our way to heaven
      
      * Update src/transformers/models/t5/tokenization_t5.py
      Co-authored-by: default avatarLysandre Debut <hi@lysand.re>
      
      * fixup
      
      * Update src/transformers/tokenization_utils_fast.py
      Co-authored-by: default avatarLeo Tronchon <leo.tronchon@gmail.com>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarLeo Tronchon <leo.tronchon@gmail.com>
      
      * fix phobert
      
      * skip some things, test more
      
      * nits
      
      * fixup
      
      * fix deberta
      
      * update
      
      * update
      
      * more updates
      
      * skip one test
      
      * more updates
      
      * fix camembert
      
      * can't test this one
      
      * more good fixes
      
      * kind of a major update
      
      - seperate what is only done in fast in fast init and refactor
      - add_token(AddedToken(..., speicla = True)) ignores it in fast
      - better loading
      
      * fixup
      
      * more fixups
      
      * fix pegasus and mpnet
      
      * remove skipped tests
      
      * fix phoneme tokenizer if self.verbose
      
      * fix individual models
      
      * update common tests
      
      * update testing files
      
      * all over again
      
      * nits
      
      * skip test for markup lm
      
      * fixups
      
      * fix order of addition in fast by sorting the added tokens decoder
      
      * proper defaults for deberta
      
      * correct default for fnet
      
      * nits on add tokens, string initialized to special if special
      
      * skip irrelevant herbert tests
      
      * main fixes
      
      * update test added_tokens_serialization
      
      * the fix for bart like models and class instanciating
      
      * update bart
      
      * nit!
      
      * update idefix test
      
      * fix whisper!
      
      * some fixup
      
      * fixups
      
      * revert some of the wrong chanegs
      
      * fixup
      
      * fixup
      
      * skip marian
      
      * skip the correct tests
      
      * skip for tf and flax as well
      
      ---------
      Co-authored-by: default avatarLysandre Debut <hi@lysand.re>
      Co-authored-by: default avatarLeo Tronchon <leo.tronchon@gmail.com>
      ef7e9369
  9. 17 Oct, 2023 2 commits
  10. 16 Oct, 2023 3 commits
  11. 13 Oct, 2023 4 commits
    • NielsRogge's avatar
      Add OWLv2, bis (#26668) · 762af3e3
      NielsRogge authored
      * First draft
      
      * Update conversion script
      
      * Update copied from statements
      
      * Fix style
      
      * Add copied from to config
      
      * Add copied from to processor
      
      * Run make fixup
      
      * Add docstring
      
      * Update docstrings
      
      * Add method
      
      * Improve docstrings
      
      * Fix docstrings
      
      * Improve docstrings
      
      * Remove onnx
      
      * Add flag
      
      * Address comments
      
      * Add copied from to model tests
      
      * Add flag to conversion script
      
      * Add code snippet
      
      * Address more comments
      
      * Address comment
      
      * Improve conversion script
      
      * More improvements
      
      * Add expected objectness logits
      
      * Skip test
      
      * Improve conversion script
      
      * Extend conversion script
      
      * Convert large checkpoint
      
      * Fix doc tests
      
      * Convert all checkpoints, update integration tests
      
      * Add checkpoint_path arg
      
      * Fix repo_id
      762af3e3
    • Matt's avatar
      Fix Falcon generation test (#26770) · bdb391e9
      Matt authored
      bdb391e9
    • Matt's avatar
      Disable default system prompt for LLaMA (#26765) · c9785d95
      Matt authored
      * Disable default system prompt for LLaMA
      
      * Update test to not expect default prompt
      c9785d95
    • Yih-Dar's avatar
  12. 12 Oct, 2023 6 commits
  13. 11 Oct, 2023 4 commits
    • Patrick von Platen's avatar
      [Assistant Generation] Improve Encoder Decoder (#26701) · da69de17
      Patrick von Platen authored
      * [Assistant Generation] Improve enc dec
      
      * save more
      
      * Fix logit processor checks
      
      * Clean
      
      * make style
      
      * fix deprecation
      
      * fix generation test
      
      * Apply suggestions from code review
      
      * fix biogpt
      
      * make style
      da69de17
    • Yih-Dar's avatar
      `Copied from` for test files (#26713) · 5334796d
      Yih-Dar authored
      
      
      * copied statement for test files
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      5334796d
    • Billy Bradley's avatar
      In assisted decoding, pass model_kwargs to model's forward call (fix... · dcc49d8a
      Billy Bradley authored
      In assisted decoding, pass model_kwargs to model's forward call (fix prepare_input_for_generation in all models) (#25242)
      
      * In assisted decoding, pass model_kwargs to model's forward call
      
      Previously, assisted decoding would ignore any additional kwargs
      that it doesn't explicitly handle. This was inconsistent with other
      generation methods, which pass the model_kwargs through
      prepare_inputs_for_generation and forward the returned dict to the
      model's forward call.
      
      The prepare_inputs_for_generation method needs to be amended in all
      models, as previously it only kept the last input ID when a past_key_values
      was passed.
      
      * Improve variable names in _extend_attention_mask
      
      * Refactor extending token_type_ids into a function
      
      * Replace deepcopy with copy to optimize performance
      
      * Update new persimmon model with llama changes for assisted generation
      
      * Update new mistral model for assisted generation with prepare_inputs_for_generation
      
      * Update position_ids creation in falcon prepare_inputs_for_generation to support assisted generation
      dcc49d8a
    • Thien Tran's avatar
      Make Whisper Encoder's sinusoidal PE non-trainable by default (#26032) · 1e3c9dda
      Thien Tran authored
      
      
      * set encoder's PE as non-trainable
      
      * freeze flax
      
      * init sinusoids
      
      * add test for non-trainable embed positions
      
      * simplify TF encoder embed_pos
      
      * revert tf
      
      * clean up
      
      * add sinusoidal init for jax
      
      * make consistent sinusoidal function
      
      * fix dtype
      
      * add default dtype
      
      * use numpy for sinusoids. fix jax
      
      * add sinusoid init for TF
      
      * fix
      
      * use custom embedding
      
      * use specialized init for each impl
      
      * fix sinusoids init. add test for pytorch
      
      * fix TF dtype
      
      * simplify sinusoid init for flax and tf
      
      * add tests for TF
      
      * change default dtype to float32
      
      * add sinusoid test for flax
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * move sinusoidal init to _init_weights
      
      ---------
      Co-authored-by: default avatarsanchit-gandhi <sanchit@huggingface.co>
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      1e3c9dda
  14. 09 Oct, 2023 1 commit