1. 27 Nov, 2023 2 commits
    • jiqing-feng's avatar
      fix assisted decoding assistant model inputs (#27503) · 1d7f406e
      jiqing-feng authored
      * fix assisted decoding attention_cat
      
      * fix attention_mask for assisted decoding
      
      * fix attention_mask len
      
      * fix attn len
      
      * Use a more clean way to prepare assistant models inputs
      
      * fix param meaning
      
      * fix param name
      
      * fix assistant model inputs
      
      * update token type ids
      
      * fix assistant kwargs copy
      
      * add encoder-decoder tests of assisted decoding
      
      * check if assistant kwargs contains updated keys
      
      * revert test
      
      * fix whisper tests
      
      * fix assistant kwargs
      
      * revert whisper test
      
      * delete _extend funcs
      1d7f406e
    • Yanan Xie's avatar
      Fix mistral generate for long prompt / response (#27548) · b09912c8
      Yanan Xie authored
      * Fix mistral generate for long prompt / response
      
      * Add unit test
      
      * fix linter
      
      * fix linter
      
      * fix test
      
      * add assisted generation test for mistral and load the model in 4 bit + fa2
      b09912c8
  2. 24 Nov, 2023 3 commits
  3. 23 Nov, 2023 1 commit
  4. 22 Nov, 2023 2 commits
    • dg845's avatar
      Add UnivNet Vocoder Model for Tortoise TTS Diffusers Integration (#24799) · 7f6a804d
      dg845 authored
      * initial commit
      
      * Add inital testing files and modify __init__ files to add UnivNet imports.
      
      * Fix some bugs
      
      * Add checkpoint conversion script and add references to transformers pre-trained model.
      
      * Add UnivNet entries for auto.
      
      * Add initial docs for UnivNet.
      
      * Handle input and output shapes in UnivNetGan.forward and add initial docstrings.
      
      * Write tests and make them pass.
      
      * Write docs.
      
      * Add UnivNet doc to _toctree.yml and improve docs.
      
      * fix typo
      
      * make fixup
      
      * make fix-copies
      
      * Add upsample_rates parameter to config and improve config documentation.
      
      * make fixup
      
      * make fix-copies
      
      * Remove unused upsample_rates config parameter.
      
      * apply suggestions from review
      
      * make style
      
      * Verify and add reason for skipped tests inherited from ModelTesterMixin.
      
      * Add initial UnivNetGan integration tests
      
      * make style
      
      * Remove noise_length input to UnivNetGan and improve integration tests.
      
      * Fix bug and make style
      
      * Make UnivNet integration tests pass
      
      * Add initial code for UnivNetFeatureExtractor.
      
      * make style
      
      * Add initial tests for UnivNetFeatureExtractor.
      
      * make style
      
      * Properly initialize weights for UnivNetGan
      
      * Get feature extractor fast tests passing
      
      * make style
      
      * Get feature extractor integration tests passing
      
      * Get UnivNet integration tests passing
      
      * make style
      
      * Add UnivNetGan usage example
      
      * make style and use feature extractor from hub in integration tests
      
      * Update tips in docs
      
      * apply suggestions from review
      
      * make style
      
      * Calculate padding directly instead of using get_padding methods.
      
      * Update UnivNetFeatureExtractor.to_dict to be UnivNet-specific.
      
      * Update feature extractor to support using model(**inputs) and add the ability to generate noise and pad the end of the spectrogram in __call__.
      
      * Perform padding before generating noise to ensure the shapes are correct.
      
      * Rename UnivNetGan.forward's noise_waveform argument to noise_sequence.
      
      * make style
      
      * Add tests to test generating noise and padding the end for UnivNetFeatureExtractor.__call__.
      
      * Add tests for checking batched vs unbatched inputs for UnivNet feature extractor and model.
      
      * Add expected mean and stddev checks to the integration tests and make them pass.
      
      * make style
      
      * Make it possible to use model(**inputs), where inputs is the output of the feature extractor.
      
      * fix typo in UnivNetGanConfig example
      
      * Calculate spectrogram_zero from other config values.
      
      * apply suggestions from review
      
      * make style
      
      * Refactor UnivNet conversion script to use load_state_dict (following persimmon).
      
      * Rename UnivNetFeatureExtractor to UnivNetGanFeatureExtractor.
      
      * make style
      
      * Switch to using torch.tensor and torch.testing.assert_close for testing expected values/slices.
      
      * make style
      
      * Use config in UnivNetGan modeling blocks.
      
      * make style
      
      * Rename the spectrogram argument of UnivNetGan.forward to input_features, following Whisper.
      
      * make style
      
      * Improving padding documentation.
      
      * Add UnivNet usage example to the docs.
      
      * apply suggestions from review
      
      * Move dynamic_range_compression computation into the mel_spectrogram method of the feature extractor.
      
      * Improve UnivNetGan.forward return docstring.
      
      * Update table in docs/source/en/index.md.
      
      * make fix-copies
      
      * Rename UnivNet components to have pattern UnivNet*.
      
      * make style
      
      * make fix-copies
      
      * Update docs
      
      * make style
      
      * Increase tolerance on flaky unbatched integration test.
      
      * Remove torch.no_grad decorators from UnivNet integration tests to try to avoid flax/Tensorflow test errors.
      
      * Add padding_mask argument to UnivNetModel.forward and add batch_decode feature extractor method to remove padding.
      
      * Update documentation and clean up padding code.
      
      * make style
      
      * make style
      
      * Remove torch dependency from UnivNetFeatureExtractor.
      
      * make style
      
      * Fix UnivNetModel usage example
      
      * Clean up feature extractor code/docstrings.
      
      * apply suggestions from review
      
      * make style
      
      * Add comments for tests skipped via ModelTesterMixin flags.
      
      * Add comment for model parallel tests skipped via the test_model_parallel ModelTesterMixin flag.
      
      * Add # Copied from statements to copied UnivNetFeatureExtractionTest tests.
      
      * Simplify UnivNetFeatureExtractorTest.test_batch_decode.
      
      * Add support for unbatched padding_masks in UnivNetModel.forward.
      
      * Refactor unbatched padding_mask support.
      
      * make style
      7f6a804d
    • Patrick von Platen's avatar
      [Whisper] Add sequential longform decoding (#27492) · 4151fbb4
      Patrick von Platen authored
      * [Whisper] Add seq gen
      
      * [Whisper] Add seq gen
      
      * more debug
      
      * Fix whisper logit processor
      
      * Improve whisper code further
      
      * Fix more
      
      * more debug
      
      * more debug
      
      * Improve further
      
      * Add tests
      
      * Prep for batch size > 1
      
      * Get batch_size>1 working
      
      * Correct more
      
      * Add extensive tests
      
      * more debug
      
      * more debug
      
      * more debug
      
      * add more tests
      
      * more debug
      
      * Apply suggestions from code review
      
      * more debug
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * Add more examples
      
      * add comments to explain the code better
      
      * fix more
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      
      * correct
      
      * correct
      
      * finalize
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      4151fbb4
  5. 21 Nov, 2023 6 commits
    • fxmarty's avatar
      Explicitely specify `use_cache=True` in Flash Attention tests (#27635) · 7f043738
      fxmarty authored
      explicit use_cache=True
      7f043738
    • jiqing-feng's avatar
      TVP model (#25856) · c770600f
      jiqing-feng authored
      * tvp model for video grounding
      
      add tokenizer auto
      
      fix param in TVPProcessor
      
      add docs
      
      clear comments and enable different torch dtype
      
      add image processor test and model test and fix code style
      
      * fix conflict
      
      * fix model doc
      
      * fix image processing tests
      
      * fix tvp tests
      
      * remove torch in processor
      
      * fix grammar error
      
      * add more details on tvp.md
      
      * fix model arch for loss, grammar, and processor
      
      * add docstring and do not regard TvpTransformer, TvpVisionModel as individual model
      
      * use pad_image
      
      * update copyright
      
      * control first downsample stride
      
      * reduce first only works for ResNetBottleNeckLayer
      
      * fix param name
      
      * fix style
      
      * add testing
      
      * fix style
      
      * rm init_weight
      
      * fix style
      
      * add post init
      
      * fix comments
      
      * do not test TvpTransformer
      
      * fix warning
      
      * fix style
      
      * fix example
      
      * fix config map
      
      * add link in config
      
      * fix comments
      
      * fix style
      
      * rm useless param
      
      * change attention
      
      * change test
      
      * add notes
      
      * fix comments
      
      * fix tvp
      
      * import checkpointing
      
      * fix gradient checkpointing
      
      * Use a more accurate example in readme
      
      * update
      
      * fix copy
      
      * fix style
      
      * update readme
      
      * delete print
      
      * remove tvp test_forward_signature
      
      * remove TvpTransformer
      
      * fix test init model
      
      * merge main and make style
      
      * fix tests and others
      
      * fix image processor
      
      * fix style and model_input_names
      
      * fix tests
      c770600f
    • amyeroberts's avatar
      Fix tracing dinov2 (#27561) · 0145c682
      amyeroberts authored
      * Enable tracing with DINOv2 model
      
      * ABC
      
      * Add note to model doc
      0145c682
    • fxmarty's avatar
      Fix flash attention bugs with Mistral and Falcon (#27625) · 82cc0a79
      fxmarty authored
      * fix various bugs with flash attention
      
      * bump
      
      * fix test
      
      * fix mistral
      
      * use skiptest instead of return that may be misleading
      
      * fix on review
      82cc0a79
    • Leo Tronchon's avatar
      Idefics: Fix information leak with cross attention gate in modeling (#26839) · 851a4f70
      Leo Tronchon authored
      
      
      * fix image_attention gate in idefics modeling
      
      * update comment
      
      * cleaner gating
      
      * fix gate condition
      
      * create attention gate once
      
      * update comment
      
      * update doc of cross-attention forward
      
      * improve comment
      
      * bring back no_images
      
      * pass cross_attention_gate similarly  to no_images gate
      
      * add information on gate shape
      
      * fix no_images placement
      
      * make tests for gate
      
      * take off no_images logic
      
      * update test based on comments
      
      * raise value error if cross_attention_gate is None
      
      * send cross_attention_gate to device
      
      * Revert "send cross_attention_gate to device"
      
      This reverts commit 054f84228405bfa2e75fecc502f6a96dc83cdc0b.
      
      * send cross_attention_gate to device
      
      * fix device in test + nit
      
      * fill hidden_states with zeros instead of multiplying with the gate
      
      * style
      
      * Update src/transformers/models/idefics/modeling_idefics.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/idefics/modeling_idefics.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      851a4f70
    • Dave Berenbaum's avatar
      dvclive callback: warn instead of fail when logging non-scalars (#27608) · 8eb9e29d
      Dave Berenbaum authored
      * dvclive callback: warn instead of fail when logging non-scalars
      
      * tests: log lr as scalar
      8eb9e29d
  6. 20 Nov, 2023 2 commits
  7. 17 Nov, 2023 3 commits
  8. 16 Nov, 2023 4 commits
    • Joao Gante's avatar
      12b50c61
    • Arthur's avatar
      [`Styling`] stylify using ruff (#27144) · 651408a0
      Arthur authored
      
      
      * try to stylify using ruff
      
      * might need to remove these changes?
      
      * use ruf format andruff check
      
      * use isinstance instead of type comparision
      
      * use # fmt: skip
      
      * use # fmt: skip
      
      * nits
      
      * soem styling changes
      
      * update ci job
      
      * nits isinstance
      
      * more files update
      
      * nits
      
      * more nits
      
      * small nits
      
      * check and format
      
      * revert wrong changes
      
      * actually use formatter instead of checker
      
      * nits
      
      * well docbuilder is overwriting this commit
      
      * revert notebook changes
      
      * try to nuke docbuilder
      
      * style
      
      * fix feature exrtaction test
      
      * remve `indent-width = 4`
      
      * fixup
      
      * more nits
      
      * update the ruff version that we use
      
      * style
      
      * nuke docbuilder styling
      
      * leve the print for detected changes
      
      * nits
      
      * Remove file I/O
      Co-authored-by: default avatarcharliermarsh <charlie.r.marsh@gmail.com>
      
      * style
      
      * nits
      
      * revert notebook changes
      
      * Add # fmt skip when possible
      
      * Add # fmt skip when possible
      
      * Fix
      
      * More `  # fmt: skip` usage
      
      * More `  # fmt: skip` usage
      
      * More `  # fmt: skip` usage
      
      * NIts
      
      * more fixes
      
      * fix tapas
      
      * Another way to skip
      
      * Recommended way
      
      * Fix two more fiels
      
      * Remove asynch
      Remove asynch
      
      ---------
      Co-authored-by: default avatarcharliermarsh <charlie.r.marsh@gmail.com>
      651408a0
    • Lucain's avatar
      Set `usedforsecurity=False` in hashlib methods (FIPS compliance) (#27483) · fd65aa98
      Lucain authored
      * Set usedforsecurity=False in hashlib methods (FIPS compliance)
      
      * trigger ci
      
      * tokenizers version
      
      * deps
      
      * bump hfh version
      
      * let's try this
      fd65aa98
    • Patrick von Platen's avatar
      Revert "add attention_mask and position_ids in assisted model" (#27523) · 5603fad2
      Patrick von Platen authored
      * Revert "add attention_mask and position_ids in assisted model (#26892)"
      
      This reverts commit 184f60dc.
      
      * more debug
      5603fad2
  9. 15 Nov, 2023 6 commits
  10. 14 Nov, 2023 6 commits
  11. 13 Nov, 2023 5 commits
    • Gift Sinthong's avatar
      [time series] Add PatchTST (#25927) · 2ac5b932
      Gift Sinthong authored
      
      
      * Initial commit of PatchTST model classes
      Co-authored-by: default avatarPhanwadee Sinthong <phsinthong@gmail.com>
      Co-authored-by: default avatarNam Nguyen <namctin@gmail.com>
      Co-authored-by: default avatarVijay Ekambaram <vijaykr.e@gmail.com>
      Co-authored-by: default avatarNgoc Diep Do <55230119+diepi@users.noreply.github.com>
      Co-authored-by: default avatarWesley Gifford <79663411+wgifford@users.noreply.github.com>
      
      * Add PatchTSTForPretraining
      
      * update to include classification
      Co-authored-by: default avatarPhanwadee Sinthong <phsinthong@gmail.com>
      Co-authored-by: default avatarNam Nguyen <namctin@gmail.com>
      Co-authored-by: default avatarVijay Ekambaram <vijaykr.e@gmail.com>
      Co-authored-by: default avatarNgoc Diep Do <55230119+diepi@users.noreply.github.com>
      Co-authored-by: default avatarWesley Gifford <79663411+wgifford@users.noreply.github.com>
      
      * clean up auto files
      
      * Add PatchTSTForPrediction
      
      * Fix relative import
      
      * Replace original PatchTSTEncoder with ChannelAttentionPatchTSTEncoder
      
      * temporary adding absolute path + add PatchTSTForForecasting class
      
      * Update base PatchTSTModel + Unittest
      
      * Update ForecastHead to use the config class
      
      * edit cv_random_masking, add mask to model output
      
      * Update configuration_patchtst.py
      
      * add masked_loss to the pretraining
      
      * add PatchEmbeddings
      
      * Update configuration_patchtst.py
      
      * edit loss which considers mask in the pretraining
      
      * remove patch_last option
      
      * Add commits from internal repo
      
      * Update ForecastHead
      
      * Add model weight initilization + unittest
      
      * Update PatchTST unittest to use local import
      
      * PatchTST integration tests for pretraining and prediction
      
      * Added PatchTSTForRegression + update unittest to include label generation
      
      * Revert unrelated model test file
      
      * Combine similar output classes
      
      * update PredictionHead
      
      * Update configuration_patchtst.py
      
      * Add Revin
      
      * small edit to PatchTSTModelOutputWithNoAttention
      
      * Update modeling_patchtst.py
      
      * Updating integration test for forecasting
      
      * Fix unittest after class structure changed
      
      * docstring updates
      
      * change input_size to num_input_channels
      
      * more formatting
      
      * Remove some unused params
      
      * Add a comment for pretrained models
      
      * add channel_attention option
      
      add channel_attention option and remove unused positional encoders.
      
      * Update PatchTST models to use HF's MultiHeadAttention module
      
      * Update paper + github urls
      
      * Fix hidden_state return value
      
      * Update integration test to use PatchTSTForForecasting
      
      * Adding dataclass decorator for model output classes
      
      * Run fixup script
      
      * Rename model repos for integration test
      
      * edit argument explanation
      
      * change individual option to shared_projection
      
      * style
      
      * Rename integration test + import cleanup
      
      * Fix outpu_hidden_states return value
      
      * removed unused mode
      
      * added std, mean and nops scaler
      
      * add initial distributional loss for predition
      
      * fix typo in docs
      
      * add generate function
      
      * formatting
      
      * add num_parallel_samples
      
      * Fix a typo
      
      * copy weighted_average function, edit PredictionHead
      
      * edit PredictionHead
      
      * add distribution head to forecasting
      
      * formatting
      
      * Add generate function for forecasting
      
      * Add generate function to prediction task
      
      * formatting
      
      * use argsort
      
      * add past_observed_mask ordering
      
      * fix arguments
      
      * docs
      
      * add back test_model_outputs_equivalence test
      
      * formatting
      
      * cleanup
      
      * formatting
      
      * use ACT2CLS
      
      * formatting
      
      * fix add_start_docstrings decorator
      
      * add distribution head and generate function to regression task
      
      add distribution head and generate function to regression task. Also made add PatchTSTForForecastingOutput,  PatchTSTForRegressionOutput.
      
      * add distribution head and generate function to regression task
      
      add distribution head and generate function to regression task. Also made add PatchTSTForForecastingOutput,  PatchTSTForRegressionOutput.
      
      * fix typos
      
      * add forecast_masking
      
      * fixed tests
      
      * use set_seed
      
      * fix doc test
      
      * formatting
      
      * Update docs/source/en/model_doc/patchtst.md
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * better var names
      
      * rename PatchTSTTranspose
      
      * fix argument names and docs string
      
      * remove compute_num_patches and unused class
      
      * remove assert
      
      * renamed to PatchTSTMasking
      
      * use num_labels for classification
      
      * use num_labels
      
      * use default num_labels from super class
      
      * move model_type after docstring
      
      * renamed PatchTSTForMaskPretraining
      
      * bs -> batch_size
      
      * more review fixes
      
      * use hidden_state
      
      * rename encoder layer and block class
      
      * remove commented seed_number
      
      * edit docstring
      
      * Add docstring
      
      * formatting
      
      * use past_observed_mask
      
      * doc suggestion
      
      * make fix-copies
      
      * use Args:
      
      * add docstring
      
      * add docstring
      
      * change some variable names and add PatchTST before some class names
      
      * formatting
      
      * fix argument types
      
      * fix tests
      
      * change x variable to patch_input
      
      * format
      
      * formatting
      
      * fix-copies
      
      * Update tests/models/patchtst/test_modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * move loss to forward
      
      * Update src/transformers/models/patchtst/modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update src/transformers/models/patchtst/modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update src/transformers/models/patchtst/modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update src/transformers/models/patchtst/modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update src/transformers/models/patchtst/modeling_patchtst.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * formatting
      
      * fix a bug when pre_norm is set to True
      
      * output_hidden_states is set to False as default
      
      * set pre_norm=True as default
      
      * format docstring
      
      * format
      
      * output_hidden_states is None by default
      
      * add missing docs
      
      * better var names
      
      * docstring: remove default to False in output_hidden_states
      
      * change labels name to target_values in regression task
      
      * format
      
      * fix tests
      
      * change to forecast_mask_ratios and random_mask_ratio
      
      * change mask names
      
      * change future_values to target_values param in the prediction class
      
      * remove nn.Sequential and make PatchTSTBatchNorm class
      
      * black
      
      * fix argument name for prediction
      
      * add output_attentions option
      
      * add output_attentions to PatchTSTEncoder
      
      * formatting
      
      * Add attention output option to all classes
      
      * Remove PatchTSTEncoderBlock
      
      * create PatchTSTEmbedding class
      
      * use config in PatchTSTPatchify
      
      * Use config in PatchTSTMasking class
      
      * add channel_attn_weights
      
      * Add PatchTSTScaler class
      
      * add output_attentions arg to test function
      
      * format
      
      * Update doc with image patchtst.md
      
      * fix-copies
      
      * rename Forecast <-> Prediction
      
      * change name of a few parameters to match with PatchTSMixer.
      
      * Remove *ForForecasting class to match with other time series models.
      
      * make style
      
      * Remove PatchTSTForForecasting in the test
      
      * remove PatchTSTForForecastingOutput class
      
      * change test_forecast_head to test_prediction_head
      
      * style
      
      * fix docs
      
      * fix tests
      
      * change num_labels to num_targets
      
      * Remove PatchTSTTranspose
      
      * remove arguments in PatchTSTMeanScaler
      
      * remove arguments in PatchTSTStdScaler
      
      * add config as an argument to all the scaler classes
      
      * reformat
      
      * Add norm_eps for batchnorm and layernorm
      
      * reformat.
      
      * reformat
      
      * edit docstring
      
      * update docstring
      
      * change variable name pooling to pooling_type
      
      * fix output_hidden_states as tuple
      
      * fix bug when calling PatchTSTBatchNorm
      
      * change stride to patch_stride
      
      * create PatchTSTPositionalEncoding class and restructure the PatchTSTEncoder
      
      * formatting
      
      * initialize scalers with configs
      
      * edit output_hidden_states
      
      * style
      
      * fix forecast_mask_patches doc string
      
      ---------
      Co-authored-by: default avatarGift Sinthong <gift.sinthong@ibm.com>
      Co-authored-by: default avatarNam Nguyen <namctin@gmail.com>
      Co-authored-by: default avatarVijay Ekambaram <vijaykr.e@gmail.com>
      Co-authored-by: default avatarNgoc Diep Do <55230119+diepi@users.noreply.github.com>
      Co-authored-by: default avatarWesley Gifford <79663411+wgifford@users.noreply.github.com>
      Co-authored-by: default avatarWesley M. Gifford <wmgifford@us.ibm.com>
      Co-authored-by: default avatarnnguyen <nnguyen@us.ibm.com>
      Co-authored-by: default avatarNgoc Diep Do <diiepy@gmail.com>
      Co-authored-by: default avatarKashif Rasul <kashif.rasul@gmail.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      2ac5b932
    • Younes Belkada's avatar
      [`AWQ` ] Addresses TODO for awq tests (#27467) · 7b139023
      Younes Belkada authored
      addresses todo for awq tests
      7b139023
    • NielsRogge's avatar
      Add DINOv2 depth estimation (#26092) · 2422c38d
      NielsRogge authored
      
      
      * First draft
      
      * Fix style
      
      * More improvements
      
      * Fix tests
      
      * Fix tests
      
      * Convert checkpoint
      
      * Improve DPTImageProcessor
      
      * Remove scripts, improve conversion script
      
      * Remove print statements
      
      * Fix test
      
      * Improve docstring
      
      * More improvements
      
      * Fix style
      
      * Fix image processor
      
      * Add tests
      
      * Address comments
      
      * Address comments
      
      * Make bias backwards compatible
      
      * Address comment
      
      * Address comment
      
      * Address comment
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Address comments
      
      * Add flag
      
      * Add tests
      
      * Make tests smaller
      
      * Use regular BackboneOutput
      
      * Fix all tests
      
      * Update test
      
      * Convert more checkpoints
      
      * Convert giant checkpoints, add integration test
      
      * Rename size_divisibility to size_divisor
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      2422c38d
    • Lysandre Debut's avatar
      Fix `from_pt` flag when loading with `safetensors` (#27394) · 68ae3be7
      Lysandre Debut authored
      * Fix
      
      * Tests
      
      * Fix
      68ae3be7
    • Lysandre Debut's avatar
      Default to msgpack for safetensors (#27460) · 9dc8fe1b
      Lysandre Debut authored
      
      
      * Default to msgpack for safetensors
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      9dc8fe1b