1. 01 Mar, 2023 2 commits
  2. 22 Feb, 2023 1 commit
  3. 21 Feb, 2023 1 commit
  4. 20 Feb, 2023 2 commits
    • tanreinama's avatar
      add GPTSAN model (reopen) (#21291) · f56174ac
      tanreinama authored
      * add GPTSAN-Japanese
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN (update for review)
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * fix typo in comment text
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * add GPTSAN
      
      * fix document and comments
      
      * fix class name GPTSAN->GPTSan
      
      * fix import and test for tokenizer
      f56174ac
    • Andy Ehrenberg's avatar
      add flax whisper implementation (#20479) · 2840272c
      Andy Ehrenberg authored
      
      
      * add flax whisper implementation
      
      * rever change to setup
      
      * remove unused imports
      
      * revert generation changes
      
      * flax whisper docs
      
      * docs
      
      * import order
      
      * import sorting
      
      * isort
      
      * add dummy objects
      
      * doc formatting
      
      * formatting
      
      * remove trailing whitespaces
      
      * fix flax whisper docs
      
      * add generation logic to unlock flax whisper
      
      * remove scans
      
      * give credits to Flax Bart implementation
      
      * remove unused imports
      
      * add license
      
      * remove assert
      
      * more credits to Bart
      
      * fix style
      
      * formatting
      
      * support left padding
      
      * add flax whisper generation test
      
      * remove copied from comments whenever not a full copy
      
      * fix docstrings for logits processors
      
      * revert change to FlaxForceTokensLogitsProcessor
      
      * revert doc changes
      
      * improve generation docs
      
      * reorganize
      
      * formatting
      
      * cleanup docs
      
      * add tests
      
      * handle empty list case
      
      * fix forced decoder ids in flax tests
      
      * add flax whisper to inits
      
      * upate dummy objects
      
      * docs for FlaxAutoModelForSpeechSeq2Seq
      
      * fix decoder_position_ids computation in pretrained model decode/__call__ fns
      
      * add Copied from statements as necessary
      
      * compute position_ids only in __call__ and decode methods of pretrained model subclasses
      
      * improve readabilityof compute positional embeddings
      
      * check dimensionality of input_features instead of hidden_states
      
      * copied from statement for init_cache
      
      * formatting
      
      * fix copies
      
      * fix copies
      
      * pass attention mask to encoder layers
      
      * fix decoder module outputs
      
      * set dtype
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * smaller flax model for whisper test
      
      * Update src/transformers/generation/flax_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update tests/models/whisper/test_modeling_flax_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * cleanup
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * bias cleanup
      
      * doc fix
      
      * align style for force tokens processor
      
      * readability
      
      * fix input shape in tests
      
      * revert FlaxGenerationMixin docstring
      
      * formatting
      
      * fix tests
      
      * fix imports
      
      * consistent encoder hidden states
      
      * consistent hidden states
      
      * input shapes
      
      * typo
      
      * partial class trick
      
      * partial class for input shape
      
      * base_class with correct input shape
      
      * partial base classes
      
      * match by name
      
      * set main_input_name
      
      * compare on names
      
      * formatting
      
      * remove unused import
      
      * safer position ids computation
      
      * safer position id computation
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * remove identical inherited tests
      
      * fix prompt ids in tests
      
      * use generation config
      
      * use jnp array
      
      * better var names
      
      * more explicit bias use
      
      * import transformers
      
      * formatting
      
      * test formatting
      
      * remove unused imports
      
      * remove unused imports
      
      * formatting
      
      * isort
      
      * docs
      
      * fix ln orders for encoder hidden states
      
      * whisper unique generation stuff
      
      * flake
      
      * use finfo for attention bias
      
      * docs
      
      * Update src/transformers/generation/flax_utils.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * docs
      
      * add timestamp flax test
      
      * jit for timestamps
      
      * formatting
      
      * clean up timestamps processor
      
      * formatting
      
      * remove if_true
      
      * cleanup
      
      ---------
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      2840272c
  5. 16 Feb, 2023 2 commits
    • Arthur's avatar
      [CLAP] Add CLAP to the library (#21370) · c236a621
      Arthur authored
      
      
      * add model like clip
      
      * update
      
      * text model ok
      
      * clap text works
      
      * some refactor
      
      - `CLAPVision` to `CLAPAudio`
      - refactor kwargs of audio modules
      
      * more refactor
      
      * more refactor
      
      * more refactor
      
      * correct fusion
      
      * more refactor
      
      * new modules
      
      * add basic processor
      
      * fixup
      
      * remove whisper copioed from
      
      * audio logits match
      
      * add doc
      
      * correct filters mel and add maxlength
      
      * style
      
      * few fixes
      
      * forward passes
      
      * fixup
      
      * fixup
      
      * some clean up
      
      * remove mels form the dictionnary
      
      * pad after the repeat
      
      * update padding when dsmaller
      
      * fix padding
      
      * style
      
      * use swin patch merging
      
      * use copied from swin
      
      * processor with any tokenizer
      
      * more copied from
      
      * some clean up
      
      * more refactor
      
      * fix mel when rand_trunc
      
      * style
      
      * remove unused imports
      
      * update processing
      
      * remove image processing tests
      
      * add testing fiel
      
      * fixmodeling issues
      
      * replace with `is_longer`
      
      * clap in serialization
      
      * more refactor
      
      * `make fixup`
      
      * make fixup
      
      * fix feature extractor
      
      * update test feature extractor
      
      * `make fixup`
      
      * clean up config
      
      * more clean up
      
      * more cleanup
      
      * update tests
      
      * refactor tests and inits
      
      * removeCLAP vision config
      
      * remove CLAP from image procssing auto and dummy vision objects
      
      * update inits
      
      * style
      
      * re order classes in modeling clap
      
      * Use roberta tokenizer as the other weights are not open sourced
      
      * small cleaup
      
      * remove tokenization CLAP
      
      * processor tokenizr is roberta
      
      * update feature extraction doc
      
      * remove vclap from model zero shot
      
      * update f_min and f_max to frequency_xx
      
      * some changes
      
      - fix modeling keys
      - add `is_longer` in the forward pass
      - make fixup
      
      * make fixup
      
      * consistent behavior ebtween rand_crop and fusion
      
      * add numpy resize and bilinear and documentation
      
      * move resizing to image utils
      
      * clean feature extraction
      
      * import resize from correct file
      
      * resize in image transforms
      
      * update
      
      * style
      
      * style
      
      * nit
      
      * remove unused arguments form the feature extractor
      
      * style
      
      * few fixes + make fixup
      
      * oops
      
      * fix more tests
      
      * add zero shot audio classification pipeline
      
      * update zeroshot classification pipeline
      
      * fixup
      
      * fix copies
      
      * all CI tests pass
      
      * make fixup + fix docs
      
      * fix docs
      
      * fix docs
      
      * update tests pip;eline
      
      * update zero shot pipeline
      
      * update feature extraction clap
      
      * update tokenization auto
      
      * use nested simplify
      
      * update pipeline tests
      
      * Apply suggestions from code review
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * split in two lines
      
      * fixes
      
      * refactor
      
      * clean up
      
      * add integration tests
      
      * update config docstring
      
      * style
      
      * update processor
      
      * fix processor test
      
      * fix feat extractor tests
      
      * update docs
      
      * Apply suggestions from code review
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * fix readmes
      
      * fix tips
      
      * Update src/transformers/models/auto/configuration_auto.py
      
      * update doc and remove todo -> properly explained
      
      * fix idx and typo
      
      * typoe
      
      * cleanup config
      
      * cleanup tests, styles and doc
      
      * ignore docstyle on image transform
      
      * add conversion script
      
      * remove the `clap` indx in favor of `CLAP`
      
      * update __init
      
      * nits
      
      * Update src/transformers/pipelines/__init__.py
      
      * fix bug
      
      * clarifiy config
      
      * fix copy
      
      * fix init
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * fix model output
      
      * fix comment
      
      * make fixup
      
      * make fixup
      
      * rename to `Clap`
      
      * replace to `Clap`
      
      * replace to `Clap`
      
      * repo consistency
      
      * again repo-consistency
      
      * make fixup
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * add config
      
      * changes
      
      * update conversion
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * remove unused function
      
      * update based on code reviews
      
      * style
      
      * more comments
      
      * cleanup
      
      * clean up
      
      * style
      
      * apply suggestions
      
      * Empty commit
      
      * pipeline will be added in a different PR
      
      * update calls to audio utils functions
      
      * update pipeline init
      
      * style
      
      * style
      
      * styling again
      
      * use pad
      
      * fix repo-consistency
      
      * update utils and add doc for audio utils
      
      * clean up resize by using torch. update inits accordingly
      
      * style
      
      * CLap's  tokenizer is RobertA
      
      * add audio utils to internal toctreee
      
      * update totctree
      
      * style
      
      * update documentation and normalize naming accross audio utils and feature extraction clap
      
      * style
      
      * clean up
      
      * update doc and typos
      
      * fix doctest
      
      * update modelin code, got rid of a lot of reshaping
      
      * style on added doc audio utils
      
      * update modeling clap
      
      * style
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * docstringvariables with CLAP
      
      * rename key
      
      * update modeling CLAP
      
      * update audio utils docstring
      
      * update processing clap
      
      * fix readmes
      
      * fix toctree
      
      * udpate configuration clap
      
      * fix init
      
      * make fixup
      
      * fix
      
      * fix
      
      * update naming
      
      * update
      
      * update checkpoint path
      
      * Apply suggestions from code review
      
      * Major refactoring
      
      * Update src/transformers/models/clap/configuration_clap.py
      
      * merge
      
      ---------
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      c236a621
    • Connor Henderson's avatar
      refactor: Make direct_transformers_import util (#21652) · 0f96c26d
      Connor Henderson authored
      * refactor: Make direct_import util
      
      * edit direct import fn
      
      * add docstring
      
      * make import function specific to transformers only
      
      * edit doc string
      0f96c26d
  6. 15 Feb, 2023 3 commits
  7. 09 Feb, 2023 1 commit
    • NielsRogge's avatar
      Add BLIP-2 (#21441) · d7f1e7c0
      NielsRogge authored
      
      
      * First draft
      
      * More improvements
      
      * More improvements
      
      * Improve conversion script
      
      * Convert all weights
      
      * Make forward pass work
      
      * Make logits match
      
      * More improvements
      
      * More improvements
      
      * More improvements
      
      * Use get_input_embeddings
      
      * Improve some more
      
      * Improve model tests
      
      * Improve model tests
      
      * More improvements
      
      * Fix processor
      
      * Update files
      
      * Update prepare_inputs_for_generation
      
      * More improvements
      
      * Fix copies
      
      * More fixes
      
      * Make fixup
      
      * More improvements
      
      * Add support for seq2seq language model
      
      * More improvements
      
      * Fix test
      
      * More improvements
      
      * Improve conversion script
      
      * Remove some todo's
      
      * Fix README's
      
      * Improve conversion script
      
      * Fix generation
      
      * Fix style and remove Blip2Model
      
      * Fix model outputs
      
      * More improvements
      
      * Set eos_token_id in config
      
      * Fix quality
      
      * Small improvements
      
      * Add processor tests
      
      * More improvements
      
      * Apply suggestions
      
      * Apply suggestions
      
      * Add integration test
      
      * Update image URL
      
      * Add integration test
      
      * Fix model_type
      
      * Update style
      
      * Improve docs
      
      * Add doc tests
      
      * Fix copies
      
      * Remove tests which are passing
      
      * Improve some more
      
      * Add tests for seq2seq language models
      
      * Minor fix
      
      * Convert more checkpoints
      
      * finalize CI
      
      * Fix blip and blip2 processors
      
      * add `accelerate` support for `blip2`
      
      * clean up
      
      * make style
      
      * Update conversion script
      
      * Update conversion script some more
      
      * Update organization
      
      * revert toc file
      
      * add blip-2 to toc file
      
      * Some more improvements
      
      * Fix docstring
      
      * Improve docs
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      d7f1e7c0
  8. 06 Feb, 2023 1 commit
    • Sylvain Gugger's avatar
      Update quality tooling for formatting (#21480) · 6f79d264
      Sylvain Gugger authored
      * Result of black 23.1
      
      * Update target to Python 3.7
      
      * Switch flake8 to ruff
      
      * Configure isort
      
      * Configure isort
      
      * Apply isort with line limit
      
      * Put the right black version
      
      * adapt black in check copies
      
      * Fix copies
      6f79d264
  9. 03 Feb, 2023 1 commit
    • Matthijs Hollemans's avatar
      [WIP] add SpeechT5 model (#18922) · e4bacf66
      Matthijs Hollemans authored
      * make SpeechT5 model by copying Wav2Vec2
      
      * add paper to docs
      
      * whoops added docs in wrong file
      
      * remove SpeechT5Tokenizer + put CTC back in the name
      
      * remove deprecated class
      
      * remove unused docstring
      
      * delete SpeechT5FeatureExtractor, use Wav2Vec2FeatureExtractor instead
      
      * remove classes we don't need right now
      
      * initial stab at speech encoder prenet
      
      * add more speech encoder prenet stuff
      
      * improve SpeechEncoderPrenet
      
      * add encoder (not finished yet)
      
      * add relative position bias to self-attention
      
      * add encoder CTC layers
      
      * fix formatting
      
      * add decoder from BART, doesn't work yet
      
      * make it work with generate loop
      
      * wrap the encoder into a speech encoder class
      
      * wrap the decoder in a text decoder class
      
      * changed my mind
      
      * changed my mind again ;-)
      
      * load decoder weights, make it work
      
      * add weights for text decoder postnet
      
      * add SpeechT5ForCTC model that uses only the encoder
      
      * clean up EncoderLayer and DecoderLayer
      
      * implement _init_weights in SpeechT5PreTrainedModel
      
      * cleanup config + Encoder and Decoder
      
      * add head + cross attention masks
      
      * improve doc comments
      
      * fixup
      
      * more cleanup
      
      * more fixup
      
      * TextDecoderPrenet works now, thanks Kendall
      
      * add CTC loss
      
      * add placeholders for other pre/postnets
      
      * add type annotation
      
      * fix freeze_feature_encoder
      
      * set padding tokens to 0 in decoder attention mask
      
      * encoder attention mask downsampling
      
      * remove features_pen calculation
      
      * disable the padding tokens thing again
      
      * fixup
      
      * more fixup
      
      * code review fixes
      
      * rename encoder/decoder wrapper classes
      
      * allow checkpoints to be loaded into SpeechT5Model
      
      * put encoder into wrapper for CTC model
      
      * clean up conversion script
      
      * add encoder for TTS model
      
      * add speech decoder prenet
      
      * add speech decoder post-net
      
      * attempt to reconstruct the generation loop
      
      * add speech generation loop
      
      * clean up generate_speech
      
      * small tweaks
      
      * fix forward pass
      
      * enable always dropout on speech decoder prenet
      
      * sort declaration
      
      * rename models
      
      * fixup
      
      * fix copies
      
      * more fixup
      
      * make consistency checker happy
      
      * add Seq2SeqSpectrogramOutput class
      
      * doc comments
      
      * quick note about loss and labels
      
      * add HiFi-GAN implementation (from Speech2Speech PR)
      
      * rename file
      
      * add vocoder to TTS model
      
      * improve vocoder
      
      * working on tokenizer
      
      * more better tokenizer
      
      * add CTC tokenizer
      
      * fix decode and batch_code in CTC tokenizer
      
      * fix processor
      
      * two processors and feature extractors
      
      * use SpeechT5WaveformFeatureExtractor instead of Wav2Vec2
      
      * cleanup
      
      * more cleanup
      
      * even more fixup
      
      * notebooks
      
      * fix log-mel spectrograms
      
      * support reduction factor
      
      * fixup
      
      * shift spectrograms to right to create decoder inputs
      
      * return correct labels
      
      * add labels for stop token prediction
      
      * fix doc comments
      
      * fixup
      
      * remove SpeechT5ForPreTraining
      
      * more fixup
      
      * update copyright headers
      
      * add usage examples
      
      * add SpeechT5ProcessorForCTC
      
      * fixup
      
      * push unofficial checkpoints to hub
      
      * initial version of tokenizer unit tests
      
      * add slow test
      
      * fix failing tests
      
      * tests for CTC tokenizer
      
      * finish CTC tokenizer tests
      
      * processor tests
      
      * initial test for feature extractors
      
      * tests for spectrogram feature extractor
      
      * fixup
      
      * more fixup
      
      * add decorators
      
      * require speech for tests
      
      * modeling tests
      
      * more tests for ASR model
      
      * fix imports
      
      * add fake tests for the other models
      
      * fixup
      
      * remove jupyter notebooks
      
      * add missing SpeechT5Model tests
      
      * add missing tests for SpeechT5ForCTC
      
      * add missing tests for SpeechT5ForTextToSpeech
      
      * sort tests by name
      
      * fix Hi-Fi GAN tests
      
      * fixup
      
      * add speech-to-speech model
      
      * refactor duplicate speech generation code
      
      * add processor for SpeechToSpeech model
      
      * add usage example
      
      * add tests for speech-to-speech model
      
      * fixup
      
      * enable gradient checkpointing for SpeechT5FeatureEncoder
      
      * code review
      
      * push_to_hub now takes repo_id
      
      * improve doc comments for HiFi-GAN config
      
      * add missing test
      
      * add integration tests
      
      * make number of layers in speech decoder prenet configurable
      
      * rename variable
      
      * rename variables
      
      * add auto classes for TTS and S2S
      
      * REMOVE CTC!!!
      
      * S2S processor does not support save/load_pretrained
      
      * fixup
      
      * these models are now in an auto mapping
      
      * fix doc links
      
      * rename HiFiGAN to HifiGan, remove separate config file
      
      * REMOVE auto classes
      
      * there can be only one
      
      * fixup
      
      * replace assert
      
      * reformat
      
      * feature extractor can process input and target at same time
      
      * update checkpoint names
      
      * fix commit hash
      e4bacf66
  10. 31 Jan, 2023 1 commit
    • NielsRogge's avatar
      Add DETA (#20983) · 5451f889
      NielsRogge authored
      * First draft
      
      * Add initial draft of conversion script
      
      * Convert all weights
      
      * Fix config
      
      * Add image processor
      
      * Fix DetaImageProcessor
      
      * Run make fix copies
      
      * Remove timm dependency
      
      * Fix dummy objects
      
      * Improve loss function
      
      * Remove conv_encoder attribute
      
      * Update conversion scripts
      
      * Improve postprocessing + docs
      
      * Fix copied from statements
      
      * Add tests
      
      * Improve postprocessing
      
      * Improve postprocessing
      
      * Update READMEs
      
      * More improvements
      
      * Fix rebase
      
      * Add is_torchvision_available
      
      * Add torchvision dependency
      
      * Fix typo and README
      
      * Fix bug
      
      * Add copied from
      
      * Fix style
      
      * Apply suggestions
      
      * Fix thanks to @ydshieh
      
      * Fix another dependency check
      
      * Simplify image processor
      
      * Add scipy
      
      * Improve code
      
      * Add threshold argument
      
      * Fix bug
      
      * Set default threshold
      
      * Improve integration test
      
      * Add another integration test
      
      * Update setup.py
      
      * Address review
      
      * Improve deformable attention function
      
      * Improve copied from
      
      * Use relative imports
      
      * Address review
      
      * Replace assertions
      
      * Address review
      
      * Update dummies
      
      * Remove dummies
      
      * Address comments, update READMEs
      
      * Remove custom kernel code
      
      * Add image processor tests
      
      * Add requires_backends
      
      * Add minor comment
      
      * Update scripts
      
      * Update organization name
      
      * Fix defaults, add doc tests
      
      * Add id2label for object 365
      
      * Fix tests
      
      * Update task guide
      5451f889
  11. 25 Jan, 2023 1 commit
    • Anahita Bhiwandiwalla's avatar
      Add BridgeTower model (#20775) · 3a6e4a22
      Anahita Bhiwandiwalla authored
      
      
      * Commit with BTModel and latest HF code
      
      * Placeholder classes for BTForMLM and BTForITR
      
      * Importing Bert classes from transformers
      
      * Removed objectives.py and dist_utils.py
      
      * Removed swin_transformer.py
      
      * Add image normalization, BridgeTowerForImageAndTextRetrieval
      
      * Add center_crop
      
      * Removing bert tokenizer and LCI references
      
      * Tested config loading from HF transformers hub
      
      * Removed state_dict updates and added path to hub
      
      * Enable center crop
      
      * Getting image_size from config, renaming num_heads and num_layers
      
      * Handling max_length in BridgeTowerProcessor
      
      * Add BridgeTowerForMaskedLM
      
      * Add doc string for BridgeTowerConfig
      
      * Add doc strings for BT config, processor, image processor
      
      * Adding docs, removed swin
      
      * Removed convert_bridgetower_original_to_pytorch.py
      
      * Added doc files for bridgetower, removed is_vision
      
      * Add support attention_mask=None and BridgeTowerModelOutput
      
      * Fix formatting
      
      * Fixes with 'make style', 'make quality', 'make fixup'
      
      * Remove downstream tasks from BridgeTowerModel
      
      * Formatting fixes, add return_dict to BT models
      
      * Clean up after doc_test
      
      * Update BTModelOutput return type, fix todo in doc
      
      * Remove loss_names from init
      
      * implement tests and update tuples returned by models
      
      * Add image reference to bridgetower.mdx
      
      * after make fix-copies, make fixup, make style, make quality, make repo-consistency
      
      * Rename class names with BridgeTower prefix
      
      * Fix for image_size in BTImageProcessor
      
      * implement feature extraction bridgetower tests
      
      * Update image_mean and image_std to be list
      
      * remove unused import
      
      * Removed old comments
      
      * Rework CLIP
      
      * update config in tests followed config update
      
      * Formatting fixes
      
      * Add copied from for BridgeTowerPredictionHeadTransform
      
      * Update bridgetower.mdx
      
      * Update test_feature_extraction_bridgetower.py
      
      * Update bridgetower.mdx
      
      * BridgeTowerForMaskedLM is conditioned on image too
      
      * Add BridgeTowerForMaskedLM
      
      * Fixes
      
      * Call post_init to init weights
      
      * Move freeze layers into method
      
      * Remove BTFeatureExtractor, add BT under multimodal models
      
      * Remove BTFeatureExtractor, add BT under multimodal models
      
      * Code review feedback - cleanup
      
      * Rename variables
      
      * Formatting and style to PR review feedback
      
      * Move center crop after resize
      
      * Use named parameters
      
      * Style fix for modeling_bridgetower.py
      
      * Update docs/source/en/model_doc/bridgetower.mdx
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/bridgetower.mdx
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/bridgetower.mdx
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/models/bridgetower/modeling_bridgetower.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/models/bridgetower/modeling_bridgetower.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/bridgetower.mdx
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      
      * Update src/transformers/models/bridgetower/modeling_bridgetower.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Rename config params, copy BERT classes, clean comments
      
      * Cleanup irtr
      
      * Replace Roberta imports, add BTTextConfig and Model
      
      * Update docs, add visionconfig, consistent arg names
      
      * make fixup
      
      * Comments for forward in BTModel and make fixup
      
      * correct tests
      
      * Remove inconsistent roberta copied from
      
      * Add BridgeTowerTextModel to dummy_pt_objects.py
      
      * Add BridgeTowerTextModel to IGNORE_NON_TESTED
      
      * Update docs for BT Text and Vision Configs
      
      * Treat BridgeTowerTextModel as a private model
      
      * BridgeTowerTextModel as private
      
      * Run make fix-copies
      
      * Adding BTTextModel to PRIVATE_MODELS
      
      * Fix for issue with BT Text and Image configs
      
      * make style changes
      
      * Update README_ja.md
      
      Add 銇嬨倝 to BridgeTower's description
      
      * Clean up config, .mdx and arg names
      
      * Fix init_weights. Remove nn.Sequential
      
      * Formatting and style fixes
      
      * Re-add tie_word_embeddings in config
      
      * update test implementation
      
      * update style
      
      * remove commented out
      
      * fix style
      
      * Update README with abs for BridgeTower
      
      * fix style
      
      * fix mdx file
      
      * Update bridgetower.mdx
      
      * Update img src in bridgetower.mdx
      
      * Update README.md
      
      * Update README.md
      
      * resolve style failed
      
      * Update _toctree.yml
      
      * Update README_ja.md
      
      * Removed mlp_ratio, rename feats, rename BTCLIPModel
      
      * Replace BTCLIP with BTVisionModel,pass in vision_config to BTVisionModel
      
      * Add test_initialization support
      
      * Add support for output_hidden_states
      
      * Update support for output_hidden_states
      
      * Add support for output_attentions
      
      * Add docstring for output_hidden_states
      
      * update tests
      
      * add bridgetowervisionmodel as private model
      
      * rerun the PR test
      
      * Remove model_type, pass configs to classes, renames
      
      * Change self.device to use weight device
      
      * Remove image_size
      
      * Style check fixes
      
      * Add hidden_size and num_hidden_layers to BridgeTowerTransformer
      
      * Update device setting
      
      * cosmetic update
      
      * trigger test again
      
      * trigger tests again
      
      * Update test_modeling_bridgetower.py
      
      trigger tests again
      
      * Update test_modeling_bridgetower.py
      
      * minor update
      
      * re-trigger tests
      
      * Update docs/source/en/model_doc/bridgetower.mdx
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Remove pad, update max_text_len, doc cleanup, pass eps to LayerNorm
      
      * Added copied to, some more review feedback
      
      * make fixup
      
      * Use BridgeTowerVisionEmbeddings
      
      * Code cleanup
      
      * Fixes for BridgeTowerVisionEmbeddings
      
      * style checks
      
      * re-tests
      
      * fix embedding
      
      * address comment on init file
      
      * retrigger tests
      
      * update import prepare_image_inputs
      
      * update test_image_processing_bridgetower.py to reflect test_image_processing_common.py
      
      * retrigger tests
      Co-authored-by: default avatarShaoyen Tseng <shao-yen.tseng@intel.com>
      Co-authored-by: default avatarTiep Le <tiep.le@intel.com>
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarTiep Le <97980157+tileintel@users.noreply.github.com>
      3a6e4a22
  12. 19 Jan, 2023 1 commit
    • Cl茅mentine Fourrier's avatar
      Graphormer model for Graph Classification (#20968) · 87208a05
      Cl茅mentine Fourrier authored
      
      
      * [FT] First commit for graphormer architecture.
      
      The model has no tokenizer, as it uses a collator and preprocessing function for its input management.
      Architecture to be tested against original one.
      The arch might need to be changed to fit the checkpoint, but a revert to the original arch will make the code less nice to read.
      TODO: doc
      
      * [FIX] removed test model
      
      * [FIX] import error
      
      * [FIX] black and flake
      
      * [DOC] added paper refs
      
      * [FIX] [DOC]
      
      * [FIX] black
      
      * [DOC] Updated READMEs
      
      * [FIX] Order of imports + rm Tokenizer calls
      
      * [FIX] Moved assert in class to prevent doc build failure
      
      * [FIX] make fix-copies
      
      * [Doc] update from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * [FIX] Removed Graphormer from Sequence classification model list
      
      * [DOC] Added HF copyright to Cython file
      
      * [DOC] Fixed comments
      
      * [FIX] typos in class doc + removed config classes.
      
      Todo: update doc from paper definitions
      
      * [FIX] Removed dependency to fairseq, and replaced all asserts with Exception management
      
      * [FIX] Homogeneized initialization of weights to pretrained constructor
      
      * [FIX] [CP] Updated multi_hop parameter to get same results as in original implementation
      
      * [DOC] Relevant parameter description in the configuration file
      
      * [DOC] Updated doc and comments in main graphormer file
      
      * [FIX] make style and quality checks
      
      * [DOC] Fix doc format
      
      * [FIX] [WIP] Updated part of the tests, though still a wip
      
      * [FIX] [WIP]
      
      * [FIX] repo consistency
      
      * [FIX] Changed input names for more understandability
      
      * [FIX] [BUG] updated num_classes params for propagation in the model
      
      * simplified collator
      
      * [FIX] Updated tests to follow new naming pattern
      
      * [TESTS] Updated test suite along with model
      
      * |FIX] rm tokenizer import
      
      * [DOC] add link to graphormerdoc
      
      * Changed section in doc from text model to graph model
      
      * Apply suggestions from code review
      
      Spacing, inits
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * [DOC] Explain algos_graphormer functions
      
      * Cython soft import protection
      
      * Rm call to Callable in configuration graphormer
      
      * [FIX] replaced asserts with Exceptions
      
      * Add org to graphormer checkpoints
      
      * Prefixed classes with Graphormer
      
      * Management of init functions
      
      * format
      
      * fixes
      
      * fix length file
      
      * update indent
      
      * relaunching ci
      
      * Errors for missing cython imports
      
      * fix style
      
      * fix style doc
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      87208a05
  13. 16 Jan, 2023 1 commit
    • NielsRogge's avatar
      Add UperNet (#20648) · 4ed89d48
      NielsRogge authored
      
      
      * First draft
      
      * More improvements
      
      * Add convnext backbone
      
      * Add conversion script
      
      * Add more improvements
      
      * Comment out to_dict
      
      * Add to_dict method
      
      * Add default config
      
      * Fix config
      
      * Fix backbone
      
      * Fix backbone some more
      
      * Add docs, auto mapping, tests
      
      * Fix some tests
      
      * Fix more tests
      
      * Fix more tests
      
      * Add conversion script
      
      * Improve conversion script
      
      * Add support for getting reshaped undownsampled hidden states
      
      * Fix forward pass
      
      * Add print statements
      
      * Comment out set_shift_and_window_size
      
      * More improvements
      
      * Correct downsampling layers conversion
      
      * Fix style
      
      * First draft
      
      * Fix conversion script
      
      * Remove config attribute
      
      * Fix more tests
      
      * Update READMEs
      
      * Update ConvNextBackbone
      
      * Fix ConvNext tests
      
      * Align ConvNext with Swin
      
      * Remove files
      
      * Fix index
      
      * Improve docs
      
      * Add output_attentions to model forward
      
      * Add backbone mixin, improve tests
      
      * More improvements
      
      * Update init_weights
      
      * Fix interpolation of logits
      
      * Add UperNetImageProcessor
      
      * Improve image processor
      
      * Fix image processor
      
      * Remove print statements
      
      * Remove script
      
      * Update import
      
      * Add image processor tests
      
      * Remove print statements
      
      * Fix test
      
      * Add integration test
      
      * Add convnext integration test
      
      * Update docstring
      
      * Fix README
      
      * Simplify config
      
      * Apply suggestions
      
      * Improve docs
      
      * Rename class
      
      * Fix test_initialization
      
      * Fix import
      
      * Address review
      
      * Fix confg
      
      * Convert all checkpoints
      
      * Fix default backbone
      
      * Usage same processor as segformer
      
      * Apply suggestions
      
      * Fix init_weights, update conversion scripts
      
      * Improve config
      
      * Use Auto API instead of creating a new image processor
      
      * Fix docs
      
      * Add doctests
      
      * Remove ResNetConfig dependency
      
      * Add always_partition argument
      
      * Fix rebase茅
      
      * Improve docs
      
      * Convert checkpoints
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MBP.localdomain>
      4ed89d48
  14. 04 Jan, 2023 2 commits
  15. 03 Jan, 2023 1 commit
    • NielsRogge's avatar
      Add GIT (GenerativeImage2Text) (#20295) · 9c6f7485
      NielsRogge authored
      
      
      * First draft
      
      * Make model instantiation work
      
      * Fix copied from statement
      
      * More fixes
      
      * Add correct output head
      
      * Improve configuration
      
      * Add conversion script
      
      * Improve conversion script
      
      * Remove token_type_ids
      
      * Fix conversion of projection layers
      
      * Convert all weights
      
      * Use cats image
      
      * Make logits match
      
      * Generate caption on cats image
      
      * Add GITProcessor
      
      * Update conversion script
      
      * Add support for more checkpoints
      
      * Fix conversion script
      
      * Add initial tests
      
      * Remove cross-attention
      
      * More improvements
      
      * Remove is_decoder
      
      * Improve model tests
      
      * Improve tests
      
      * Improve model outputs
      
      * Fix model outputs equivalence
      
      * Fix more tests
      
      * Remove unused code
      
      * Use generate to generate text, no use of cache for now
      
      * Use generate more appropriately
      
      * Fix config tests
      
      * Fix style
      
      * Add support for use_cache
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Fix style
      
      * Fix GIT vision encoder
      
      * Update README
      
      * Fix integration test
      
      * Set bos and eos token ids
      
      * Improve docs
      
      * Improve code
      
      * Add support for provided attention_mask
      
      * Add copied from statement
      
      * Fix gradient checkpointing test
      
      * Set model_input_names
      
      * Investigate model_input_names
      
      * Remove script
      
      * Fix model inputs
      
      * Fix docstring
      
      * Rename GIT to Git
      
      * Support more models
      
      * Add support for textvqa model
      
      * Add video support
      
      * Extend conversion script for video
      
      * Add support for large variant
      
      * Add support for more models
      
      * Fix config archive map
      
      * Update integration test
      
      * Fix README
      
      * Fix CLIP mean and std
      
      * Update processor
      
      * Fix use_cache for video, thanks @gante
      
      * Remove print statements
      
      * Remove assertion
      
      * Add processor tests
      
      * Fix model_input_names
      
      * Use Auto API for processor
      
      * Fix processor tests
      
      * Fix integration test
      
      * Fix pipeline test
      
      * Make tests faster
      
      * Update conversion script
      
      * Update conversion script
      
      * Convert more checkpoints
      
      * Update conversion script
      
      * Fix typo
      
      * Update docstrings
      
      * Improve code snippets
      
      * Fix doc tests
      
      * Add more code examples茅
      
      * Fix doc tests
      
      * Add integration tests
      
      * Fix unused variable
      
      * revert
      
      * Add GIT to Japanese README
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      9c6f7485
  16. 21 Dec, 2022 1 commit
  17. 19 Dec, 2022 1 commit
  18. 16 Dec, 2022 2 commits
    • NielsRogge's avatar
      Add Swin2SR (#19784) · 26dd041c
      NielsRogge authored
      
      
      * First draft
      
      * Add more improvements
      
      * Improve forward pass
      
      * Fix layernorm
      
      * Add upscaler
      
      * More improvements
      
      * More improvements
      
      * More improvements
      
      * Improve conversion script
      
      * Add preprocessing
      
      * Make output match original implementation
      
      * Add additional attributes
      
      * Add support for more models
      
      * Support more models
      
      * Add support for real world sr
      
      * Add initial Swin2SRFeatureExtractor
      
      * Add ImageSuperResolutionOutput
      
      * Make more tests pass
      
      * Use BaseModelOutput
      
      * Fix one more test
      
      * Fix more tests
      
      * Fix another test
      
      * Fix all tests
      
      * Rename to Swin2SRImageProcessor
      
      * Fix toctree
      
      * Fix toctree
      
      * Fix rebase
      
      * Improve Swin2SRImageProcessor
      
      * Remove feature extractor file
      
      * Improve model
      
      * Improve conversion script
      
      * Fix integration test
      
      * Fix init
      
      * Fix conversion script
      
      * Address comments
      
      * Improve upsampler
      
      * Add NearestConvUpsampler
      
      * Improve pixel shuffle upsampler
      
      * Improve auxiliary upsampler
      
      * Improve conversion script
      
      * Rename conv_last to final_convolution
      
      * Fix rebase
      
      * Improve upsample module
      
      * Add padding to image processor
      
      * Fix bug
      
      * Update padding
      
      * Remove print statement and fix integration test
      
      * Improve docs
      
      * Add image processor tests
      
      * Convert all checkpoints, fix tests茅
      
      * Remove print statements
      
      * Fix import
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      26dd041c
    • NielsRogge's avatar
      Add Universal Segmentation class + mapping (#20766) · 7f998612
      NielsRogge authored
      
      
      * Add mapping
      
      * Add mapping to pipeline
      
      * Apply suggestions
      
      * Fix feature extractor tests
      
      * Use ForInstance, add model to universal mapping
      
      * More fixes
      
      * Remove model from deprecated objects茅
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      7f998612
  19. 14 Dec, 2022 1 commit
  20. 13 Dec, 2022 1 commit
  21. 12 Dec, 2022 1 commit
    • Ariel Ekgren's avatar
      Add gpt-sw3 model to transformers (#20209) · 5f94855d
      Ariel Ekgren authored
      
      
      * Add templates for gpt-sw3
      
      * Add templates for gpt-sw3
      
      * Added sentencepiece tokenizer
      
      * intermediate commit with many changes
      
      * fixed conflicts
      
      * Init commit for tokenization port
      
      * Tokenization progress
      
      * Remove fast tokenizer
      
      * Clean up and rename spm.model -> spiece.model
      
      * Remove TF -> PT conversion script template, Clean up Megatron -> PT script
      
      * Optimize encode & decode performance
      
      * added new attention
      
      * added new attention
      
      * attention for gpt-sw3 working
      
      * attention good
      
      * Cache is now working
      
      * fixed attention mask so that it works with causal attention
      
      * fixed badbmm bug for cpu and caching
      
      * updated config with correct parameters
      
      * Refactor and leave optimizations as separate functions to avoid breaking expected functionality
      
      * Fix special tokens mapping for both tokenizers
      
      * cleaning up of code and comments
      
      * HF compatible attention outputs
      
      * Tokenizer now passing tests, add documentation
      
      * Update documentation
      
      * reverted back to base implementation after checking that it is identical to pretrained model
      
      * updated gpt-sw3 config
      
      * updated conversion script
      
      * aligned parameters with gpt-sw3 config
      
      * changed default scale_attn_by_inverse_layer_idx to true
      
      * removed flag from conversion script
      
      * added temporary model path
      
      * reverted back to functioning convert script
      
      * small changes to default config
      
      * updated tests for gpt-sw3
      
      * make style, make quality, minor cleanup
      
      * Change local paths to testing online repository
      
      * Change name: GptSw3 -> GPTSw3
      
      * Remove GPTSw3TokenizerFast references
      
      * Use official model repository and add more model sizes
      
      * Added reference to 6.7b model
      
      * Add GPTSw3DoubleHeadsModel to IGNORE_NON_AUTO_CONFIGURED, like GPT2DoubleHeadsModel
      
      * Remove pointers to non-existing TFGPTSw3
      
      * Add GPTSw3 to docs/_toctree.yml
      
      * Remove TF artifacts from GPTSw3 in __init__ files
      
      * Update README:s with 'make fix-copies'
      
      * Add 20b model to archive list
      
      * Add documentation for GPT-Sw3
      
      * Fix typo in documentation for GPT-Sw3
      
      * Do 'make fix-copies' again after having updated docs
      
      * Fix some typos in docs
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update tests/models/gpt_sw3/test_tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Resolve comments from PR feedback
      
      * Resolve more comments from PR feedback, also set use_cache=True in convert script
      
      * Add '# Copied from' comments for GPTSw3 modeling
      
      * Set 'is_parallelizable = False'
      
      * Remove '# Copied from' where code was modified and add 'with x->y' when appropriate
      
      * Remove parallelize in mdx
      
      * make style, make quality
      
      * Update GPTSw3Config default values and corresponding documentation
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Clean up and protect GPTSw3Tokenizer imports with is_sentencepiece_available
      
      * Make style, make quality
      
      * Add dummy object for GPTSw3Tokenizer via 'make fix-copies'
      
      * make fix-copies
      
      * Remove GPTSw3 modeling classes
      
      * make style, make quality
      
      * Add GPTSw3 auto-mappings for other GPT2 heads
      
      * Update docs/source/en/model_doc/gpt-sw3.mdx
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Remove old TODO-comment
      
      * Add example usage to GPTSw3Tokenizer docstring
      
      * make style, make quality
      
      * Add implementation details and example usage to gpt-sw3.mdx
      Co-authored-by: default avatarJoeyOhman <joeyoh@kth.se>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      5f94855d
  22. 07 Dec, 2022 1 commit
  23. 30 Nov, 2022 1 commit
    • Yang An's avatar
      Add Chinese-CLIP implementation (#20368) · 72176402
      Yang An authored
      
      
      * init chinese-clip model from clip
      
      * init model tests and docs
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * update usecase example in model implementation
      
      * fix codestyle
      
      * fix model_type typo in readme
      
      * add placeholder in doc
      
      * add placeholder in doc
      
      * update the init script
      
      * update usecase
      
      * fix codestyle
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * forward the convert_rgb
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * merge the recent update from clip about model_input_name property
      
      * update the doc
      
      * update the doc
      
      * update the doc
      
      * update the doc
      
      * remove unused imports
      
      * reformat code style
      
      * update the doc
      
      * fix isort style
      
      * bypass a weird failed unit test which is unrelated with my PR
      
      * update the doc
      
      * implement independent vision config class
      
      * implement independent vision model class
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * fix refactor bug
      
      * doc-build restyle
      
      * implement independent text config class
      
      * implement independent text model class
      
      * implement independent text model class
      
      * make style
      
      * make fix-copies
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * update doc
      
      * black and isort
      
      * update doc
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/auto/tokenization_auto.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * modify the model type from chinese-clip to chinese_clip
      
      * format the example comment of ChineseCLIPVisionConfig
      
      * correct the copyright comment
      
      * fix the tokenizer specification
      
      * add copied from for loss function
      
      * remove unused class
      
      * update CHINESE_CLIP_TEXT_INPUTS_DOCSTRING
      
      * update CHINESE_CLIP_INPUTS_DOCSTRING
      
      * update doc
      
      * update doc
      
      * update code comment in config
      
      * update copied from statement
      
      * make style
      
      * rename the doc file
      
      * add copied statement
      
      * remove unused attention_mask, causal_attention_mask in ChineseCLIPVisionEncoder
      
      * remove ChineseCLIPTextPreTrainedModel
      
      * fix bug
      
      * fix bug
      
      * fix bug
      
      * update doc
      
      * make style
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * update ChineseCLIPImageProcessor in image_processing_auto
      
      * fix config_class of chinesecliptextmodel
      
      * fix the test case
      
      * update the docs
      
      * remove the copied from comment for ChineseCLIPTextModel, since it has diverged from BertModel with customed config_class
      
      * update the testcase
      
      * final fix
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      72176402
  24. 28 Nov, 2022 2 commits
  25. 17 Nov, 2022 1 commit
    • NielsRogge's avatar
      Add AutoBackbone + ResNetBackbone (#20229) · 6b217c52
      NielsRogge authored
      
      
      * Add ResNetBackbone
      
      * Define channels and strides as property
      
      * Remove file
      
      * Add test for backbone
      
      * Update BackboneOutput class
      
      * Remove strides property
      
      * Fix docstring
      
      * Add backbones to SHOULD_HAVE_THEIR_OWN_PAGE
      
      * Fix auto mapping name
      
      * Add sanity check for out_features
      
      * Set stage names based on depths
      
      * Update to tuple
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      6b217c52
  26. 15 Nov, 2022 2 commits
    • Suraj Patil's avatar
      [CLIP] allow loading projection layer in vision and text model (#18962) · 7f744338
      Suraj Patil authored
      
      
      * allow loading projection in text and vision model
      
      * begin tests
      
      * finish test for CLIPTextModelTest
      
      * style
      
      * add slow tests
      
      * add new classes for projection heads
      
      * remove with_projection
      
      * add in init
      
      * add in doc
      
      * fix tests
      
      * fix some more tests
      
      * fix copies
      
      * fix docs
      
      * remove leftover from fix-copies
      
      * add the head models in IGNORE_NON_AUTO_CONFIGURED
      
      * fix docstr
      
      * fix tests
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * add docstr for models
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      7f744338
    • Younes Belkada's avatar
      Add Switch transformers (#19323) · 163ac3d3
      Younes Belkada authored
      
      
      * first commit
      
      * add more comments
      
      * add router v1
      
      * clean up
      
      - remove `tf` modeling files
      
      * clean up
      
      - remove `tf` modeling files
      
      * clean up
      
      * v0 routers
      
      * added more router
      
      - Implemented `ExpertsChooseMaskedRouter`
      
      - added tests
      - 2 more routers to implement
      
      * last router
      
      * improved docstring
      
      - completed the docstring in `router.py`
      - added more args in the config
      
      * v0 sparse mlp
      
      * replace wrong naming
      
      * forward pass run
      
      * update MOE layer
      
      * small router update
      
      * fixup
      
      * consistency
      
      * remove scatter router
      
      * remove abstract layer
      
      * update test and model for integration testing
      
      * v1 conversion
      
      * update
      
      * hardcode hack
      
      * all keys match
      
      * add gin conversion, without additional libraries
      
      * update conversion sctipy
      
      * delete router file
      
      * update tests wrt router deletion
      
      * fix router issues
      
      * update expert code
      
      * update, logits match, code needsREFACTORING
      
      * Refactor code
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * add generate tests
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      
      * add support for router loss
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * fix forward error
      
      * refactor a bit
      
      * remove `FlaxSwitchTransformers` modules
      
      * more tests pass
      
      * Update code
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * fixup
      
      * fix tests
      
      * fix doc
      
      * fix doc + tokenization
      
      * fix tokenizer test
      
      * fix test
      
      * fix loss output
      
      * update code for backward pass
      
      * add loss support
      
      * update documentation
      
      * fix documentation, clean tokenizer
      
      * more doc fix, cleanup example_switch
      
      * fix failing test
      
      * fix test
      
      * fix test
      
      * fix loss issue
      
      * move layer
      
      * update doc and fix router capacity usage
      
      * fixup
      
      * add sparse mlp index for documentation on hub
      
      * fixup
      
      * test sparse mix architecture
      
      * Apply suggestions from code review
      
      * Update docs/source/en/model_doc/switch_transformers.mdx
      
      * fixup on update
      
      * fix tests
      
      * fix another test
      
      * attempt fix
      
      * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/switch_transformers/convert_switch_transformers_original_flax_checkpoint_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * try
      
      * all tests pass
      
      * fix jitter noise
      
      * Apply suggestions from code review
      
      * doc tests pass
      
      * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * remove assert
      
      * change config order
      
      * fix readme japanese
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * remove parallelizable tests + add one liners
      
      * remove ONNX config
      
      * fix nits
      
      - add `T5Tokenizer` in auto mapping
      - remove `Switch Transformers` from ONNX supported models
      
      * remove `_get_router`
      
      * remove asserts
      
      * add check in test for `router_dtype`
      
      * add `SwitchTransformersConfig` in `run_pipeline_test`
      
      * Update tests/pipelines/test_pipelines_summarization.py
      
      * add huge model conversion script
      
      * fix slow tests
      
      - add better casting for `Linear8bitLt`
      - remove `torchscript` tests
      
      * add make dir
      
      * style on new script
      
      * fix nits
      
      - doctest
      - remove `_keys_to_ignore_on_load_unexpected`
      
      * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py
      
      * add google as authors
      
      * fix year
      
      * remove last `assert` statements
      
      * standardize vertical spaces
      
      * fix failing import
      
      * fix another failing test
      
      * Remove strange 脿uthorized_keys`
      
      * removing todo and padding that is never used
      Co-authored-by: default avatarArthur Zucker <arthur.zucker@gmail.com>
      Co-authored-by: default avatarybelkada <younes@huggingface.co>
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarArthur Zucker <arthur@huggingface.co>
      163ac3d3
  27. 10 Nov, 2022 1 commit
  28. 08 Nov, 2022 1 commit
    • NielsRogge's avatar
      Add CLIPSeg (#20066) · 25896306
      NielsRogge authored
      
      
      * Add first draft
      
      * Update conversion script
      
      * Improve conversion script
      
      * Improve conversion script some more
      
      * Add conditional embeddings
      
      * Add initial decoder
      
      * Fix activation function of decoder
      
      * Make decoder outputs match original implementation
      
      * Make decoder outputs match original implementation
      
      * Add more copied from statements
      
      * Improve model outputs
      
      * Fix auto tokenizer file
      
      * Fix more tests
      
      * Add test
      
      * Improve README and docs, improve conditional embeddings
      
      * Fix more tests
      
      * Remove print statements
      
      * Remove initial embeddings
      
      * Improve conversion script
      
      * Add interpolation of position embeddings
      
      * Finish addition of interpolation of position embeddings
      
      * Add support for refined checkpoint
      
      * Fix refined checkpoint
      
      * Remove unused parameter
      
      * Improve conversion script
      
      * Add support for training
      
      * Fix conversion script
      
      * Add CLIPSegFeatureExtractor
      
      * Fix processor
      
      * Fix CLIPSegProcessor
      
      * Fix conversion script
      
      * Fix most tests
      
      * Fix equivalence test
      
      * Fix README
      
      * Add model to doc tests
      
      * Use better variable name
      
      * Convert other checkpoint as well
      
      * Update config, add link to paper
      
      * Add docs
      
      * Update organization
      
      * Replace base_model_prefix with clip
      
      * Fix base_model_prefix
      
      * Fix checkpoint of config
      
      * Fix config checkpoint
      
      * Remove file
      
      * Use logits for output
      
      * Fix tests
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      25896306
  29. 01 Nov, 2022 1 commit
    • Matt's avatar
      Add ESMFold (#19977) · 7f9b7b3f
      Matt authored
      
      
      * initial commit
      
      * First draft that gets outputs without crashing!
      
      * Add all the ported openfold dependencies
      
      * testing
      
      * Restructure config files for ESMFold
      
      * Debugging to find output discrepancies
      
      * Mainly style
      
      * Make model runnable without extra deps
      
      * Remove utils and merge them to the modeling file
      
      * Use correct gelu and remove some debug prints
      
      * More cleanup
      
      * Update esm docs
      
      * Update conversion script to support ESMFold properly
      
      * Port some top-level changes from ESMFold repo
      
      * Expand EsmFold docstrings
      
      * Make attention_mask optional (default to all 1s)
      
      * Add inference test for ESMFold
      
      * Use config and not n kwargs
      
      * Add modeling output class
      
      * Remove einops
      
      * Remove chunking in ESM FFN
      
      * Update tests for ESMFold
      
      * Quality
      
      * REpo consistency
      
      * Remove tree dependency from ESMFold
      
      * make fixup
      
      * Add an error in case my structure map function breaks later
      
      * Remove needless code
      
      * Stop auto-casting the LM to float16 so CPU tests pass
      
      * Stop auto-casting the LM to float16 so CPU tests pass
      
      * Final test updates
      
      * Split test file
      
      * Copyright and quality
      
      * Unpin PyTorch to see built doc
      
      * Fix config file to_dict() method
      
      * Add some docstrings to the output
      
      * Skip TF checkpoint tests for ESM until we reupload those
      
      * make fixup
      
      * More docstrings
      
      * Unpin to get even with main
      
      * Flag example to write
      Co-authored-by: default avatarSylvain Gugger <Sylvain.gugger@gmail.com>
      7f9b7b3f
  30. 18 Oct, 2022 1 commit
    • NielsRogge's avatar
      Add table transformer [v2] (#19614) · dd523da5
      NielsRogge authored
      * First draft
      
      * Add conversion script
      
      * Make conversion work
      
      * Upload checkpoints
      
      * Add final fixes
      
      * Revert changes of conditional and deformable detr
      
      * Fix toctree, add and remove copied from
      
      * Use model type
      
      * Improve docs
      
      * Improve code example
      
      * Update copies
      
      * Add copied formt
      
      * Don't update conditional detr
      
      * Don't update deformable detr
      dd523da5
  31. 10 Oct, 2022 1 commit
    • amyeroberts's avatar
      Add TF whisper (#19378) · e3f028f3
      amyeroberts authored
      
      
      * simplify loop
      
      * add featur extractor
      
      * add model
      
      * start conversion
      
      * add dropout
      
      * initial commit of test files
      
      * copnversion for all models
      
      * update processor for correct padding
      
      * update feature extraction
      
      * update integration test logits match
      
      * fmnt: off for the logits
      
      * on the fly mel bank
      
      * small nit
      
      * update test
      
      * update tokenizer
      
      * nit feature extraction
      
      * update
      
      * update tokenizer test
      
      * adds logit processor and update tokenizer to get supress tokens
      
      * style
      
      * clean convert
      
      * revert to original modeling tf utils
      
      * Update
      
      * update
      
      * nit
      
      * clean convert file
      
      * update tests and nits
      
      * quality
      
      * slow generation test
      
      * ffn_dim to allow customization
      
      * update readme
      
      * add to toctreee
      
      * start fixing integration tests
      
      * update tests and code
      
      * fix feature extractor
      
      * fix config tests common
      
      * update code to fix tests
      
      * fix feature exctractor
      
      * nit feature extraction
      
      * update test for new feature extractor
      
      * style
      
      * add absrtact
      
      * large logits wioth custom decoder input ids
      
      * wraap around is otrch available
      
      * fix feature extractor
      
      * correct logits for whisper small.en
      
      * nit
      
      * fix encoder_attentino_mask
      
      * some fixes
      
      * remove unnecessary inputs
      
      * nits
      
      * add normalizer file
      
      * update etst tokenization
      
      * fix attention mask not defined
      
      * fix generate
      
      * remove uncoder attention mask useless
      
      * update test modeling whisper
      
      * update condfig to add second non supress tokens
      
      * nits on feature exrtactor
      
      * nit for test tokenizers
      
      * update etsts
      
      * update tests
      
      * update tokenization test
      
      * fixup
      
      * invalidated hf token. Clean convert openai to whisper
      
      * fix logit tests
      
      * fixup
      
      * Add model to README
      
      * Fix doc tests
      
      * clean merge
      
      * revert toc_tree changes
      
      * remove useless LogitProcessor
      
      * Update whisper .mdx
      
      * update config file doc
      
      * update configuration docstring
      
      * update test tokenization
      
      * update test tokenization
      
      * update tokenization whisper
      Added copied from where needed
      
      * update feature extraction
      
      * nit test name
      
      * style
      
      * quality
      
      * remove get suppress tokens and update non_speech tokens global variables
      
      * Update src/transformers/models/whisper/feature_extraction_whisper.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * clean modeling whisper and test
      Removed the attention mask arguments that are deprecated
      
      * fix large test
      
      * Add multilingual audio test, and translate test
      
      * style
      
      * fix larg multilingual test
      
      * nits
      
      * add copied from for attention layer
      
      * remove attention masks in doc
      
      * add english normalizer
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * update tokenization test
      
      * remove copied from in whisper attention : no bias in k_proj only
      
      * wrap around dependencies in english normalizer
      
      * style
      
      * correct import generation logits
      
      * for now, wrap feature extractor with torch
      
      * remove torch depencies for feature extraction and style
      
      * Update src/transformers/models/whisper/convert_openai_whisper_to_tfms.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * fixup
      
      * nit
      
      * update logitds
      
      * style
      
      * nit
      
      * nits and fix final tests
      
      * add `is_more_itertools_available` to utils
      
      * quality
      
      * add begin supress tokens, supress tokens to generate args and config
      
      * clean supressTokensLogitProcessor in generation logits
      
      * Nit naming
      
      * add supressTokensAtBegin
      
      * udpate tests, supress tokens to None or correct values
      
      * nit and style
      
      * update RAG to fit test and generate_logit
      
      * add copy pasted statment on english normalizer
      
      * add arguments to config_common_kwargs
      
      * Update src/transformers/generation_utils.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/generation_logits_process.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * revert changes based on reviews
      
      * update doc and nits
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * more nits
      
      * last nits
      
      * update test configuration common
      
      * add BART name in decoder attention mask documentation
      
      * Update src/transformers/models/whisper/modeling_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * style
      
      * nit
      
      * nit
      
      * add english.json file to git
      
      * nits on documentation
      
      * nit
      
      * nits
      
      * last styling
      
      * add main toctree file
      
      * remove sentence piece dependency
      
      * clean init file
      
      * fix tokenizer that has no dependencies on sentencepiece
      
      * update whisper init file, nit
      
      * remove english.json file
      
      * add get decoder prompt id
      
      * All weights loading
      
      * Remove hanging pdb
      
      * Fixup and tidy up
      
      * Use same copied from as PT model
      
      * Remove whitespace changes
      
      * Remove torch references
      
      * Tie embeddings
      
      * Remove logits processor input to generate
      
      * Update logit values
      
      * revert changes and add forced logit processor
      
      * nit
      
      * clean normalizer
      
      * remove protected
      
      * Add logit processors and update generation code & tests
      
      * Some tidy up
      
      * Update docstring
      
      * update
      
      * update based on review
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update to reflect changes on the PT model branch
      
      * Tidy up
      
      * Remove extra whitespace
      
      * Fix test - make input ids small enough we can append
      
      * Include upstream changes on main
      
      * PR comments - add batch tests, remove comments & defaults
      
      * Fix model output imports
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/generation_tf_logits_process.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update tests/models/whisper/test_modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update docstring example
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Remove changes to adjust_logits_during_generation function
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Tidy up imports that don't require TF
      
      * Update tests - skip and no more skip
      
      * Update tests/generation/test_generation_tf_logits_process.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Add training flags
      
      * Add (skipped) XLA generation tests
      
      * Add embedding correctness test
      
      * Add constant ids for generation tests
      
      * Make logits finding a bit tidier
      
      * Remove unused args
      
      * xla generation enabled
      
      * Don't skip XLA tests anymore
      
      * Fix tests - add position ids to expected signature and update rag generation
      
      * Undo method reorder
      
      * Remove added whitespace
      
      * Remove copy-paste gradient checkopint ref
      
      * Remove
      
      * Trigger CI - (issue with refs when pulling)
      Co-authored-by: default avatarArthur Zucker <arthur.zucker@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <niels.rogge1@gmail.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      Co-authored-by: default avatarJoao Gante <joao@huggingface.co>
      e3f028f3