1. 23 Jan, 2023 1 commit
  2. 20 Jan, 2023 1 commit
  3. 19 Jan, 2023 2 commits
    • Cl茅mentine Fourrier's avatar
      Graphormer model for Graph Classification (#20968) · 87208a05
      Cl茅mentine Fourrier authored
      
      
      * [FT] First commit for graphormer architecture.
      
      The model has no tokenizer, as it uses a collator and preprocessing function for its input management.
      Architecture to be tested against original one.
      The arch might need to be changed to fit the checkpoint, but a revert to the original arch will make the code less nice to read.
      TODO: doc
      
      * [FIX] removed test model
      
      * [FIX] import error
      
      * [FIX] black and flake
      
      * [DOC] added paper refs
      
      * [FIX] [DOC]
      
      * [FIX] black
      
      * [DOC] Updated READMEs
      
      * [FIX] Order of imports + rm Tokenizer calls
      
      * [FIX] Moved assert in class to prevent doc build failure
      
      * [FIX] make fix-copies
      
      * [Doc] update from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * [FIX] Removed Graphormer from Sequence classification model list
      
      * [DOC] Added HF copyright to Cython file
      
      * [DOC] Fixed comments
      
      * [FIX] typos in class doc + removed config classes.
      
      Todo: update doc from paper definitions
      
      * [FIX] Removed dependency to fairseq, and replaced all asserts with Exception management
      
      * [FIX] Homogeneized initialization of weights to pretrained constructor
      
      * [FIX] [CP] Updated multi_hop parameter to get same results as in original implementation
      
      * [DOC] Relevant parameter description in the configuration file
      
      * [DOC] Updated doc and comments in main graphormer file
      
      * [FIX] make style and quality checks
      
      * [DOC] Fix doc format
      
      * [FIX] [WIP] Updated part of the tests, though still a wip
      
      * [FIX] [WIP]
      
      * [FIX] repo consistency
      
      * [FIX] Changed input names for more understandability
      
      * [FIX] [BUG] updated num_classes params for propagation in the model
      
      * simplified collator
      
      * [FIX] Updated tests to follow new naming pattern
      
      * [TESTS] Updated test suite along with model
      
      * |FIX] rm tokenizer import
      
      * [DOC] add link to graphormerdoc
      
      * Changed section in doc from text model to graph model
      
      * Apply suggestions from code review
      
      Spacing, inits
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * [DOC] Explain algos_graphormer functions
      
      * Cython soft import protection
      
      * Rm call to Callable in configuration graphormer
      
      * [FIX] replaced asserts with Exceptions
      
      * Add org to graphormer checkpoints
      
      * Prefixed classes with Graphormer
      
      * Management of init functions
      
      * format
      
      * fixes
      
      * fix length file
      
      * update indent
      
      * relaunching ci
      
      * Errors for missing cython imports
      
      * fix style
      
      * fix style doc
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      87208a05
    • Jitesh Jain's avatar
      Add OneFormer Model (#20577) · 5b949623
      Jitesh Jain authored
      * Add Oneformer Model
      
      * Add OneFormer Tests
      
      * Add UNIVERSAL_SEGMENTATION_MAPPING
      
      * Fix config
      
      * 馃悰 Fix error encountered while writing tests
      
      * 馃敤 Fix instance segmentation post processing
      
      * Format Files and Add Documentation
      
      * Add Documentation mdx file
      
      * Run make fixup
      
      * Run make fix-copies
      
      * Remove unnecessary code
      
      * Format modeling_oneformer.py
      
      * Add OneFormer to ImageSegmentationPipeline
      
      * Format files
      
      * Add Demo link to Readme
      
      * Fix fomatting errors
      
      * Fix test failures
      
      * Update Table in index.mdx
      
      * Fix version
      
      * Fix style
      
      * Remove OneFormer from TF
      
      * Fix Imports
      
      * Fix dummy objects
      
      * Fix tests
      
      * Add newline
      
      * Remove OneFormerFeatureExtractor
      
      * Remove CUDA Kernels
      
      * Use AutoBackbone for Swin
      
      * Fix description
      
      * Use Image Processor
      
      * Fix copies
      
      * Fix formatting
      
      * Fix import order
      
      * Fix flake8 errors
      
      * Fix doc errors
      
      * Add Hindi Readme entry
      
      * Update supported backbones
      
      * Update supported backbones
      
      * Undo Changes
      
      * Fix type of config
      
      * Fix isort
      
      * Fix auto.mdx
      
      * Fix swin config
      
      * Replace DinatBackbone with AutoBackbone
      
      * Use SwinBackbone
      
      * Use SwinBackbone
      
      * Fix conversion script
      
      * Fix arguments
      
      * Add argument description
      
      * Fix style
      
      * Add OneFormerProcessor
      
      * Fix OneFormerProcessor Tests
      
      * Fix mapping
      
      * Fix imports
      
      * Fix inits
      
      * Fix style
      
      * Fix comment
      
      * Fix docstring
      
      * Move OneFormer to MultiModal
      
      * Fix Copies
      
      * Remove size divisor
      
      * Fix check_repo.py
      
      * Fix copies
      
      * Add Processor for Testing Pipeline
      
      * Fix padding for tokens
      
      * Fix variables
      
      * Fix formatting with correct black version
      
      * Add Image Processor Test
      
      * Apply suggestions
      
      * Revert common modeling
      
      * Add check for task
      
      * Fix conversion script
      
      * Fix initialization order
      
      * Fix tests
      
      * Undo Pipeline Changes
      
      * Fix layers in MLP
      
      * Fix copies
      
      * Update image paths
      
      * Fix copies
      
      * Apply suggestions
      5b949623
  4. 17 Jan, 2023 1 commit
  5. 16 Jan, 2023 2 commits
    • Alara Dirik's avatar
    • NielsRogge's avatar
      Add UperNet (#20648) · 4ed89d48
      NielsRogge authored
      
      
      * First draft
      
      * More improvements
      
      * Add convnext backbone
      
      * Add conversion script
      
      * Add more improvements
      
      * Comment out to_dict
      
      * Add to_dict method
      
      * Add default config
      
      * Fix config
      
      * Fix backbone
      
      * Fix backbone some more
      
      * Add docs, auto mapping, tests
      
      * Fix some tests
      
      * Fix more tests
      
      * Fix more tests
      
      * Add conversion script
      
      * Improve conversion script
      
      * Add support for getting reshaped undownsampled hidden states
      
      * Fix forward pass
      
      * Add print statements
      
      * Comment out set_shift_and_window_size
      
      * More improvements
      
      * Correct downsampling layers conversion
      
      * Fix style
      
      * First draft
      
      * Fix conversion script
      
      * Remove config attribute
      
      * Fix more tests
      
      * Update READMEs
      
      * Update ConvNextBackbone
      
      * Fix ConvNext tests
      
      * Align ConvNext with Swin
      
      * Remove files
      
      * Fix index
      
      * Improve docs
      
      * Add output_attentions to model forward
      
      * Add backbone mixin, improve tests
      
      * More improvements
      
      * Update init_weights
      
      * Fix interpolation of logits
      
      * Add UperNetImageProcessor
      
      * Improve image processor
      
      * Fix image processor
      
      * Remove print statements
      
      * Remove script
      
      * Update import
      
      * Add image processor tests
      
      * Remove print statements
      
      * Fix test
      
      * Add integration test
      
      * Add convnext integration test
      
      * Update docstring
      
      * Fix README
      
      * Simplify config
      
      * Apply suggestions
      
      * Improve docs
      
      * Rename class
      
      * Fix test_initialization
      
      * Fix import
      
      * Address review
      
      * Fix confg
      
      * Convert all checkpoints
      
      * Fix default backbone
      
      * Usage same processor as segformer
      
      * Apply suggestions
      
      * Fix init_weights, update conversion scripts
      
      * Improve config
      
      * Use Auto API instead of creating a new image processor
      
      * Fix docs
      
      * Add doctests
      
      * Remove ResNetConfig dependency
      
      * Add always_partition argument
      
      * Fix rebase茅
      
      * Improve docs
      
      * Convert checkpoints
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MBP.localdomain>
      4ed89d48
  6. 04 Jan, 2023 2 commits
  7. 03 Jan, 2023 1 commit
    • NielsRogge's avatar
      Add GIT (GenerativeImage2Text) (#20295) · 9c6f7485
      NielsRogge authored
      
      
      * First draft
      
      * Make model instantiation work
      
      * Fix copied from statement
      
      * More fixes
      
      * Add correct output head
      
      * Improve configuration
      
      * Add conversion script
      
      * Improve conversion script
      
      * Remove token_type_ids
      
      * Fix conversion of projection layers
      
      * Convert all weights
      
      * Use cats image
      
      * Make logits match
      
      * Generate caption on cats image
      
      * Add GITProcessor
      
      * Update conversion script
      
      * Add support for more checkpoints
      
      * Fix conversion script
      
      * Add initial tests
      
      * Remove cross-attention
      
      * More improvements
      
      * Remove is_decoder
      
      * Improve model tests
      
      * Improve tests
      
      * Improve model outputs
      
      * Fix model outputs equivalence
      
      * Fix more tests
      
      * Remove unused code
      
      * Use generate to generate text, no use of cache for now
      
      * Use generate more appropriately
      
      * Fix config tests
      
      * Fix style
      
      * Add support for use_cache
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Fix style
      
      * Fix GIT vision encoder
      
      * Update README
      
      * Fix integration test
      
      * Set bos and eos token ids
      
      * Improve docs
      
      * Improve code
      
      * Add support for provided attention_mask
      
      * Add copied from statement
      
      * Fix gradient checkpointing test
      
      * Set model_input_names
      
      * Investigate model_input_names
      
      * Remove script
      
      * Fix model inputs
      
      * Fix docstring
      
      * Rename GIT to Git
      
      * Support more models
      
      * Add support for textvqa model
      
      * Add video support
      
      * Extend conversion script for video
      
      * Add support for large variant
      
      * Add support for more models
      
      * Fix config archive map
      
      * Update integration test
      
      * Fix README
      
      * Fix CLIP mean and std
      
      * Update processor
      
      * Fix use_cache for video, thanks @gante
      
      * Remove print statements
      
      * Remove assertion
      
      * Add processor tests
      
      * Fix model_input_names
      
      * Use Auto API for processor
      
      * Fix processor tests
      
      * Fix integration test
      
      * Fix pipeline test
      
      * Make tests faster
      
      * Update conversion script
      
      * Update conversion script
      
      * Convert more checkpoints
      
      * Update conversion script
      
      * Fix typo
      
      * Update docstrings
      
      * Improve code snippets
      
      * Fix doc tests
      
      * Add more code examples茅
      
      * Fix doc tests
      
      * Add integration tests
      
      * Fix unused variable
      
      * revert
      
      * Add GIT to Japanese README
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      9c6f7485
  8. 21 Dec, 2022 1 commit
  9. 19 Dec, 2022 1 commit
  10. 16 Dec, 2022 1 commit
    • NielsRogge's avatar
      Add Swin2SR (#19784) · 26dd041c
      NielsRogge authored
      
      
      * First draft
      
      * Add more improvements
      
      * Improve forward pass
      
      * Fix layernorm
      
      * Add upscaler
      
      * More improvements
      
      * More improvements
      
      * More improvements
      
      * Improve conversion script
      
      * Add preprocessing
      
      * Make output match original implementation
      
      * Add additional attributes
      
      * Add support for more models
      
      * Support more models
      
      * Add support for real world sr
      
      * Add initial Swin2SRFeatureExtractor
      
      * Add ImageSuperResolutionOutput
      
      * Make more tests pass
      
      * Use BaseModelOutput
      
      * Fix one more test
      
      * Fix more tests
      
      * Fix another test
      
      * Fix all tests
      
      * Rename to Swin2SRImageProcessor
      
      * Fix toctree
      
      * Fix toctree
      
      * Fix rebase
      
      * Improve Swin2SRImageProcessor
      
      * Remove feature extractor file
      
      * Improve model
      
      * Improve conversion script
      
      * Fix integration test
      
      * Fix init
      
      * Fix conversion script
      
      * Address comments
      
      * Improve upsampler
      
      * Add NearestConvUpsampler
      
      * Improve pixel shuffle upsampler
      
      * Improve auxiliary upsampler
      
      * Improve conversion script
      
      * Rename conv_last to final_convolution
      
      * Fix rebase
      
      * Improve upsample module
      
      * Add padding to image processor
      
      * Fix bug
      
      * Update padding
      
      * Remove print statement and fix integration test
      
      * Improve docs
      
      * Add image processor tests
      
      * Convert all checkpoints, fix tests茅
      
      * Remove print statements
      
      * Fix import
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      26dd041c
  11. 12 Dec, 2022 1 commit
    • Ariel Ekgren's avatar
      Add gpt-sw3 model to transformers (#20209) · 5f94855d
      Ariel Ekgren authored
      
      
      * Add templates for gpt-sw3
      
      * Add templates for gpt-sw3
      
      * Added sentencepiece tokenizer
      
      * intermediate commit with many changes
      
      * fixed conflicts
      
      * Init commit for tokenization port
      
      * Tokenization progress
      
      * Remove fast tokenizer
      
      * Clean up and rename spm.model -> spiece.model
      
      * Remove TF -> PT conversion script template, Clean up Megatron -> PT script
      
      * Optimize encode & decode performance
      
      * added new attention
      
      * added new attention
      
      * attention for gpt-sw3 working
      
      * attention good
      
      * Cache is now working
      
      * fixed attention mask so that it works with causal attention
      
      * fixed badbmm bug for cpu and caching
      
      * updated config with correct parameters
      
      * Refactor and leave optimizations as separate functions to avoid breaking expected functionality
      
      * Fix special tokens mapping for both tokenizers
      
      * cleaning up of code and comments
      
      * HF compatible attention outputs
      
      * Tokenizer now passing tests, add documentation
      
      * Update documentation
      
      * reverted back to base implementation after checking that it is identical to pretrained model
      
      * updated gpt-sw3 config
      
      * updated conversion script
      
      * aligned parameters with gpt-sw3 config
      
      * changed default scale_attn_by_inverse_layer_idx to true
      
      * removed flag from conversion script
      
      * added temporary model path
      
      * reverted back to functioning convert script
      
      * small changes to default config
      
      * updated tests for gpt-sw3
      
      * make style, make quality, minor cleanup
      
      * Change local paths to testing online repository
      
      * Change name: GptSw3 -> GPTSw3
      
      * Remove GPTSw3TokenizerFast references
      
      * Use official model repository and add more model sizes
      
      * Added reference to 6.7b model
      
      * Add GPTSw3DoubleHeadsModel to IGNORE_NON_AUTO_CONFIGURED, like GPT2DoubleHeadsModel
      
      * Remove pointers to non-existing TFGPTSw3
      
      * Add GPTSw3 to docs/_toctree.yml
      
      * Remove TF artifacts from GPTSw3 in __init__ files
      
      * Update README:s with 'make fix-copies'
      
      * Add 20b model to archive list
      
      * Add documentation for GPT-Sw3
      
      * Fix typo in documentation for GPT-Sw3
      
      * Do 'make fix-copies' again after having updated docs
      
      * Fix some typos in docs
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update tests/models/gpt_sw3/test_tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Resolve comments from PR feedback
      
      * Resolve more comments from PR feedback, also set use_cache=True in convert script
      
      * Add '# Copied from' comments for GPTSw3 modeling
      
      * Set 'is_parallelizable = False'
      
      * Remove '# Copied from' where code was modified and add 'with x->y' when appropriate
      
      * Remove parallelize in mdx
      
      * make style, make quality
      
      * Update GPTSw3Config default values and corresponding documentation
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Clean up and protect GPTSw3Tokenizer imports with is_sentencepiece_available
      
      * Make style, make quality
      
      * Add dummy object for GPTSw3Tokenizer via 'make fix-copies'
      
      * make fix-copies
      
      * Remove GPTSw3 modeling classes
      
      * make style, make quality
      
      * Add GPTSw3 auto-mappings for other GPT2 heads
      
      * Update docs/source/en/model_doc/gpt-sw3.mdx
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Remove old TODO-comment
      
      * Add example usage to GPTSw3Tokenizer docstring
      
      * make style, make quality
      
      * Add implementation details and example usage to gpt-sw3.mdx
      Co-authored-by: default avatarJoeyOhman <joeyoh@kth.se>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      5f94855d
  12. 07 Dec, 2022 1 commit
  13. 06 Dec, 2022 1 commit
  14. 05 Dec, 2022 1 commit
    • Kamal Raj Kanakarajan's avatar
      Add BioGPT (#20420) · 13e73668
      Kamal Raj Kanakarajan authored
      * biogpt initial commit
      
      * updated init
      
      * fix faster decoding with use_cache
      
      * 1. fix input_ids and input_embeds with correct device
      2. added _keys_to_ignore_on_load_missing
      3. updated prepare_inputs_for_generation
      
      * add activation_dropout and scale_embedding
      
      * replace fsmt attention with bart attention
      
      * added test
      
      * run make fix-copies
      
      * doc init and fix build
      
      * updated README with proper information
      
      * 1. added tips to docs
      2. updated BioGptTokenizer func
      
      * 1. added tokenizer test
      2. refactor tokenizer
      
      * make fixup
      
      * add biogpt fairseq to hf converter
      
      * updated layer names more
      similar to original checkpoints
      
      * config update doc string and set defaults
      
      * added "#copied" from bart model and
      updated doc strings
      
      * enable model_input_names in tokenizer
      
      * 1.  positionalembedding depending on attention_mask
      2. added attention mask to prepare for generation
      
      * added test to verify past and generation
      
      * BioGptLMHeadModel -> BioGptForCausalLM
      
      * fix typo
      
      * tokenization and test
      Copyright and updated assertion
      
      * updated Copyright and
      one func at time in line
      
      * Copyright updates and
      minor doc fix
      
      * replace assertion with ValueError
      
      * rm extra space
      
      * added code syntax
      
      * revert cmnt position change
      
      * add tokenizer to auto
      
      * updated doc string
      
      * tokenizer doc string update
      
      * biogpt hub model update to microsoft/biogpt
      
      * make fixup
      
      * rm cmnt to fix flake8 5.0.4 vs 6 error
      13e73668
  15. 02 Dec, 2022 1 commit
    • fatih's avatar
      [New Model] Add TimeSformer model (#18908) · cc3d0e1b
      fatih authored
      * init timesformer
      
      * apply fix-copies
      
      * reformat style
      
      * revert back some incoorect style updates
      
      * init timesformer
      
      * apply fix-copies
      
      * reformat style
      
      * revert back some incoorect style updates
      
      * update timseformer doc
      
      * add some functions and classes
      
      * add new config params
      
      * implement multiple classes
      
      * update TimeSformerLayer
      
      * update TimeSformerModel, TimeSformerPreTrainedModel, TimeSformerEncoder
      
      * several fixes
      
      * reformat
      
      * temporary update
      
      * fix some typos
      
      * fix weight converter
      
      * more fixes
      
      * fix a typo
      
      * fix typo
      
      * remove redundant params
      
      * fix for latest hf-hub
      
      * merge fix
      
      * fix some checks
      
      * video classification works with einops
      
      * add paper info to docs
      
      * merge fix
      
      * remove redundant line
      
      * remove redundant docstring
      
      * update config
      
      * fix some typos
      
      * fix converter
      
      * update some test constants
      
      * refactor einops functions
      
      * reformat
      
      * fix a comment
      
      * remove redundat imports
      
      * reformat
      
      * fix a typo
      
      * remove comment
      
      * remove unused imports
      
      * remove redundant doc line
      
      * reformat
      
      * add missing line
      
      * fix docs
      
      * fix timesformer auto feat ext
      
      * add unittests
      
      * reformat
      
      * fix docs
      
      * some fixes and updates
      
      * fix readme
      
      * fix modeling
      
      * fix readme
      
      * update index
      
      * revert _toctree.yml changes
      
      * update timseformer.mdx
      
      * update drop_path_prob to drop_path_rate
      
      * add dosctring for drop_path_rate
      
      * update TimeSformerPatchEmbed naming
      
      * remove to_2tuple
      
      * explicit use of nn.functional
      
      * reformat
      
      * many updates from review comments
      
      * fix a typo
      
      * reformat
      
      * remove assert, better variable name
      
      * make variable names more explicit
      
      * add some adapted from
      
      * more explicit variable names
      
      * remove redundant docstring
      
      * fix initilaization
      
      * move permute inside embedding
      
      * update class names
      
      * remove unused imports
      
      * add test for video classification
      
      * update PretrainedModel with PreTrainedModel
      
      * remove double permute
      
      * update based on sylvain's review
      
      * aply auto fix
      
      * update image_processing_auto for timesformer
      
      * update hub urls
      
      * reformat
      
      * remove duplicate import
      
      * update doc link
      cc3d0e1b
  16. 01 Dec, 2022 1 commit
  17. 30 Nov, 2022 1 commit
    • Yang An's avatar
      Add Chinese-CLIP implementation (#20368) · 72176402
      Yang An authored
      
      
      * init chinese-clip model from clip
      
      * init model tests and docs
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * implement chinese-clip into hf
      
      * update usecase example in model implementation
      
      * fix codestyle
      
      * fix model_type typo in readme
      
      * add placeholder in doc
      
      * add placeholder in doc
      
      * update the init script
      
      * update usecase
      
      * fix codestyle
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * forward the convert_rgb
      
      * update testcase
      
      * update testcase
      
      * update testcase
      
      * merge the recent update from clip about model_input_name property
      
      * update the doc
      
      * update the doc
      
      * update the doc
      
      * update the doc
      
      * remove unused imports
      
      * reformat code style
      
      * update the doc
      
      * fix isort style
      
      * bypass a weird failed unit test which is unrelated with my PR
      
      * update the doc
      
      * implement independent vision config class
      
      * implement independent vision model class
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * fix refactor bug
      
      * fix refactor bug
      
      * doc-build restyle
      
      * implement independent text config class
      
      * implement independent text model class
      
      * implement independent text model class
      
      * make style
      
      * make fix-copies
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * fix refactor bug
      
      * make style
      
      * update doc
      
      * black and isort
      
      * update doc
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/auto/tokenization_auto.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * modify the model type from chinese-clip to chinese_clip
      
      * format the example comment of ChineseCLIPVisionConfig
      
      * correct the copyright comment
      
      * fix the tokenizer specification
      
      * add copied from for loss function
      
      * remove unused class
      
      * update CHINESE_CLIP_TEXT_INPUTS_DOCSTRING
      
      * update CHINESE_CLIP_INPUTS_DOCSTRING
      
      * update doc
      
      * update doc
      
      * update code comment in config
      
      * update copied from statement
      
      * make style
      
      * rename the doc file
      
      * add copied statement
      
      * remove unused attention_mask, causal_attention_mask in ChineseCLIPVisionEncoder
      
      * remove ChineseCLIPTextPreTrainedModel
      
      * fix bug
      
      * fix bug
      
      * fix bug
      
      * update doc
      
      * make style
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * update ChineseCLIPImageProcessor in image_processing_auto
      
      * fix config_class of chinesecliptextmodel
      
      * fix the test case
      
      * update the docs
      
      * remove the copied from comment for ChineseCLIPTextModel, since it has diverged from BertModel with customed config_class
      
      * update the testcase
      
      * final fix
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      72176402
  18. 21 Nov, 2022 3 commits
    • NielsRogge's avatar
      Add Audio Spectogram Transformer (#19981) · 4973d2a0
      NielsRogge authored
      
      
      * First draft
      
      * Make conversion script work
      
      * Add id2label mapping, run code quality
      
      * Fix copies
      
      * Add first draft of feature extractor
      
      * Update conversion script to use feature extractor
      
      * Make more tests pass
      
      * Add docs
      
      * update input_features to input_values + pad by default to max length
      
      * Fix doc tests
      
      * Add feature extractor tests
      
      * Add proper padding/truncation to feature extractor
      
      * Add support for conversion of all audioset checkpoints
      
      * Improve docs and extend conversion script
      
      * Fix README
      
      * Rename spectogram to spectrogram
      
      * Fix copies
      
      * Add integration test
      
      * Remove dummy conv
      
      * Update to ast
      
      * Update organization
      
      * Fix init
      
      * Rename model to AST
      
      * Add require_torchaudio annotator
      
      * Move import of ASTFeatureExtractor under a is_speech_available
      
      * Fix rebase
      
      * Add pipeline config
      
      * Update name of classifier head
      
      * Rename time_dimension and frequency_dimension for clarity
      
      * Remove print statement
      
      * Fix pipeline test
      
      * Fix pipeline test
      
      * Fix index table
      
      * Fix init
      
      * Fix conversion script
      
      * Rename to ForAudioClassification
      
      * Fix index table
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      4973d2a0
    • Matthijs Hollemans's avatar
      add MobileNetV1 model (#17799) · d21c97cc
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      rename files for MobileNetV1
      
      initial implementation of MobileNetV1
      
      fix conversion script
      
      cleanup
      
      write docs
      
      tweaks
      
      fix conversion script
      
      extract hidden states
      
      fix test cases
      
      make fixup
      
      fixup it all
      
      remove main from doc link
      
      fixes
      
      fix tests
      
      fix up
      
      use google org
      
      fix weird assert
      
      * fixup
      
      * use google organization for checkpoints
      d21c97cc
    • Raj Rajhans's avatar
      fix: "BigSicence" typo in docs (#20331) · 22d7161a
      Raj Rajhans authored
      22d7161a
  19. 18 Nov, 2022 1 commit
    • Ali Hassani's avatar
      Add Neighborhood Attention Transformer (NAT) and Dilated NAT (DiNAT) models (#20219) · fc4a993e
      Ali Hassani authored
      * Add DiNAT
      
      * Adds DiNAT + tests
      
      * Minor fixes
      
      * Added HF model
      
      * Add natten to dependencies.
      
      * Cleanup
      
      * Minor fixup
      
      * Reformat
      
      * Optional NATTEN import.
      
      * Reformat & add doc to _toctree
      
      * Reformat (finally)
      
      * Dummy objects for DiNAT
      
      * Add NAT + minor changes
      
      Adds NAT as its own independent model + docs, tests
      Adds NATTEN to ext deps to ensure ci picks it up.
      
      * Remove natten from `all` and `dev-torch` deps, add manual pip install to ci tests
      
      * Minor fixes.
      
      * Fix READMEs.
      
      * Requested changes to docs + minor fixes.
      
      * Requested changes.
      
      * Add NAT/DiNAT tests to layoutlm_job
      
      * Correction to Dinat doc.
      
      * Requested changes.
      fc4a993e
  20. 15 Nov, 2022 1 commit
    • Younes Belkada's avatar
      Add Switch transformers (#19323) · 163ac3d3
      Younes Belkada authored
      
      
      * first commit
      
      * add more comments
      
      * add router v1
      
      * clean up
      
      - remove `tf` modeling files
      
      * clean up
      
      - remove `tf` modeling files
      
      * clean up
      
      * v0 routers
      
      * added more router
      
      - Implemented `ExpertsChooseMaskedRouter`
      
      - added tests
      - 2 more routers to implement
      
      * last router
      
      * improved docstring
      
      - completed the docstring in `router.py`
      - added more args in the config
      
      * v0 sparse mlp
      
      * replace wrong naming
      
      * forward pass run
      
      * update MOE layer
      
      * small router update
      
      * fixup
      
      * consistency
      
      * remove scatter router
      
      * remove abstract layer
      
      * update test and model for integration testing
      
      * v1 conversion
      
      * update
      
      * hardcode hack
      
      * all keys match
      
      * add gin conversion, without additional libraries
      
      * update conversion sctipy
      
      * delete router file
      
      * update tests wrt router deletion
      
      * fix router issues
      
      * update expert code
      
      * update, logits match, code needsREFACTORING
      
      * Refactor code
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * add generate tests
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      
      * add support for router loss
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * fix forward error
      
      * refactor a bit
      
      * remove `FlaxSwitchTransformers` modules
      
      * more tests pass
      
      * Update code
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      
      * fixup
      
      * fix tests
      
      * fix doc
      
      * fix doc + tokenization
      
      * fix tokenizer test
      
      * fix test
      
      * fix loss output
      
      * update code for backward pass
      
      * add loss support
      
      * update documentation
      
      * fix documentation, clean tokenizer
      
      * more doc fix, cleanup example_switch
      
      * fix failing test
      
      * fix test
      
      * fix test
      
      * fix loss issue
      
      * move layer
      
      * update doc and fix router capacity usage
      
      * fixup
      
      * add sparse mlp index for documentation on hub
      
      * fixup
      
      * test sparse mix architecture
      
      * Apply suggestions from code review
      
      * Update docs/source/en/model_doc/switch_transformers.mdx
      
      * fixup on update
      
      * fix tests
      
      * fix another test
      
      * attempt fix
      
      * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/switch_transformers/convert_switch_transformers_original_flax_checkpoint_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * try
      
      * all tests pass
      
      * fix jitter noise
      
      * Apply suggestions from code review
      
      * doc tests pass
      
      * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * remove assert
      
      * change config order
      
      * fix readme japanese
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * remove parallelizable tests + add one liners
      
      * remove ONNX config
      
      * fix nits
      
      - add `T5Tokenizer` in auto mapping
      - remove `Switch Transformers` from ONNX supported models
      
      * remove `_get_router`
      
      * remove asserts
      
      * add check in test for `router_dtype`
      
      * add `SwitchTransformersConfig` in `run_pipeline_test`
      
      * Update tests/pipelines/test_pipelines_summarization.py
      
      * add huge model conversion script
      
      * fix slow tests
      
      - add better casting for `Linear8bitLt`
      - remove `torchscript` tests
      
      * add make dir
      
      * style on new script
      
      * fix nits
      
      - doctest
      - remove `_keys_to_ignore_on_load_unexpected`
      
      * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py
      
      * add google as authors
      
      * fix year
      
      * remove last `assert` statements
      
      * standardize vertical spaces
      
      * fix failing import
      
      * fix another failing test
      
      * Remove strange 脿uthorized_keys`
      
      * removing todo and padding that is never used
      Co-authored-by: default avatarArthur Zucker <arthur.zucker@gmail.com>
      Co-authored-by: default avatarybelkada <younes@huggingface.co>
      Co-authored-by: default avatarYounes Belkada <younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarArthur Zucker <arthur@huggingface.co>
      163ac3d3
  21. 14 Nov, 2022 1 commit
    • Matthijs Hollemans's avatar
      add MobileNetV2 model (#17845) · f711d683
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      * rename files for MobileNetV1
      
      * initial implementation of MobileNetV1
      
      * fix conversion script
      
      * cleanup
      
      * write docs
      
      * tweaks
      
      * fix conversion script
      
      * extract hidden states
      
      * fix test cases
      
      * make fixup
      
      * fixup it all
      
      * rename V1 to V2
      
      * fix checkpoints
      
      * fixup
      
      * implement first block + weight conversion
      
      * add remaining layers
      
      * add output stride and dilation
      
      * fixup
      
      * add tests
      
      * add deeplabv3+ head
      
      * a bit of fixup
      
      * finish deeplab conversion
      
      * add link to doc
      
      * fix issue with JIT trace
      
      in_height and in_width would be Tensor objects during JIT trace, which caused Core ML conversion to fail on the remainder op. By making them ints, the result of the padding calculation becomes a constant value.
      
      * cleanup
      
      * fix order of models
      
      * fix rebase error
      
      * remove main from doc link
      
      * add image processor
      
      * remove old feature extractor
      
      * fix converter + other issues
      
      * fixup
      
      * fix unit test
      
      * add to onnx tests (but these appear broken now)
      
      * add post_process_semantic_segmentation
      
      * use google org
      
      * remove unused imports
      
      * move args
      
      * replace weird assert
      f711d683
  22. 10 Nov, 2022 1 commit
  23. 08 Nov, 2022 2 commits
    • Weiwe Shi's avatar
      Add RocBert (#20013) · efa889d2
      Weiwe Shi authored
      
      
      * add roc_bert
      
      * update roc_bert readme
      
      * code style
      
      * change name and delete unuse file
      
      * udpate model file
      
      * delete unuse log file
      
      * delete tokenizer fast
      
      * reformat code and change model file path
      
      * add RocBertForPreTraining
      
      * update docs
      
      * delete wrong notes
      
      * fix copies
      
      * fix make repo-consistency error
      
      * fix files are not present in the table of contents error
      
      * change RocBert -> RoCBert
      
      * add doc, add detail test
      Co-authored-by: default avatarweiweishi <weiweishi@tencent.com>
      efa889d2
    • NielsRogge's avatar
      Add CLIPSeg (#20066) · 25896306
      NielsRogge authored
      
      
      * Add first draft
      
      * Update conversion script
      
      * Improve conversion script
      
      * Improve conversion script some more
      
      * Add conditional embeddings
      
      * Add initial decoder
      
      * Fix activation function of decoder
      
      * Make decoder outputs match original implementation
      
      * Make decoder outputs match original implementation
      
      * Add more copied from statements
      
      * Improve model outputs
      
      * Fix auto tokenizer file
      
      * Fix more tests
      
      * Add test
      
      * Improve README and docs, improve conditional embeddings
      
      * Fix more tests
      
      * Remove print statements
      
      * Remove initial embeddings
      
      * Improve conversion script
      
      * Add interpolation of position embeddings
      
      * Finish addition of interpolation of position embeddings
      
      * Add support for refined checkpoint
      
      * Fix refined checkpoint
      
      * Remove unused parameter
      
      * Improve conversion script
      
      * Add support for training
      
      * Fix conversion script
      
      * Add CLIPSegFeatureExtractor
      
      * Fix processor
      
      * Fix CLIPSegProcessor
      
      * Fix conversion script
      
      * Fix most tests
      
      * Fix equivalence test
      
      * Fix README
      
      * Add model to doc tests
      
      * Use better variable name
      
      * Convert other checkpoint as well
      
      * Update config, add link to paper
      
      * Add docs
      
      * Update organization
      
      * Replace base_model_prefix with clip
      
      * Fix base_model_prefix
      
      * Fix checkpoint of config
      
      * Fix config checkpoint
      
      * Remove file
      
      * Use logits for output
      
      * Fix tests
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      25896306
  24. 04 Nov, 2022 1 commit
  25. 01 Nov, 2022 2 commits
  26. 26 Oct, 2022 2 commits
  27. 18 Oct, 2022 1 commit
    • NielsRogge's avatar
      Add table transformer [v2] (#19614) · dd523da5
      NielsRogge authored
      * First draft
      
      * Add conversion script
      
      * Make conversion work
      
      * Upload checkpoints
      
      * Add final fixes
      
      * Revert changes of conditional and deformable detr
      
      * Fix toctree, add and remove copied from
      
      * Use model type
      
      * Improve docs
      
      * Improve code example
      
      * Update copies
      
      * Add copied formt
      
      * Don't update conditional detr
      
      * Don't update deformable detr
      dd523da5
  28. 17 Oct, 2022 2 commits
  29. 12 Oct, 2022 1 commit
    • NielsRogge's avatar
      Add LiLT (#19450) · 4d367a3c
      NielsRogge authored
      
      
      * First draft
      
      * Fix more things
      
      * Improve more things
      
      * Remove some head models
      
      * Fix more things
      
      * Add missing layers
      
      * Remove tokenizer
      
      * Fix more things
      
      * Fix copied from statements
      
      * Make all tests pass
      
      * Remove print statements
      
      * Remove files
      
      * Fix README and docs
      
      * Add integration test and fix organization
      
      * Add tips
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Make tests faster, improve docs
      
      * Fix doc tests
      
      * Add model to toctree
      
      * Add docs
      
      * Add note about creating new checkpoint
      
      * Remove is_decoder
      
      * Make tests smaller, add docs
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      4d367a3c
  30. 11 Oct, 2022 1 commit
  31. 10 Oct, 2022 1 commit