1. 29 Nov, 2022 3 commits
  2. 28 Nov, 2022 2 commits
  3. 21 Nov, 2022 3 commits
    • NielsRogge's avatar
      Add Audio Spectogram Transformer (#19981) · 4973d2a0
      NielsRogge authored
      
      
      * First draft
      
      * Make conversion script work
      
      * Add id2label mapping, run code quality
      
      * Fix copies
      
      * Add first draft of feature extractor
      
      * Update conversion script to use feature extractor
      
      * Make more tests pass
      
      * Add docs
      
      * update input_features to input_values + pad by default to max length
      
      * Fix doc tests
      
      * Add feature extractor tests
      
      * Add proper padding/truncation to feature extractor
      
      * Add support for conversion of all audioset checkpoints
      
      * Improve docs and extend conversion script
      
      * Fix README
      
      * Rename spectogram to spectrogram
      
      * Fix copies
      
      * Add integration test
      
      * Remove dummy conv
      
      * Update to ast
      
      * Update organization
      
      * Fix init
      
      * Rename model to AST
      
      * Add require_torchaudio annotator
      
      * Move import of ASTFeatureExtractor under a is_speech_available
      
      * Fix rebase
      
      * Add pipeline config
      
      * Update name of classifier head
      
      * Rename time_dimension and frequency_dimension for clarity
      
      * Remove print statement
      
      * Fix pipeline test
      
      * Fix pipeline test
      
      * Fix index table
      
      * Fix init
      
      * Fix conversion script
      
      * Rename to ForAudioClassification
      
      * Fix index table
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      4973d2a0
    • Matthijs Hollemans's avatar
      add MobileNetV1 model (#17799) · d21c97cc
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      rename files for MobileNetV1
      
      initial implementation of MobileNetV1
      
      fix conversion script
      
      cleanup
      
      write docs
      
      tweaks
      
      fix conversion script
      
      extract hidden states
      
      fix test cases
      
      make fixup
      
      fixup it all
      
      remove main from doc link
      
      fixes
      
      fix tests
      
      fix up
      
      use google org
      
      fix weird assert
      
      * fixup
      
      * use google organization for checkpoints
      d21c97cc
    • Joao Gante's avatar
  4. 18 Nov, 2022 1 commit
    • Ali Hassani's avatar
      Add Neighborhood Attention Transformer (NAT) and Dilated NAT (DiNAT) models (#20219) · fc4a993e
      Ali Hassani authored
      * Add DiNAT
      
      * Adds DiNAT + tests
      
      * Minor fixes
      
      * Added HF model
      
      * Add natten to dependencies.
      
      * Cleanup
      
      * Minor fixup
      
      * Reformat
      
      * Optional NATTEN import.
      
      * Reformat & add doc to _toctree
      
      * Reformat (finally)
      
      * Dummy objects for DiNAT
      
      * Add NAT + minor changes
      
      Adds NAT as its own independent model + docs, tests
      Adds NATTEN to ext deps to ensure ci picks it up.
      
      * Remove natten from `all` and `dev-torch` deps, add manual pip install to ci tests
      
      * Minor fixes.
      
      * Fix READMEs.
      
      * Requested changes to docs + minor fixes.
      
      * Requested changes.
      
      * Add NAT/DiNAT tests to layoutlm_job
      
      * Correction to Dinat doc.
      
      * Requested changes.
      fc4a993e
  5. 17 Nov, 2022 1 commit
    • NielsRogge's avatar
      Add AutoBackbone + ResNetBackbone (#20229) · 6b217c52
      NielsRogge authored
      
      
      * Add ResNetBackbone
      
      * Define channels and strides as property
      
      * Remove file
      
      * Add test for backbone
      
      * Update BackboneOutput class
      
      * Remove strides property
      
      * Fix docstring
      
      * Add backbones to SHOULD_HAVE_THEIR_OWN_PAGE
      
      * Fix auto mapping name
      
      * Add sanity check for out_features
      
      * Set stage names based on depths
      
      * Update to tuple
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      6b217c52
  6. 15 Nov, 2022 3 commits
  7. 14 Nov, 2022 2 commits
    • Bartosz Szmelczynski's avatar
      Fix tapas scatter (#20149) · 78a471ff
      Bartosz Szmelczynski authored
      
      
      * First draft
      
      * Remove scatter dependency
      
      * Add require_torch
      
      * update vectorized sum test, add clone call
      
      * remove artifacts
      
      * fix style
      
      * fix style v2
      
      * remove "scatter" mentions from the code base
      
      * fix isort error
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      78a471ff
    • Matthijs Hollemans's avatar
      add MobileNetV2 model (#17845) · f711d683
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      * rename files for MobileNetV1
      
      * initial implementation of MobileNetV1
      
      * fix conversion script
      
      * cleanup
      
      * write docs
      
      * tweaks
      
      * fix conversion script
      
      * extract hidden states
      
      * fix test cases
      
      * make fixup
      
      * fixup it all
      
      * rename V1 to V2
      
      * fix checkpoints
      
      * fixup
      
      * implement first block + weight conversion
      
      * add remaining layers
      
      * add output stride and dilation
      
      * fixup
      
      * add tests
      
      * add deeplabv3+ head
      
      * a bit of fixup
      
      * finish deeplab conversion
      
      * add link to doc
      
      * fix issue with JIT trace
      
      in_height and in_width would be Tensor objects during JIT trace, which caused Core ML conversion to fail on the remainder op. By making them ints, the result of the padding calculation becomes a constant value.
      
      * cleanup
      
      * fix order of models
      
      * fix rebase error
      
      * remove main from doc link
      
      * add image processor
      
      * remove old feature extractor
      
      * fix converter + other issues
      
      * fixup
      
      * fix unit test
      
      * add to onnx tests (but these appear broken now)
      
      * add post_process_semantic_segmentation
      
      * use google org
      
      * remove unused imports
      
      * move args
      
      * replace weird assert
      f711d683
  8. 10 Nov, 2022 1 commit
  9. 09 Nov, 2022 1 commit
  10. 08 Nov, 2022 3 commits
    • amyeroberts's avatar
      AutoImageProcessor (#20111) · 4eb918e6
      amyeroberts authored
      * AutoImageProcessor skeleton
      
      * Update references
      
      * Add mapping in init
      
      * Add model image processors to __init__ for importing
      
      * Add AutoImageProcessor tests
      
      * Fix up
      
      * Image Processor documentation
      
      * Remove pdb
      
      * Update docs/source/en/model_doc/mobilevit.mdx
      
      * Update docs
      
      * Don't add whitespace on json files
      
      * Remove fixtures
      
      * Move checking model config down
      
      * Fix up
      
      * Add check for image processor
      
      * Remove FeatureExtractorMixin in docstrings
      
      * Rename model_tmpfile to config_tmpfile
      
      * Don't make None if not in image processor map
      4eb918e6
    • Weiwe Shi's avatar
      Add RocBert (#20013) · efa889d2
      Weiwe Shi authored
      
      
      * add roc_bert
      
      * update roc_bert readme
      
      * code style
      
      * change name and delete unuse file
      
      * udpate model file
      
      * delete unuse log file
      
      * delete tokenizer fast
      
      * reformat code and change model file path
      
      * add RocBertForPreTraining
      
      * update docs
      
      * delete wrong notes
      
      * fix copies
      
      * fix make repo-consistency error
      
      * fix files are not present in the table of contents error
      
      * change RocBert -> RoCBert
      
      * add doc, add detail test
      Co-authored-by: default avatarweiweishi <weiweishi@tencent.com>
      efa889d2
    • NielsRogge's avatar
      Add CLIPSeg (#20066) · 25896306
      NielsRogge authored
      
      
      * Add first draft
      
      * Update conversion script
      
      * Improve conversion script
      
      * Improve conversion script some more
      
      * Add conditional embeddings
      
      * Add initial decoder
      
      * Fix activation function of decoder
      
      * Make decoder outputs match original implementation
      
      * Make decoder outputs match original implementation
      
      * Add more copied from statements
      
      * Improve model outputs
      
      * Fix auto tokenizer file
      
      * Fix more tests
      
      * Add test
      
      * Improve README and docs, improve conditional embeddings
      
      * Fix more tests
      
      * Remove print statements
      
      * Remove initial embeddings
      
      * Improve conversion script
      
      * Add interpolation of position embeddings
      
      * Finish addition of interpolation of position embeddings
      
      * Add support for refined checkpoint
      
      * Fix refined checkpoint
      
      * Remove unused parameter
      
      * Improve conversion script
      
      * Add support for training
      
      * Fix conversion script
      
      * Add CLIPSegFeatureExtractor
      
      * Fix processor
      
      * Fix CLIPSegProcessor
      
      * Fix conversion script
      
      * Fix most tests
      
      * Fix equivalence test
      
      * Fix README
      
      * Add model to doc tests
      
      * Use better variable name
      
      * Convert other checkpoint as well
      
      * Update config, add link to paper
      
      * Add docs
      
      * Update organization
      
      * Replace base_model_prefix with clip
      
      * Fix base_model_prefix
      
      * Fix checkpoint of config
      
      * Fix config checkpoint
      
      * Remove file
      
      * Use logits for output
      
      * Fix tests
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      25896306
  11. 01 Nov, 2022 2 commits
    • Sylvain Gugger's avatar
      v4.25.0.dev0 · c3a93d8d
      Sylvain Gugger authored
      c3a93d8d
    • Matt's avatar
      Add ESMFold (#19977) · 7f9b7b3f
      Matt authored
      
      
      * initial commit
      
      * First draft that gets outputs without crashing!
      
      * Add all the ported openfold dependencies
      
      * testing
      
      * Restructure config files for ESMFold
      
      * Debugging to find output discrepancies
      
      * Mainly style
      
      * Make model runnable without extra deps
      
      * Remove utils and merge them to the modeling file
      
      * Use correct gelu and remove some debug prints
      
      * More cleanup
      
      * Update esm docs
      
      * Update conversion script to support ESMFold properly
      
      * Port some top-level changes from ESMFold repo
      
      * Expand EsmFold docstrings
      
      * Make attention_mask optional (default to all 1s)
      
      * Add inference test for ESMFold
      
      * Use config and not n kwargs
      
      * Add modeling output class
      
      * Remove einops
      
      * Remove chunking in ESM FFN
      
      * Update tests for ESMFold
      
      * Quality
      
      * REpo consistency
      
      * Remove tree dependency from ESMFold
      
      * make fixup
      
      * Add an error in case my structure map function breaks later
      
      * Remove needless code
      
      * Stop auto-casting the LM to float16 so CPU tests pass
      
      * Stop auto-casting the LM to float16 so CPU tests pass
      
      * Final test updates
      
      * Split test file
      
      * Copyright and quality
      
      * Unpin PyTorch to see built doc
      
      * Fix config file to_dict() method
      
      * Add some docstrings to the output
      
      * Skip TF checkpoint tests for ESM until we reupload those
      
      * make fixup
      
      * More docstrings
      
      * Unpin to get even with main
      
      * Flag example to write
      Co-authored-by: default avatarSylvain Gugger <Sylvain.gugger@gmail.com>
      7f9b7b3f
  12. 18 Oct, 2022 1 commit
    • NielsRogge's avatar
      Add table transformer [v2] (#19614) · dd523da5
      NielsRogge authored
      * First draft
      
      * Add conversion script
      
      * Make conversion work
      
      * Upload checkpoints
      
      * Add final fixes
      
      * Revert changes of conditional and deformable detr
      
      * Fix toctree, add and remove copied from
      
      * Use model type
      
      * Improve docs
      
      * Improve code example
      
      * Update copies
      
      * Add copied formt
      
      * Don't update conditional detr
      
      * Don't update deformable detr
      dd523da5
  13. 17 Oct, 2022 2 commits
  14. 12 Oct, 2022 3 commits
  15. 11 Oct, 2022 3 commits
    • Mathieu Jouffroy's avatar
      [CvT] Tensorflow implementation (#18597) · 5ca131f3
      Mathieu Jouffroy authored
      
      
      * implemented TFCvtModel and TFCvtForImageClassification and modified relevant files, added an exception in convert_tf_weight_name_to_pt_weight_name, added quick testing file to compare with pytorch model
      
      * added docstring + testing file in transformers testing suite
      
      * added test in testing file, modified docs to pass repo-consistency, passed formatting test
      
      * refactoring + passing all test
      
      * small refacto, removing unwanted comments
      
      * improved testing config
      
      * corrected import error
      
      * modified acces to pretrained model archive list, to pass tf_test
      
      * corrected import structure in init files
      
      * modified testing for keras_fit with cpu
      
      * correcting PR issues + Refactoring
      
      * Refactoring : improving readability and reducing the number of permutations
      
      * corrected momentum value + cls_token initialization
      
      * removed from_pt as weights were added to the hub
      
      * Update tests/models/cvt/test_modeling_tf_cvt.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      5ca131f3
    • h's avatar
      Decouples `XLMProphet` model from `Prophet` (#19406) · ab856f68
      h authored
      * decouples xlm_prophet from prophet and adds copy patterns that pass the copy check
      
      * adds copy patterns to copied docstrings too
      
      * restores autodoc for XLMProphetNetModel
      
      * removes all-casing in a bunch of places to ensure that the model is compatible with all checkpoints on the hub
      
      * adds missing model to main init
      
      * adds autodocs to make document checker happy
      
      * adds missing pretrained model import
      
      * adds missing pretrained model import to main init
      
      * adds XLMProphetNetPreTrainedModel to the dummy pt objects
      
      * removes examples from the source-doc file since docstrings contain them already
      
      * adds a missing new line to make check_repo happy
      ab856f68
    • Sofia Oliveira's avatar
      Make `XLMRoberta` model and config independent from `Roberta` (#19359) · e38cf93e
      Sofia Oliveira authored
      * remove config dependence
      
      * remove dependencies from xlm_roberta
      
      * Fix style
      
      * Fix comments
      
      * various fixes
      
      * Fix pre-trained model name
      e38cf93e
  16. 10 Oct, 2022 3 commits
    • Lysandre's avatar
      Dev version · 10100979
      Lysandre authored
      10100979
    • amyeroberts's avatar
      Add TF whisper (#19378) · e3f028f3
      amyeroberts authored
      
      
      * simplify loop
      
      * add featur extractor
      
      * add model
      
      * start conversion
      
      * add dropout
      
      * initial commit of test files
      
      * copnversion for all models
      
      * update processor for correct padding
      
      * update feature extraction
      
      * update integration test logits match
      
      * fmnt: off for the logits
      
      * on the fly mel bank
      
      * small nit
      
      * update test
      
      * update tokenizer
      
      * nit feature extraction
      
      * update
      
      * update tokenizer test
      
      * adds logit processor and update tokenizer to get supress tokens
      
      * style
      
      * clean convert
      
      * revert to original modeling tf utils
      
      * Update
      
      * update
      
      * nit
      
      * clean convert file
      
      * update tests and nits
      
      * quality
      
      * slow generation test
      
      * ffn_dim to allow customization
      
      * update readme
      
      * add to toctreee
      
      * start fixing integration tests
      
      * update tests and code
      
      * fix feature extractor
      
      * fix config tests common
      
      * update code to fix tests
      
      * fix feature exctractor
      
      * nit feature extraction
      
      * update test for new feature extractor
      
      * style
      
      * add absrtact
      
      * large logits wioth custom decoder input ids
      
      * wraap around is otrch available
      
      * fix feature extractor
      
      * correct logits for whisper small.en
      
      * nit
      
      * fix encoder_attentino_mask
      
      * some fixes
      
      * remove unnecessary inputs
      
      * nits
      
      * add normalizer file
      
      * update etst tokenization
      
      * fix attention mask not defined
      
      * fix generate
      
      * remove uncoder attention mask useless
      
      * update test modeling whisper
      
      * update condfig to add second non supress tokens
      
      * nits on feature exrtactor
      
      * nit for test tokenizers
      
      * update etsts
      
      * update tests
      
      * update tokenization test
      
      * fixup
      
      * invalidated hf token. Clean convert openai to whisper
      
      * fix logit tests
      
      * fixup
      
      * Add model to README
      
      * Fix doc tests
      
      * clean merge
      
      * revert toc_tree changes
      
      * remove useless LogitProcessor
      
      * Update whisper .mdx
      
      * update config file doc
      
      * update configuration docstring
      
      * update test tokenization
      
      * update test tokenization
      
      * update tokenization whisper
      Added copied from where needed
      
      * update feature extraction
      
      * nit test name
      
      * style
      
      * quality
      
      * remove get suppress tokens and update non_speech tokens global variables
      
      * Update src/transformers/models/whisper/feature_extraction_whisper.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * clean modeling whisper and test
      Removed the attention mask arguments that are deprecated
      
      * fix large test
      
      * Add multilingual audio test, and translate test
      
      * style
      
      * fix larg multilingual test
      
      * nits
      
      * add copied from for attention layer
      
      * remove attention masks in doc
      
      * add english normalizer
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * update tokenization test
      
      * remove copied from in whisper attention : no bias in k_proj only
      
      * wrap around dependencies in english normalizer
      
      * style
      
      * correct import generation logits
      
      * for now, wrap feature extractor with torch
      
      * remove torch depencies for feature extraction and style
      
      * Update src/transformers/models/whisper/convert_openai_whisper_to_tfms.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * fixup
      
      * nit
      
      * update logitds
      
      * style
      
      * nit
      
      * nits and fix final tests
      
      * add `is_more_itertools_available` to utils
      
      * quality
      
      * add begin supress tokens, supress tokens to generate args and config
      
      * clean supressTokensLogitProcessor in generation logits
      
      * Nit naming
      
      * add supressTokensAtBegin
      
      * udpate tests, supress tokens to None or correct values
      
      * nit and style
      
      * update RAG to fit test and generate_logit
      
      * add copy pasted statment on english normalizer
      
      * add arguments to config_common_kwargs
      
      * Update src/transformers/generation_utils.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/generation_logits_process.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * revert changes based on reviews
      
      * update doc and nits
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * more nits
      
      * last nits
      
      * update test configuration common
      
      * add BART name in decoder attention mask documentation
      
      * Update src/transformers/models/whisper/modeling_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * style
      
      * nit
      
      * nit
      
      * add english.json file to git
      
      * nits on documentation
      
      * nit
      
      * nits
      
      * last styling
      
      * add main toctree file
      
      * remove sentence piece dependency
      
      * clean init file
      
      * fix tokenizer that has no dependencies on sentencepiece
      
      * update whisper init file, nit
      
      * remove english.json file
      
      * add get decoder prompt id
      
      * All weights loading
      
      * Remove hanging pdb
      
      * Fixup and tidy up
      
      * Use same copied from as PT model
      
      * Remove whitespace changes
      
      * Remove torch references
      
      * Tie embeddings
      
      * Remove logits processor input to generate
      
      * Update logit values
      
      * revert changes and add forced logit processor
      
      * nit
      
      * clean normalizer
      
      * remove protected
      
      * Add logit processors and update generation code & tests
      
      * Some tidy up
      
      * Update docstring
      
      * update
      
      * update based on review
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update to reflect changes on the PT model branch
      
      * Tidy up
      
      * Remove extra whitespace
      
      * Fix test - make input ids small enough we can append
      
      * Include upstream changes on main
      
      * PR comments - add batch tests, remove comments & defaults
      
      * Fix model output imports
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/generation_tf_logits_process.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update tests/models/whisper/test_modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update docstring example
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Remove changes to adjust_logits_during_generation function
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Tidy up imports that don't require TF
      
      * Update tests - skip and no more skip
      
      * Update tests/generation/test_generation_tf_logits_process.py
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Add training flags
      
      * Add (skipped) XLA generation tests
      
      * Add embedding correctness test
      
      * Add constant ids for generation tests
      
      * Make logits finding a bit tidier
      
      * Remove unused args
      
      * xla generation enabled
      
      * Don't skip XLA tests anymore
      
      * Fix tests - add position ids to expected signature and update rag generation
      
      * Undo method reorder
      
      * Remove added whitespace
      
      * Remove copy-paste gradient checkopint ref
      
      * Remove
      
      * Trigger CI - (issue with refs when pulling)
      Co-authored-by: default avatarArthur Zucker <arthur.zucker@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <niels.rogge1@gmail.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      Co-authored-by: default avatarJoao Gante <joao@huggingface.co>
      e3f028f3
    • APAVOU Clément's avatar
      Add `OPTForQuestionAnswering` (#19402) · af69360b
      APAVOU Clément authored
      * Add `OPTForQuestionAnswering`
      
      - added `OPTForQuestionAnswering` class based on `BloomForQuestionAnswering`
      - added `OPTForQuestionAnswering` in common tests
      - all common tests pass
      - make fixup done
      
      * added docstrings for OPTForQuestionAnswering
      
      * Fix docstrings for OPTForQuestionAnswering
      af69360b
  17. 07 Oct, 2022 2 commits
    • mustapha ajeghrir's avatar
      Make `Camembert` TF version independent from `Roberta` (#19364) · 34e0cc6d
      mustapha ajeghrir authored
      
      
      * camembert tf version independent
      
      * fixup
      
      * fixup, all working
      
      * remove comments
      
      * Adding copied from roberta
      Co-authored-by: default avatarMustapha AJEGHRIR <mustapha.ajeghrir@kleegroup.com>
      34e0cc6d
    • Amrit Sahu's avatar
      [WIP] Add ZeroShotObjectDetectionPipeline (#18445) (#18930) · e9a49bab
      Amrit Sahu authored
      * Add ZeroShotObjectDetectionPipeline (#18445)
      
      * Add AutoModelForZeroShotObjectDetection task
      
      This commit also adds the following
      
      - Add explicit _processor method for ZeroShotObjectDetectionPipeline.
        This is necessary as pipelines don't auto infer processors yet and
        `OwlVitProcessor` wraps tokenizer and feature_extractor together, to
        process multiple images at once
      
      - Add auto tests and other tests for ZeroShotObjectDetectionPipeline
      
      * Add AutoModelForZeroShotObjectDetection task
      
      This commit also adds the following
      
      - Add explicit _processor method for ZeroShotObjectDetectionPipeline.
        This is necessary as pipelines don't auto infer processors yet and
        `OwlVitProcessor` wraps tokenizer and feature_extractor together, to
        process multiple images at once
      
      - Add auto tests and other tests for ZeroShotObjectDetectionPipeline
      
      * Add batching for ZeroShotObjectDetectionPipeline
      
      * Fix doc-string ZeroShotObjectDetectionPipeline
      
      * Fix output format: ZeroShotObjectDetectionPipeline
      e9a49bab
  18. 05 Oct, 2022 2 commits
    • Arthur's avatar
      Add WhisperModel to transformers (#19166) · 45e14038
      Arthur authored
      
      
      * simplify loop
      
      * add featur extractor
      
      * add model
      
      * start conversion
      
      * add dropout
      
      * initial commit of test files
      
      * copnversion for all models
      
      * update processor for correct padding
      
      * update feature extraction
      
      * update integration test logits match
      
      * fmnt: off for the logits
      
      * on the fly mel bank
      
      * small nit
      
      * update test
      
      * update tokenizer
      
      * nit feature extraction
      
      * update
      
      * update tokenizer test
      
      * adds logit processor and update tokenizer to get supress tokens
      
      * style
      
      * clean convert
      
      * revert to original modeling tf utils
      
      * Update
      
      * update
      
      * nit
      
      * clean convert file
      
      * update tests and nits
      
      * quality
      
      * slow generation test
      
      * ffn_dim to allow customization
      
      * update readme
      
      * add to toctreee
      
      * start fixing integration tests
      
      * update tests and code
      
      * fix feature extractor
      
      * fix config tests common
      
      * update code to fix tests
      
      * fix feature exctractor
      
      * nit feature extraction
      
      * update test for new feature extractor
      
      * style
      
      * add absrtact
      
      * large logits wioth custom decoder input ids
      
      * wraap around is otrch available
      
      * fix feature extractor
      
      * correct logits for whisper small.en
      
      * nit
      
      * fix encoder_attentino_mask
      
      * some fixes
      
      * remove unnecessary inputs
      
      * nits
      
      * add normalizer file
      
      * update etst tokenization
      
      * fix attention mask not defined
      
      * Add model to README
      
      * Fix doc tests
      
      * fix generate
      
      * remove uncoder attention mask useless
      
      * update test modeling whisper
      
      * update condfig to add second non supress tokens
      
      * nits on feature exrtactor
      
      * nit for test tokenizers
      
      * update etsts
      
      * update tests
      
      * update tokenization test
      
      * fixup
      
      * invalidated hf token. Clean convert openai to whisper
      
      * fix logit tests
      
      * fixup
      
      * clean merge
      
      * revert toc_tree changes
      
      * remove useless LogitProcessor
      
      * Update whisper .mdx
      
      * update config file doc
      
      * update configuration docstring
      
      * update test tokenization
      
      * update test tokenization
      
      * update tokenization whisper
      Added copied from where needed
      
      * update feature extraction
      
      * nit test name
      
      * style
      
      * quality
      
      * remove get suppress tokens and update non_speech tokens global variables
      
      * Update src/transformers/models/whisper/feature_extraction_whisper.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * clean modeling whisper and test
      Removed the attention mask arguments that are deprecated
      
      * fix large test
      
      * Add multilingual audio test, and translate test
      
      * style
      
      * fix larg multilingual test
      
      * nits
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * add copied from for attention layer
      
      * remove attention masks in doc
      
      * add english normalizer
      
      * update tokenization test
      
      * remove copied from in whisper attention : no bias in k_proj only
      
      * wrap around dependencies in english normalizer
      
      * style
      
      * correct import generation logits
      
      * for now, wrap feature extractor with torch
      
      * Update src/transformers/models/whisper/convert_openai_whisper_to_tfms.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/whisper.mdx
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * remove torch depencies for feature extraction and style
      
      * fixup
      
      * nit
      
      * update logitds
      
      * style
      
      * nit
      
      * nits and fix final tests
      
      * add `is_more_itertools_available` to utils
      
      * quality
      
      * add begin supress tokens, supress tokens to generate args and config
      
      * clean supressTokensLogitProcessor in generation logits
      
      * Nit naming
      
      * add supressTokensAtBegin
      
      * udpate tests, supress tokens to None or correct values
      
      * nit and style
      
      * update RAG to fit test and generate_logit
      
      * add copy pasted statment on english normalizer
      
      * add arguments to config_common_kwargs
      
      * Update src/transformers/generation_utils.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/generation_logits_process.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * revert changes based on reviews
      
      * update doc and nits
      
      * more nits
      
      * last nits
      
      * update test configuration common
      
      * add BART name in decoder attention mask documentation
      
      * Update src/transformers/models/whisper/modeling_whisper.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * style
      
      * nit
      
      * nit
      
      * add english.json file to git
      
      * nits on documentation
      
      * nit
      
      * nits
      
      * last styling
      
      * add main toctree file
      
      * remove sentence piece dependency
      
      * clean init file
      
      * fix tokenizer that has no dependencies on sentencepiece
      
      * update whisper init file, nit
      
      * remove english.json file
      
      * add get decoder prompt id
      
      * revert changes and add forced logit processor
      
      * nit
      
      * clean normalizer
      
      * remove protected
      
      * update
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * update based on review
      
      * Update src/transformers/models/whisper/configuration_whisper.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * add batched tests
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarNielsRogge <niels.rogge1@gmail.com>
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      45e14038
    • mustapha ajeghrir's avatar
  19. 04 Oct, 2022 1 commit
  20. 30 Sep, 2022 1 commit
    • Kashif Rasul's avatar
      time series forecasting model (#17965) · 5cd16f01
      Kashif Rasul authored
      
      
      * initial files
      
      * initial model via cli
      
      * typos
      
      * make a start on the model config
      
      * ready with configuation
      
      * remove tokenizer ref.
      
      * init the transformer
      
      * added initial model forward to return dec_output
      
      * require gluonts
      
      * update dep. ver table and add as extra
      
      * fixed typo
      
      * add type for prediction_length
      
      * use num_time_features
      
      * use config
      
      * more config
      
      * typos
      
      * opps another typo
      
      * freq can be none
      
      * default via transformation is 1
      
      * initial transformations
      
      * fix imports
      
      * added transform_start_field
      
      * add helper to create pytorch dataloader
      
      * added inital val and test data loader
      
      * added initial distr head and loss
      
      * training working
      
      * remove TimeSeriesTransformerTokenizer
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/__init__.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/__init__.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * fixed copyright
      
      * removed docs
      
      * remove time series tokenizer
      
      * fixed docs
      
      * fix text
      
      * fix second
      
      * fix default
      
      * fix order
      
      * use config directly
      
      * undo change
      
      * fix comment
      
      * fix year
      
      * fix import
      
      * add additional arguments for training vs. test
      
      * initial greedy inference loop
      
      * fix inference
      
      * comment out token inputs to enc dec
      
      * Use HF encoder/decoder
      
      * fix inference
      
      * Use Seq2SeqTSModelOutput output
      
      * return Seq2SeqTSPredictionOutput
      
      * added default arguments
      
      * fix return_dict true
      
      * scale is a tensor
      
      * output static_features for inference
      
      * clean up some unused bits
      
      * fixed typo
      
      * set return_dict if none
      
      * call model once for both train/predict
      
      * use cache if future_target is none
      
      * initial generate func
      
      * generate arguments
      
      * future_time_feat is required
      
      * return SampleTSPredictionOutput
      
      * removed unneeded classes
      
      * fix when params is none
      
      * fix return dict
      
      * fix num_attention_heads
      
      * fix arguments
      
      * remove unused shift_tokens_right
      
      * add different dropout configs
      
      * implement FeatureEmbedder, Scaler and weighted_average
      
      * remove gluonts dependency
      
      * fix class names
      
      * avoid _variable names
      
      * remove gluonts dependency
      
      * fix imports
      
      * remove gluonts from configuration
      
      * fix docs
      
      * fixed typo
      
      * move utils to examples
      
      * add example requirements
      
      * config has no freq
      
      * initial run_ts_no_trainer
      
      * remove from ignore
      
      * fix output_attentions and removed unsued getters/setters
      
      * removed unsed tests
      
      * add dec seq len
      
      * add test_attention_outputs
      
      * set has_text_modality=False
      
      * add config attribute_map
      
      * make style
      
      * make fix-copies
      
      * add encoder_outputs to TimeSeriesTransformerForPrediction forward
      
      * Improve docs, add model to README
      
      * added test_forward_signature
      
      * More improvements
      
      * Add more copied from
      
      * Fix README
      
      * Fix remaining quality issues
      
      * updated encoder and decoder
      
      * fix generate
      
      * output_hidden_states and use_cache are optional
      
      * past key_values returned too
      
      * initialize weights of distribution_output module
      
      * fixed more tests
      
      * update test_forward_signature
      
      * fix return_dict outputs
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * removed commented out tests
      
      * added neg. bin and normal output
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * move to one line
      
      * Add docstrings
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * add try except for assert and raise
      
      * try and raise exception
      
      * fix the documentation formatting
      
      * fix assert call
      
      * fix docstring formatting
      
      * removed input_ids from DOCSTRING
      
      * Update input docstring
      
      * Improve variable names
      
      * Update order of inputs
      
      * Improve configuration
      
      * Improve variable names
      
      * Improve docs
      
      * Remove key_length from tests
      
      * Add extra docs
      
      * initial unittests
      
      * added test_inference_no_head test
      
      * added test_inference_head
      
      * add test_seq_to_seq_generation
      
      * make style
      
      * one line
      
      * assert mean prediction
      
      * removed comments
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      
      * fix order of args
      
      * make past_observed_mask optional as well
      
      * added Amazon license header
      
      * updated utils with new fieldnames
      
      * make style
      
      * cleanup
      
      * undo position of past_observed_mask
      
      * fix import
      
      * typo
      
      * more typo
      
      * rename example files
      
      * remove example for now
      
      * Update docs/source/en/_toctree.yml
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update modeling_time_series_transformer.py
      
      fix style
      
      * fixed typo
      
      * fix typo and grammer
      
      * fix style
      Co-authored-by: default avatarNielsRogge <48327001+NielsRogge@users.noreply.github.com>
      Co-authored-by: default avatarNielsRogge <niels.rogge1@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      5cd16f01