"docs/vscode:/vscode.git/clone" did not exist on "3a43794dd65c198f1979a3ddd04179f58b29097c"
  1. 23 Dec, 2021 2 commits
    • lewtun's avatar
      Add ONNX support for MarianMT models (#14586) · 6b655cc6
      lewtun authored
      * First commit to add MarianMT to ONNX
      
      * Now MarianModel.forward() automatically generates decoder_input_ids, like BartModel.forward()
      
      * Adjusted MarianOnnxConfig.inputs and outputs to work with seq2seq-lm feature
      
      * Style fix
      
      * Added support for other features for already supported models
      
      * Partial support for causal and seq2seq models
      
      * Partial support for causal and seq2seq models
      
      * Add default task for MarianMT ONNX
      
      * Remove automatic creation of decoder_input_ids
      
      * Extend inputs and outputs for MarianMT ONNX config
      
      * Add MarianMT to ONNX unit tests
      
      * Refactor
      
      * OnnxSeq2SeqConfigWithPast to support seq2seq models
      
      * Parameterized the onnx tests
      
      * Restored run_mlm.py
      
      * Restored run_mlm.py
      
      * [WIP] BART update
      
      * BART and MBART
      
      * Add past_key_values and fix dummy decoder inputs
      
      Using a sequence length of 1 in generate_dummy_outputs() produces large discrepancies, presumably due to some hidden optimisations.
      
      * Refactor MarianOnnxConfig to remove custom past_key_values logic
      
      * Fix quality
      
      * Revert "Revert "Added support for other features for already supported models (#14358)" (#14679)"
      
      This reverts commit 0f4e39c5.
      
      * is_torch_available test to avoid failing imports
      
      * sorting parameterize parameters to solve ERROR gw0 gw1
      
      * tests fix
      
      * tests fix
      
      * GPT2 with past fix
      
      * Fixed stateful class attribute change that was breaking things when converting multiple models sequentially
      
      * Removed onnx file
      
      * Refactor Marian export to account for base changes
      
      * Fix copies
      
      * Implemented suggestions
      
      * Extend support for causal LM
      
      * Revert "Revert "Added support for other features for already supported models (#14358)" (#14679)"
      
      This reverts commit 0f4e39c5.
      
      * is_torch_available test to avoid failing imports
      
      * sorting parameterize parameters to solve ERROR gw0 gw1
      
      * tests fix
      
      * tests fix
      
      * GPT2 with past fix
      
      * Fixed stateful class attribute change that was breaking things when converting multiple models sequentially
      
      * Removed onnx file
      
      * Implemented suggestions
      
      * Fixed __init__ to resolve conflict with master
      
      * Revert "Revert "Added support for other features for already supported models (#14358)" (#14679)"
      
      This reverts commit 0f4e39c5
      
      .
      
      * is_torch_available test to avoid failing imports
      
      * sorting parameterize parameters to solve ERROR gw0 gw1
      
      * tests fix
      
      * tests fix
      
      * GPT2 with past fix
      
      * Fixed stateful class attribute change that was breaking things when converting multiple models sequentially
      
      * Removed onnx file
      
      * Implemented suggestions
      
      * Fixed __init__ to resolve conflict with master
      
      * Remove commented import
      
      * Remove ONNX model
      
      * Remove redundant class method
      
      * Tidy up imports
      
      * Fix quality
      
      * Refactor dummy input function
      
      * Add copied from statements to Marian config functions
      
      * Remove false copied from comments
      
      * Fix copy from comment
      Co-authored-by: default avatarMassimiliano Bruni <massimiliano.bruni@hcl.com>
      Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
      6b655cc6
    • Henrik Holm's avatar
  2. 22 Dec, 2021 3 commits
    • Michael Benayoun's avatar
      Onnx enable tasks for supported models (part 2) (#14700) · 13504dcb
      Michael Benayoun authored
      * Revert "Revert "Added support for other features for already supported models (#14358)" (#14679)"
      
      This reverts commit 0f4e39c5.
      
      * is_torch_available test to avoid failing imports
      
      * sorting parameterize parameters to solve ERROR gw0 gw1
      
      * tests fix
      
      * tests fix
      
      * GPT2 with past fix
      
      * Fixed stateful class attribute change that was breaking things when converting multiple models sequentially
      
      * Removed onnx file
      
      * Implemented suggestions
      
      * Fixed __init__ to resolve conflict with master
      
      * Remove commented import
      13504dcb
    • Ryokan RI's avatar
      Feature/fix slow test in mluke (#14749) · 824fd44f
      Ryokan RI authored
      * make MLukeTokenizerTest fast
      
      * make LukeTokenizerTest fast
      
      * add entry to _toctree.yaml
      824fd44f
    • SaulLu's avatar
      update the arguments `add_prefix_space` and `trim_offsets` in... · c94c1b89
      SaulLu authored
      update the arguments `add_prefix_space` and `trim_offsets` in `backend_tokenizer.post_processor` of `RobertaTokenizerFast` (#14752)
      
      * add tests
      
      * change post-processor, pre-tokenizer and decoder (can't update decoder)
      
      * update test (remove decoder which doesn't depend on trim and add_prefix)
      
      * just update the post_processor
      
      * fix change
      
      * `trim_offsets` has no influence on `pre_tokenizer`
      
      * remove a test that need some input from the `tokenizers` lib maintainers
      
      * format
      
      * add new test offsets roberta
      
      * polish comments
      c94c1b89
  3. 21 Dec, 2021 2 commits
  4. 20 Dec, 2021 6 commits
  5. 17 Dec, 2021 5 commits
  6. 16 Dec, 2021 4 commits
  7. 15 Dec, 2021 1 commit
    • Matt's avatar
      TF model cards (#14720) · 48d48276
      Matt authored
      * Initial commit for Keras model cards
      
      * Revert accidental change
      
      * make style
      
      * make style
      
      * make style
      
      * Fix PR comments
      
      * Move repo creation to __init__
      
      * Fixes to README.md creation
      
      * Partial progress for proper card creation on `push_to_hub`
      
      * Proper card creation from `push_to_hub` plus fixes for malformed model cards
      
      * Fixes for model card creation outside the callback
      
      * Adding a model card creation test
      
      * Putting the model card creation test in the right file.
      Good job, Matt.
      
      * make style
      
      * Fix model card test temp dir usage
      
      * Fix model card creation when no optimizer present
      
      * Fixes for when training history not present
      
      * Fix accidental edit to test_modeling_common
      48d48276
  8. 14 Dec, 2021 2 commits
    • Nicolas Patry's avatar
      Adding support for multiple mask tokens. (#14716) · e7ed7ffd
      Nicolas Patry authored
      * Adding support for multiple mask tokens.
      
      - Original implem: https://github.com/huggingface/transformers/pull/10222
      
      Co-authored-by: default avatarnjafer <naveen.jafer@oracle.com>
      
      * In order to accomodate optionally multimodal models like Perceiver
      
      we add information to the tasks to specify tasks where we know for sure
      if we need the tokenizer/feature_extractor or not.
      
      * Adding info in the documentation about multi masks.
      
      + marked as experimental.
      
      * Add a copy() to prevent overriding the same tensor over and over.
      
      * Fixup.
      
      * Adding small test for multi mask with real values..
      Co-authored-by: default avatarnjafer <naveen.jafer@oracle.com>
      e7ed7ffd
    • Nicolas Patry's avatar
      Fixing tests for Perceiver (#14739) · 546a91ab
      Nicolas Patry authored
      * Adding some slow test to check for perceiver at least from a high level.
      
      * Re-enabling fast tests for Perceiver ImageClassification.
      
      * Perceiver might try to run without Tokenizer (Fast doesn't exist) and
      with FeatureExtractor some text only pipelines.
      
      * Oops.
      
      * Adding a comment for `update_config_with_model_class`.
      
      * Remove `model_architecture` to get `tiny_config`.
      
      * Finalize rebase.
      
      * Smarter way to handle undefined FastTokenizer.
      
      * Remove old code.
      
      * Addressing some nits.
      
      * Don't instantiate `None`.
      546a91ab
  9. 13 Dec, 2021 3 commits
    • NielsRogge's avatar
      Improve perceiver (#14750) · e926ea2b
      NielsRogge authored
      * First draft
      
      * Improve docstring + clean up tests
      
      * Remove unused code
      
      * Add check in case one doesn't provide a preprocessor
      e926ea2b
    • Yih-Dar's avatar
      Fix: change tooslow to slow (#14734) · 12d9b957
      Yih-Dar authored
      
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      12d9b957
    • Lysandre Debut's avatar
      Fixing tests for Perceiver (#14745) · 3d66146a
      Lysandre Debut authored
      
      
      - Do not run image-classification pipeline (_CHECKPOINT_FOR_DOC uses the checkpoint for
      langage, which cannot load a FeatureExtractor so current logic fails).
      - Add a safeguard to not run tests when `tokenizer_class` or
      `feature_extractor_class` **are** defined, but cannot be loaded
      This happens for Perceiver for the "FastTokenizer" (which doesn't exist
      so None) and FeatureExtractor (which does exist but cannot be loaded
      because the checkpoint doesn't define one which is reasonable for the
      said checkpoint)
      - Added `get_vocab` function to `PerceiverTokenizer` since it is used by
      `fill-mask` pipeline when the argument `targets` is used to narrow a
      subset of possible values.
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      3d66146a
  10. 11 Dec, 2021 1 commit
  11. 09 Dec, 2021 1 commit
  12. 08 Dec, 2021 7 commits
    • Sylvain Gugger's avatar
      Make MLuke tokenizer tests slow (#14690) · e6219320
      Sylvain Gugger authored
      e6219320
    • lewtun's avatar
      0f4e39c5
    • Michael Benayoun's avatar
      Added support for other features for already supported models (#14358) · 0c70f145
      Michael Benayoun authored
      * Added support for other features for already supported models
      
      * Partial support for causal and seq2seq models
      
      * Partial support for causal and seq2seq models
      
      * OnnxSeq2SeqConfigWithPast to support seq2seq models
      
      * Parameterized the onnx tests
      
      * Restored run_mlm.py
      
      * Restored run_mlm.py
      
      * [WIP] BART update
      
      * BART and MBART
      
      * Added comments
      
      * Another sequence length of the past_key_values
      0c70f145
    • Patrick von Platen's avatar
      [AutoProcessor] Add Wav2Vec2WithLM & small fix (#14675) · ee4fa2e4
      Patrick von Platen authored
      * [AutoProcessor] Add Wav2Vec2WithLM & small fix
      
      * revert line removal
      
      * Update src/transformers/__init__.py
      
      * add test
      
      * up
      
      * up
      
      * small fix
      ee4fa2e4
    • NielsRogge's avatar
      Add Perceiver IO (#14487) · 65b20b73
      NielsRogge authored
      * First draft
      
      * Style and remove mlm
      
      * Make forward pass work
      
      * More improvements
      
      * More improvements
      
      * Fix bug
      
      * More improvements
      
      * More improvements
      
      * Add PerceiverTokenizer first draft
      
      * Improve conversion script
      
      * More improvements
      
      * Make conversion script work for the encoder
      
      * Make conversion script work with local pickle files
      
      * Style & quality, fix-copies
      
      * Add dummy input to conversion script
      
      * Add absolute position embeddings to TextPreProcessor
      
      * Make forward pass of encoder work
      
      * More improvements
      
      * Move text preprocessor to separate script
      
      * More improvements
      
      * More improvements
      
      * Add post processor
      
      * Make MLM model work
      
      * Style
      
      * Add PerceiverForMaskedLM
      
      * Add PerceiverImagePreprocessor
      
      * Make style
      
      * Make PerceiverForImageClassification work
      
      * More improvements
      
      * More improvements
      
      * Use tokenizer in conversion script
      
      * Use PerceiverForMaskedLM in conversion script
      
      * Define custom PerceiverModelOutput
      
      * Improve PerceiverAttention to make it work for both MLM and image classification
      
      * More improvements
      
      * More improvements
      
      * More improvements to the conversion script
      
      * Make conversion script work for both MLM and image classification
      
      * Add PerceiverFeatureExtractor
      
      * More improvements
      
      * Style and quality
      
      * Add center cropping
      
      * Fix bug
      
      * Small fix
      
      * Add print statement
      
      * Fix bug in image preprocessor
      
      * Fix bug with conversion script
      
      * Make output position embeddings an nn.Parameter layer instead of nn.Embedding
      
      * Comment out print statements
      
      * Add position encoding classes
      
      * More improvements
      
      * Use position_encoding_kwargs
      
      * Add PerceiverForImageClassificationFourier
      
      * Make style & quality
      
      * Add PerceiverForImageClassificationConvProcessing
      
      * Style & quality
      
      * Add flow model
      
      * Move processors to modeling file
      
      * Make position encodings modular
      
      * Make basic decoder use modular position encodings
      
      * Add PerceiverForOpticalFlow to conversion script
      
      * Add AudioPreprocessor
      
      * Make it possible for the basic decoder to use Fourier position embeddings
      
      * Add PerceiverForMultimodalAutoencoding
      
      * Improve model for optical flow
      
      * Improve _build_network_inputs method
      
      * Add print statement
      
      * Fix device issue
      
      * Fix device of Fourier embeddings
      
      * Add print statements for debugging
      
      * Add another print statement
      
      * Add another print statement
      
      * Add another print statement
      
      * Add another print statement
      
      * Improve PerceiverAudioPreprocessor
      
      * Improve conversion script for multimodal modal
      
      * More improvements
      
      * More improvements
      
      * Improve multimodal model
      
      * Make forward pass multimodal model work
      
      * More improvements
      
      * Improve tests
      
      * Fix some more tests
      
      * Add output dataclasses
      
      * Make more tests pass
      
      * Add print statements for debuggin
      
      * Add tests for image classification
      
      * Add PerceiverClassifierOutput
      
      * More improvements
      
      * Make more tests pass for the optical flow model
      
      * Make style & quality
      
      * Small improvements
      
      * Don't support training for optical flow model for now
      
      * Fix _prepare_for_class for tests
      
      * Make more tests pass, add some docs
      
      * Add multimodal model to tests
      
      * Minor fixes
      
      * Fix tests
      
      * Improve conversion script
      
      * Make fixup
      
      * Remove pos_dim argument
      
      * Fix device issue
      
      * Potential fix for OOM
      
      * Revert previous commit
      
      * Fix test_initialization
      
      * Add print statements for debugging
      
      * Fix print statement
      
      * Add print statement
      
      * Add print statement
      
      * Add print statement
      
      * Add print statement
      
      * Add print statement
      
      * Add print statement
      
      * Remove need for output_shape
      
      * Comment out output_shape
      
      * Remove unnecessary code
      
      * Improve docs
      
      * Fix make fixup
      
      * Remove PerceiverTextProcessor from init
      
      * Improve docs
      
      * Small improvement
      
      * Apply first batch of suggestions from code review
      
      * Apply more suggestions from code review
      
      * Update docstrings
      
      * Define dicts beforehand for readability
      
      * Rename task to architecture in conversion script, include PerceiverModel in tests
      
      * Add print statements for debugging
      
      * Fix tests on GPU
      
      * Remove preprocessors, postprocessors and decoders from main init
      
      * Add integration test
      
      * Fix docs
      
      * Replace einops by torch
      
      * Update for new docs frontend
      
      * Rename PerceiverForImageClassification
      
      * Improve docs
      
      * Improve docs
      
      * Improve docs of PerceiverModel
      
      * Fix some more tests
      
      * Improve center_crop
      
      * Add PerceiverForSequenceClassification
      
      * Small improvements
      
      * Fix tests
      
      * Add integration test for optical flow model
      
      * Clean up
      
      * Add tests for tokenizer
      
      * Fix tokenizer by adding special tokens properly
      
      * Fix CI
      65b20b73
    • Patrick von Platen's avatar
      [Wav2Vec2] PyCTCDecode Integration to support language model boosted decoding (#14339) · 961732c2
      Patrick von Platen authored
      
      
      * up
      
      * up
      
      * up
      
      * make it cleaner
      
      * correct
      
      * make styhahalal
      
      * add more tests
      
      * finish
      
      * small fix
      
      * make style
      
      * up
      
      * tryout to solve cicrle ci
      
      * up
      
      * fix more tests
      
      * fix more tests
      
      * apply sylvains suggestions
      
      * fix import
      
      * correct docs
      
      * add pyctcdecode only to speech tests
      
      * fix more tests
      
      * add tf, flax and pt tests
      
      * add pt
      
      * fix last tests
      
      * fix more tests
      
      * Apply suggestions from code review
      
      * change lines
      
      * Apply suggestions from code review
      Co-authored-by: default avatarAnton Lozhkov <aglozhkov@gmail.com>
      
      * correct tests
      
      * correct tests
      
      * add doc string
      Co-authored-by: default avatarAnton Lozhkov <aglozhkov@gmail.com>
      961732c2
    • Nicolas Patry's avatar
      Fixing Dataset for TQA + token-classification. (#14658) · 2e12d90b
      Nicolas Patry authored
      * Fixing Dataset for TQA + token-classification.
      
      * Fixing the tests.
      
      * Making sure `offset_mappings` is a valid argument.
      2e12d90b
  13. 07 Dec, 2021 2 commits
    • Stas Bekman's avatar
      [deepspeed] fix --load_best_model_at_end (#14652) · b66c5ab2
      Stas Bekman authored
      * [deepspeed] fix load_best_model_at_end
      
      * try with pull_request_target
      
      * revert: try with pull_request_target
      
      * style
      
      * add test
      
      * cleanup
      b66c5ab2
    • Ryokan RI's avatar
      Add mLUKE (#14640) · 30646a0a
      Ryokan RI authored
      * implement MLukeTokenizer and LukeForMaskedLM
      
      * update tests
      
      * update docs
      
      * add LukeForMaskedLM to check_repo.py
      
      * update README
      
      * fix test and specify the entity pad id in tokenization_(m)luke
      
      * fix EntityPredictionHeadTransform
      30646a0a
  14. 06 Dec, 2021 1 commit
    • Yih-Dar's avatar
      Use cross_attention_hidden_size in Encoder-Decoder models (#14378) · 4cdb67ca
      Yih-Dar authored
      
      
      * add cross_attention_hidden_size to text-2-text encoder-decoder models (PT/Flax)
      
      * for TFEncoderDecoderModel
      
      * add equivalence test for TFEncoderDecoderModel
      
      * fix
      
      * fix failed equivalence tests
      
      * remove unused import
      
      * add detailed comment
      
      * Fix check_equivalence_tf_to_pt by using encoder/decoder
      
      * cleaning
      
      * Use cross_attention_hidden_size in speech-to-text
      
      * clean fast init logging msg in encoder decoder models
      
      * increase tol from 1e-5 to 1e-3 for tf test
      
      * style
      
      * style
      
      * make sure projection layer can run
      
      * remove type conversion + add check
      
      * fix conflict (config.output_hidden_size)
      
      * Remove TF -> PT in check_pt_tf_equivalence for TFEncoderDecoderModel
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      4cdb67ca