1. 04 Jan, 2023 1 commit
  2. 03 Jan, 2023 5 commits
    • Motoki Wu's avatar
      Add custom stop token ids for generation (#20727) · 45da7cec
      Motoki Wu authored
      * Add StopIdStoppingCriteria
      
      * add a working test for stop id criteria
      
      * add to global scope
      
      * add stop_ids to generate
      
      * add pipeline test
      
      * use tokenizer encode in test
      
      * add test to generation utils
      
      * reformat
      
      * fixup
      
      * make-fix-copies
      
      * rename to stop_token_id
      
      * use stop_tokens instead
      
      * add to text to text generation
      
      * make fixup
      
      * make repo-consistency
      
      * Add support for list of ints for eos_token_id inside generation/utils.py
      
      * Instead of having if elses, cast the eos_token_id into a List[int]
      
      * Add List[int] support for logits_process.py
      
      * add List[int] for beam_search.py
      
      * add List[int] for forced_eos_token_id
      
      * revert stop token id stopping criteria changes
      
      * make fixup
      
      * fix tests
      
      * add eos_token_id to generation/utils.py and added tests test_utils.py
      
      * add eos_token_id type hints and fix for pad tokens
      
      * add comments
      
      * remove some prints and remove forced false test
      
      * fix
      
      * put back test_stop_sequence_stopping_criteria
      
      * remove unused import and make fixup
      
      * add a none check
      
      * update docstring
      
      * add more docstring for list ints
      
      * make fixup
      45da7cec
    • Alara Dirik's avatar
      Improve OWL-ViT postprocessing (#20980) · cd245780
      Alara Dirik authored
      * add post_process_object_detection method
      
      * style changes
      cd245780
    • samuelpullely's avatar
      Enable `decoder_attention_mask` in `generate` function (#20726) · 15c68c67
      samuelpullely authored
      * Enable `decoder_attention_mask` in `generate` function
      
      * Make style corrections
      
      * Run `make repo-consistency`
      
      * Add integration test
      15c68c67
    • NielsRogge's avatar
      Add GIT (GenerativeImage2Text) (#20295) · 9c6f7485
      NielsRogge authored
      
      
      * First draft
      
      * Make model instantiation work
      
      * Fix copied from statement
      
      * More fixes
      
      * Add correct output head
      
      * Improve configuration
      
      * Add conversion script
      
      * Improve conversion script
      
      * Remove token_type_ids
      
      * Fix conversion of projection layers
      
      * Convert all weights
      
      * Use cats image
      
      * Make logits match
      
      * Generate caption on cats image
      
      * Add GITProcessor
      
      * Update conversion script
      
      * Add support for more checkpoints
      
      * Fix conversion script
      
      * Add initial tests
      
      * Remove cross-attention
      
      * More improvements
      
      * Remove is_decoder
      
      * Improve model tests
      
      * Improve tests
      
      * Improve model outputs
      
      * Fix model outputs equivalence
      
      * Fix more tests
      
      * Remove unused code
      
      * Use generate to generate text, no use of cache for now
      
      * Use generate more appropriately
      
      * Fix config tests
      
      * Fix style
      
      * Add support for use_cache
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Fix style
      
      * Fix GIT vision encoder
      
      * Update README
      
      * Fix integration test
      
      * Set bos and eos token ids
      
      * Improve docs
      
      * Improve code
      
      * Add support for provided attention_mask
      
      * Add copied from statement
      
      * Fix gradient checkpointing test
      
      * Set model_input_names
      
      * Investigate model_input_names
      
      * Remove script
      
      * Fix model inputs
      
      * Fix docstring
      
      * Rename GIT to Git
      
      * Support more models
      
      * Add support for textvqa model
      
      * Add video support
      
      * Extend conversion script for video
      
      * Add support for large variant
      
      * Add support for more models
      
      * Fix config archive map
      
      * Update integration test
      
      * Fix README
      
      * Fix CLIP mean and std
      
      * Update processor
      
      * Fix use_cache for video, thanks @gante
      
      * Remove print statements
      
      * Remove assertion
      
      * Add processor tests
      
      * Fix model_input_names
      
      * Use Auto API for processor
      
      * Fix processor tests
      
      * Fix integration test
      
      * Fix pipeline test
      
      * Make tests faster
      
      * Update conversion script
      
      * Update conversion script
      
      * Convert more checkpoints
      
      * Update conversion script
      
      * Fix typo
      
      * Update docstrings
      
      * Improve code snippets
      
      * Fix doc tests
      
      * Add more code examples茅
      
      * Fix doc tests
      
      * Add integration tests
      
      * Fix unused variable
      
      * revert
      
      * Add GIT to Japanese README
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      9c6f7485
    • Konstantin Kotik's avatar
      `MinNewTokensLengthLogitsProcessor` for `.generate` method #20814 (#20892) · 367fdf33
      Konstantin Kotik authored
      
      
      * feat: add min new length logit processor
      
      * test: add min new length logit processor
      
      * docs: add MinNewTokensLengthLogitsProcessor
      
      * feat: import MinNewTokensLengthLogitsProcessor
      
      * fix: update pytorch dummy objects
      
      * refactor & fix: rename attributes and var and get rid of dynamic attribute
      
      * tests: align test with new interface
      
      * docs: fix typo
      
      * docs: minor clarification
      
      * Empty-Commit
      
      * empty commit
      
      * run automated quality edits
      Co-authored-by: default avatarJoao Gante <joao@huggingface.co>
      367fdf33
  3. 31 Dec, 2022 2 commits
  4. 29 Dec, 2022 1 commit
  5. 27 Dec, 2022 1 commit
  6. 23 Dec, 2022 3 commits
  7. 22 Dec, 2022 1 commit
  8. 21 Dec, 2022 4 commits
  9. 20 Dec, 2022 3 commits
  10. 19 Dec, 2022 1 commit
  11. 16 Dec, 2022 3 commits
    • NielsRogge's avatar
      Add Swin2SR (#19784) · 26dd041c
      NielsRogge authored
      
      
      * First draft
      
      * Add more improvements
      
      * Improve forward pass
      
      * Fix layernorm
      
      * Add upscaler
      
      * More improvements
      
      * More improvements
      
      * More improvements
      
      * Improve conversion script
      
      * Add preprocessing
      
      * Make output match original implementation
      
      * Add additional attributes
      
      * Add support for more models
      
      * Support more models
      
      * Add support for real world sr
      
      * Add initial Swin2SRFeatureExtractor
      
      * Add ImageSuperResolutionOutput
      
      * Make more tests pass
      
      * Use BaseModelOutput
      
      * Fix one more test
      
      * Fix more tests
      
      * Fix another test
      
      * Fix all tests
      
      * Rename to Swin2SRImageProcessor
      
      * Fix toctree
      
      * Fix toctree
      
      * Fix rebase
      
      * Improve Swin2SRImageProcessor
      
      * Remove feature extractor file
      
      * Improve model
      
      * Improve conversion script
      
      * Fix integration test
      
      * Fix init
      
      * Fix conversion script
      
      * Address comments
      
      * Improve upsampler
      
      * Add NearestConvUpsampler
      
      * Improve pixel shuffle upsampler
      
      * Improve auxiliary upsampler
      
      * Improve conversion script
      
      * Rename conv_last to final_convolution
      
      * Fix rebase
      
      * Improve upsample module
      
      * Add padding to image processor
      
      * Fix bug
      
      * Update padding
      
      * Remove print statement and fix integration test
      
      * Improve docs
      
      * Add image processor tests
      
      * Convert all checkpoints, fix tests茅
      
      * Remove print statements
      
      * Fix import
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      26dd041c
    • Nicolas Patry's avatar
      Fix object detection2 (#20798) · 3ee95820
      Nicolas Patry authored
      * Revert "Fixing object detection with `layoutlm` (#20776)"
      
      This reverts commit fca66abe.
      
      * Better fix for layoutlm object detection.
      
      * Style.
      3ee95820
    • Younes Belkada's avatar
      [Pipeline] skip feature extraction test if in `IMAGE_PROCESSOR_MAPPING` (#20790) · 4341f4e2
      Younes Belkada authored
      skip feature extraction test if in `IMAGE_PROCESSOR_MAPPING`
      4341f4e2
  12. 15 Dec, 2022 5 commits
  13. 14 Dec, 2022 4 commits
  14. 13 Dec, 2022 3 commits
  15. 12 Dec, 2022 3 commits
    • amyeroberts's avatar
      Add decorator for flaky Donut tests (#20739) · 1416b5d9
      amyeroberts authored
      * Add decorator for flaky tests
      
      * Fix up
      1416b5d9
    • Ariel Ekgren's avatar
      Add gpt-sw3 model to transformers (#20209) · 5f94855d
      Ariel Ekgren authored
      
      
      * Add templates for gpt-sw3
      
      * Add templates for gpt-sw3
      
      * Added sentencepiece tokenizer
      
      * intermediate commit with many changes
      
      * fixed conflicts
      
      * Init commit for tokenization port
      
      * Tokenization progress
      
      * Remove fast tokenizer
      
      * Clean up and rename spm.model -> spiece.model
      
      * Remove TF -> PT conversion script template, Clean up Megatron -> PT script
      
      * Optimize encode & decode performance
      
      * added new attention
      
      * added new attention
      
      * attention for gpt-sw3 working
      
      * attention good
      
      * Cache is now working
      
      * fixed attention mask so that it works with causal attention
      
      * fixed badbmm bug for cpu and caching
      
      * updated config with correct parameters
      
      * Refactor and leave optimizations as separate functions to avoid breaking expected functionality
      
      * Fix special tokens mapping for both tokenizers
      
      * cleaning up of code and comments
      
      * HF compatible attention outputs
      
      * Tokenizer now passing tests, add documentation
      
      * Update documentation
      
      * reverted back to base implementation after checking that it is identical to pretrained model
      
      * updated gpt-sw3 config
      
      * updated conversion script
      
      * aligned parameters with gpt-sw3 config
      
      * changed default scale_attn_by_inverse_layer_idx to true
      
      * removed flag from conversion script
      
      * added temporary model path
      
      * reverted back to functioning convert script
      
      * small changes to default config
      
      * updated tests for gpt-sw3
      
      * make style, make quality, minor cleanup
      
      * Change local paths to testing online repository
      
      * Change name: GptSw3 -> GPTSw3
      
      * Remove GPTSw3TokenizerFast references
      
      * Use official model repository and add more model sizes
      
      * Added reference to 6.7b model
      
      * Add GPTSw3DoubleHeadsModel to IGNORE_NON_AUTO_CONFIGURED, like GPT2DoubleHeadsModel
      
      * Remove pointers to non-existing TFGPTSw3
      
      * Add GPTSw3 to docs/_toctree.yml
      
      * Remove TF artifacts from GPTSw3 in __init__ files
      
      * Update README:s with 'make fix-copies'
      
      * Add 20b model to archive list
      
      * Add documentation for GPT-Sw3
      
      * Fix typo in documentation for GPT-Sw3
      
      * Do 'make fix-copies' again after having updated docs
      
      * Fix some typos in docs
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update tests/models/gpt_sw3/test_tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Resolve comments from PR feedback
      
      * Resolve more comments from PR feedback, also set use_cache=True in convert script
      
      * Add '# Copied from' comments for GPTSw3 modeling
      
      * Set 'is_parallelizable = False'
      
      * Remove '# Copied from' where code was modified and add 'with x->y' when appropriate
      
      * Remove parallelize in mdx
      
      * make style, make quality
      
      * Update GPTSw3Config default values and corresponding documentation
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/__init__.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Clean up and protect GPTSw3Tokenizer imports with is_sentencepiece_available
      
      * Make style, make quality
      
      * Add dummy object for GPTSw3Tokenizer via 'make fix-copies'
      
      * make fix-copies
      
      * Remove GPTSw3 modeling classes
      
      * make style, make quality
      
      * Add GPTSw3 auto-mappings for other GPT2 heads
      
      * Update docs/source/en/model_doc/gpt-sw3.mdx
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Remove old TODO-comment
      
      * Add example usage to GPTSw3Tokenizer docstring
      
      * make style, make quality
      
      * Add implementation details and example usage to gpt-sw3.mdx
      Co-authored-by: default avatarJoeyOhman <joeyoh@kth.se>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      5f94855d
    • Nicolas Patry's avatar