"vscode:/vscode.git/clone" did not exist on "dc3b93c93f74ce816aa1a893d91549ec8fd9e585"
  1. 30 Jun, 2022 1 commit
  2. 29 Jun, 2022 2 commits
  3. 28 Jun, 2022 1 commit
  4. 24 Jun, 2022 1 commit
    • rooa's avatar
      Add CodeGen model (#17443) · d6b6fb99
      rooa authored
      
      
      * Add CodeGen model
      
      * Add missing key and switch order of super()
      
      * Fix torch.ones init with uint8 instead of bool
      
      * Address comments: copy statements and doc
      
      * update tests
      
      * remove old model parallel
      
      * fix batch gen tests
      
      * fix batch gen test
      
      * update test_gpt2_sample_max_time
      
      * fix codgen test and revert gpt2 test change
      
      * Fix incorrect tie_word_embedding value, typo, URL
      
      * Fix model order in README and styling
      
      * Reorder model list alphabetically
      
      * Set tie_word_embedding to False by default
      
      * Apply suggestions from code review
      
      * Better attn mask name & remove attn masked_bias
      
      * add tokenizer for codegen
      
      * quality
      
      * doc tokenizer
      
      * fix-copies
      
      * add CodeGenTokenizer in converter
      
      * make truncation optional
      
      * add test for truncation
      
      * add copyright
      
      * fix-copies
      
      * fix fast tokenizer decode
      
      * Update src/transformers/models/codegen/tokenization_codegen.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * increase vocab_size in tests
      Co-authored-by: default avatarpatil-suraj <surajp815@gmail.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      d6b6fb99
  5. 21 Jun, 2022 1 commit
  6. 13 Jun, 2022 1 commit
    • Daniel Stancl's avatar
      Add `LongT5` model (#16792) · a72f1c9f
      Daniel Stancl authored
      
      
      * Initial commit
      
      * Make some fixes
      
      * Make PT model full forward pass
      
      * Drop TF & Flax implementation, fix copies etc
      
      * Add Flax model and update some corresponding stuff
      
      * Drop some TF things
      
      * Update config and flax local attn
      
      * Add encoder_attention_type to config
      
      * .
      
      * Update docs
      
      * Do some cleansing
      
      * Fix some issues -> make style; add some docs
      
      * Fix position_bias + mask addition + Update tests
      
      * Fix repo consistency
      
      * Fix model consistency by removing flax operation over attn_mask
      
      * [WIP] Add PT TGlobal LongT5
      
      * .
      
      * [WIP] Add flax tglobal model
      
      * [WIP] Update flax model to use the right attention type in the encoder
      
      * Fix flax tglobal model forward pass
      
      * Make the use of global_relative_attention_bias
      
      * Add test suites for TGlobal model
      
      * Fix minor bugs, clean code
      
      * Fix pt-flax equivalence though not convinced with correctness
      
      * Fix LocalAttn implementation to match the original impl. + update READMEs
      
      * Few updates
      
      * Update: [Flax] improve large model init and loading #16148
      
      * Add ckpt conversion script accoring to #16853 + handle torch device placement
      
      * Minor updates to conversion script.
      
      * Typo: AutoModelForSeq2SeqLM -> FlaxAutoModelForSeq2SeqLM
      
      * gpu support + dtype fix
      
      * Apply some suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * * Remove (de)parallelize stuff
      * Edit shape comments
      * Update README.md
      * make fix-copies
      
      * Remove caching logic for local & tglobal attention
      
      * Apply another batch of suggestions from code review
      
      * Add missing checkpoints
      * Format converting scripts
      * Drop (de)parallelize links from longT5 mdx
      
      * Fix converting script + revert config file change
      
      * Revert "Remove caching logic for local & tglobal attention"
      
      This reverts commit 2a619828f6ddc3e65bd9bb1725a12b77fa883a46.
      
      * Stash caching logic in Flax model
      
      * Make side relative bias used always
      
      * Drop caching logic in PT model
      
      * Return side bias as it was
      
      * Drop all remaining model parallel logic
      
      * Remove clamp statements
      
      * Move test files to the proper place
      
      * Update docs with new version of hf-doc-builder
      
      * Fix test imports
      
      * Make some minor improvements
      
      * Add missing checkpoints to docs
      * Make TGlobal model compatible with torch.onnx.export
      * Replace some np.ndarray with jnp.ndarray
      
      * Fix TGlobal for ONNX conversion + update docs
      
      * fix _make_global_fixed_block_ids and masked neg  value
      
      * update flax model
      
      * style and quality
      
      * fix imports
      
      * remove load_tf_weights_in_longt5 from init and fix copies
      
      * add slow test for TGlobal model
      
      * typo fix
      
      * Drop obsolete is_parallelizable and one warning
      
      * Update __init__ files to fix repo-consistency
      
      * fix pipeline test
      
      * Fix some device placements
      
      * [wip]: Update tests -- need to generate summaries to update expected_summary
      
      * Fix quality
      
      * Update LongT5 model card
      
      * Update (slow) summarization tests
      
      * make style
      
      * rename checkpoitns
      
      * finish
      
      * fix flax tests
      Co-authored-by: default avatarphungvanduy <pvduy23@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarpatil-suraj <surajp815@gmail.com>
      a72f1c9f
  7. 09 Jun, 2022 2 commits
  8. 03 Jun, 2022 1 commit
  9. 02 Jun, 2022 1 commit
  10. 01 Jun, 2022 1 commit
  11. 31 May, 2022 1 commit
    • Ritik Nandwal's avatar
      Added XLM onnx config (#17030) · 5af38953
      Ritik Nandwal authored
      * Add onnx configuration for xlm
      
      * Add supported features for xlm
      
      * Add xlm to models exportable with onnx
      
      * Add xlm architecture to test file
      
      * Modify docs
      
      * Make code quality fixes
      5af38953
  12. 18 May, 2022 1 commit
  13. 16 May, 2022 1 commit
  14. 09 May, 2022 2 commits
  15. 06 May, 2022 1 commit
  16. 26 Apr, 2022 1 commit
  17. 25 Apr, 2022 1 commit
  18. 22 Apr, 2022 1 commit
  19. 18 Apr, 2022 1 commit
  20. 12 Apr, 2022 2 commits
  21. 08 Apr, 2022 1 commit
    • NielsRogge's avatar
      Add TAPEX (#16473) · 4ef0abb7
      NielsRogge authored
      
      
      * Add TapexTokenizer
      
      * Improve docstrings and provide option to provide answer
      
      * Remove option for pretokenized inputs
      
      * Add TAPEX to README
      
      * Fix copies
      
      * Remove option for pretokenized inputs
      
      * Initial commit: add tapex fine-tuning examples on both table-based question answering and table-based fact verification.
      
      * - Draft a README file for running the script and introducing some background.
      - Remove unused code lines in tabfact script.
      - Disable the deafult `pad_to_max_length` option which is memory-consuming.
      
      * * Support `as_target_tokenizer` function for TapexTokenizer.
      * Fix the do_lower_case behaviour of TapexTokenizer.
      * Add unit tests for target scenarios and cased/uncased scenarios for both source and target.
      
      * * Replace the label BartTokenizer with TapexTokenizer's as_target_tokenizer function.
      * Fix typos in tapex example README.
      
      * * fix the evaluation script - remove the property `task_name`
      
      * * Make the label space more clear for tabfact tasks
      
      * * Using a new fine-tuning script for tapex-base on tabfact.
      
      * * Remove the lowercase code outside the tokenizer - we use the tokenizer to control whether do_lower_case
      * Guarantee the hyper-parameter can be run without out-of-memory on 16GB card and report the new reproduced number on wikisql
      
      * * Remove the default tokenizer_name option.
      * Provide evaluation command.
      
      * * Support for WikiTableQuestion dataset.
      
      * Fix a typo in README.
      
      * * Fix the datasets's key name in WikiTableQuestions
      
      * Run make fixup and move test to folder
      
      * Fix quality
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply some more suggestions from code review
      
      * Improve docstrings
      
      * Overwrite failing test
      
      * Improve comment in example scripts
      
      * Fix rebase
      
      * Add TAPEX to Auto mapping
      
      * Add TAPEX to auto config mappings
      
      * Put TAPEX higher than BART in auto mapping
      
      * Add TAPEX to doc tests
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MBP.localdomain>
      Co-authored-by: default avatarSivilTaram <qianlxc@outlook.com>
      Co-authored-by: default avatarNiels Rogge <nielsrogge@nielss-mbp.home>
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      4ef0abb7
  22. 04 Apr, 2022 1 commit
  23. 01 Apr, 2022 1 commit
  24. 25 Mar, 2022 2 commits
  25. 23 Mar, 2022 2 commits
  26. 22 Mar, 2022 1 commit
  27. 21 Mar, 2022 1 commit
  28. 15 Mar, 2022 1 commit
  29. 09 Mar, 2022 1 commit
    • lewtun's avatar
      Add ONNX export for ViT (#15658) · 50dd314d
      lewtun authored
      
      
      * Add ONNX support for ViT
      
      * Refactor to use generic preprocessor
      
      * Add vision dep to tests
      
      * Extend ONNX slow tests to ViT
      
      * Add dummy image generator
      
      * Use model_type to determine modality
      
      * Add deprecation warnings for tokenizer argument
      
      * Add warning when overwriting the preprocessor
      
      * Add optional args to docstrings
      
      * Add minimum PyTorch version to OnnxConfig
      
      * Refactor OnnxConfig class variables from CONSTANT_NAME to snake_case
      
      * Add reasonable value for default atol
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      50dd314d
  30. 02 Mar, 2022 1 commit
  31. 01 Mar, 2022 1 commit
  32. 18 Feb, 2022 1 commit
    • Gunjan Chhablani's avatar
      Add PLBart (#13269) · ae1f8350
      Gunjan Chhablani authored
      * Init PLBART
      
      * Add missing configuration file
      
      * Add conversion script and configurationf ile
      
      * Fix style
      
      * Update modeling and conversion scripts
      
      * Fix scale embedding in config
      
      * Add comment
      
      * Fix conversion script
      
      * Add classification option to conversion script
      
      * Fix vocab size in config doc
      
      * Add tokenizer files from MBart50
      
      * Allow no lang code in regular tokenizer
      
      * Add PLBart Tokenizer Converters
      
      * Remove mask from multi tokenizer
      
      * Remove mask from multi tokenizer
      
      * Change from MBart-50 to MBart tokenizer
      
      * Fix names and modify src/tgt behavior
      
      * Fix imports for tokenizer
      
      * Remove <mask> from multi tokenizer
      
      * Fix style
      
      * Change tokenizer_class to processor_class
      
      * Add attribute map to config class
      
      * Update modeling file to modified MBart code
      
      * Update configuration file to MBart style configuration
      
      * Fix tokenizer
      
      * Separate tokenizers
      
      * Fix error in tokenization auto
      
      * Copy MBart tests
      
      * Replace with MBart tokenization tests
      
      * Fix style
      
      * Fix language code in multi tokenizer
      
      * Fix configuration docs
      
      * Add entry for plbart_multi in transformers init
      
      * Add dummy objects and fix imports
      
      * Fix modeling tests
      
      * Add TODO in config
      
      * Fix copyright year
      
      * Fix modeling docs and test
      
      * Fix some tokenization tests and style
      
      * Add changes from review
      
      * Fix copies
      
      * Fix docs
      
      * Fix docs
      
      * Fix style
      
      * Fix year
      
      * Add changes from review
      
      * Remove extra changes
      
      * Fix base tokenizer and doc
      
      * Fix style
      
      * Fix modeling and slow tokenizer tests
      
      * Remove Multi-tokenizer Converter and Tests
      
      * Delete QA model and Multi Tokenizer dummy objects
      
      * Fix repo consistency and code quality issues
      
      * Fix example documentation
      
      * Fix style
      
      * Remove PLBartTokenizer from type checking in init
      
      * Fix consistency issue
      
      * Add changes from review
      
      * Fix style
      
      * Remove PLBartTokenizerFast
      
      * Remove FastTokenizer converter
      
      * Fix AutoTokenzier mapping
      
      * Add plbart to toctree and fix consistency issues
      
      * Add language codes tokenizer test
      
      * Fix styling and doc issues
      
      * Add fixes for failing tests
      
      * Fix copies
      
      * Fix failing modeling test
      
      * Change assert to assertTrue in modeling tests
      ae1f8350
  33. 10 Feb, 2022 2 commits