- 13 Jun, 2022 1 commit
-
-
Daniel Stancl authored
* Initial commit * Make some fixes * Make PT model full forward pass * Drop TF & Flax implementation, fix copies etc * Add Flax model and update some corresponding stuff * Drop some TF things * Update config and flax local attn * Add encoder_attention_type to config * . * Update docs * Do some cleansing * Fix some issues -> make style; add some docs * Fix position_bias + mask addition + Update tests * Fix repo consistency * Fix model consistency by removing flax operation over attn_mask * [WIP] Add PT TGlobal LongT5 * . * [WIP] Add flax tglobal model * [WIP] Update flax model to use the right attention type in the encoder * Fix flax tglobal model forward pass * Make the use of global_relative_attention_bias * Add test suites for TGlobal model * Fix minor bugs, clean code * Fix pt-flax equivalence though not convinced with correctness * Fix LocalAttn implementation to match the original impl. + update READMEs * Few updates * Update: [Flax] improve large model init and loading #16148 * Add ckpt conversion script accoring to #16853 + handle torch device placement * Minor updates to conversion script. * Typo: AutoModelForSeq2SeqLM -> FlaxAutoModelForSeq2SeqLM * gpu support + dtype fix * Apply some suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * * Remove (de)parallelize stuff * Edit shape comments * Update README.md * make fix-copies * Remove caching logic for local & tglobal attention * Apply another batch of suggestions from code review * Add missing checkpoints * Format converting scripts * Drop (de)parallelize links from longT5 mdx * Fix converting script + revert config file change * Revert "Remove caching logic for local & tglobal attention" This reverts commit 2a619828f6ddc3e65bd9bb1725a12b77fa883a46. * Stash caching logic in Flax model * Make side relative bias used always * Drop caching logic in PT model * Return side bias as it was * Drop all remaining model parallel logic * Remove clamp statements * Move test files to the proper place * Update docs with new version of hf-doc-builder * Fix test imports * Make some minor improvements * Add missing checkpoints to docs * Make TGlobal model compatible with torch.onnx.export * Replace some np.ndarray with jnp.ndarray * Fix TGlobal for ONNX conversion + update docs * fix _make_global_fixed_block_ids and masked neg value * update flax model * style and quality * fix imports * remove load_tf_weights_in_longt5 from init and fix copies * add slow test for TGlobal model * typo fix * Drop obsolete is_parallelizable and one warning * Update __init__ files to fix repo-consistency * fix pipeline test * Fix some device placements * [wip]: Update tests -- need to generate summaries to update expected_summary * Fix quality * Update LongT5 model card * Update (slow) summarization tests * make style * rename checkpoitns * finish * fix flax tests Co-authored-by:
phungvanduy <pvduy23@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
patil-suraj <surajp815@gmail.com>
-
- 09 Jun, 2022 2 commits
- 03 Jun, 2022 1 commit
-
-
Patrick Deutschmann authored
* Start adding perceiver support for ONNX * Fix pad token bug for fast tokenizers * Fix formatting * Make get_preprocesor more opinionated (processor priority, otherwise tokenizer/feature extractor) * Clean docs format * Minor cleanup following @sgugger's comments * Fix typo in docs * Fix another docs typo * Fix one more typo in docs * Update src/transformers/onnx/utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/onnx/utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/onnx/utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 01 Jun, 2022 1 commit
-
-
Ruihua Fang authored
* add onnx config for SqueezeBert * add test for onnx config for SqueezeBert * add automatically updated doc for onnx config for SqueezeBert * Update src/transformers/onnx/features.py Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> * Update src/transformers/models/squeezebert/configuration_squeezebert.py Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com>
-
- 31 May, 2022 1 commit
-
-
Ritik Nandwal authored
* Add onnx configuration for xlm * Add supported features for xlm * Add xlm to models exportable with onnx * Add xlm architecture to test file * Modify docs * Make code quality fixes
-
- 18 May, 2022 1 commit
-
-
Jingya HUANG authored
Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com>
-
- 12 May, 2022 1 commit
-
-
Sylvain Gugger authored
* Black preview * Fixup too! * Fix check copies * Use the same version as the CI * Bump black
-
- 09 May, 2022 1 commit
-
-
Manan Dey authored
* update docs of length_penalty * Revert "update docs of length_penalty" This reverts commit 466bf4800b75ec29bd2ff75bad8e8973bd98d01c. * add mobilebert onnx config * address suggestions * Update auto.mdx * Update __init__.py * Update features.py
-
- 06 May, 2022 1 commit
-
-
Ritik Nandwal authored
* Add onnx configuration for bigbird-pegasus * Modify docs
-
- 04 May, 2022 1 commit
-
-
lewtun authored
* Skip RoFormer ONNX test if rjieba not installed * Update deps table * Skip RoFormer serialization test * Fix RoFormer vocab * Add rjieba to CircleCI
-
- 26 Apr, 2022 1 commit
-
-
Krishna Sirumalla authored
* add roformer onnx config
-
- 25 Apr, 2022 2 commits
-
-
Rushi Chaudhari authored
* added deit onnx config
-
Thomas Chaigneau authored
Co-authored-by:ChainYo <t.chaigneau.tc@gmail.com>
-
- 22 Apr, 2022 1 commit
-
-
Thomas Chaigneau authored
* add OnnxConfig for ConvBert Co-authored-by:ChainYo <t.chaigneau.tc@gmail.com>
-
- 19 Apr, 2022 1 commit
-
-
Ella Charlaix authored
* Add export of models with a multiple-choice classification head
-
- 12 Apr, 2022 1 commit
-
-
Minh Chien Vu authored
* add Bigbird ONNX config
-
- 01 Apr, 2022 1 commit
-
-
Jim Rohrer authored
* Add beit onnx conversion support * Updated docs * Added cross reference to ViT ONNX config
-
- 25 Mar, 2022 1 commit
-
-
lewtun authored
* Add ONNX support for Blenderbot * Add BlenderbotSmall ONNX configuration * Update serialization table
-
- 23 Mar, 2022 1 commit
-
-
Sylvain Gugger authored
* Split file_utils in several submodules * Fixes * Add back more objects * More fixes * Who exactly decided to import that from there? * Second suggestion to code with code review * Revert wront move * Fix imports * Adapt all imports * Adapt all imports everywhere * Revert this import, will fix in a separate commit
-
- 14 Mar, 2022 1 commit
-
-
lewtun authored
* Make Camembert great again! * Add Camembert to TensorFlow ONNX tests
-
- 10 Mar, 2022 1 commit
-
-
lewtun authored
* Fix duplicate arguments passed to dummy inputs in ONNX export * Fix M2M100 ONNX config * Ensure we check PreTrained model only if torch is available * Remove TensorFlow tests for models without PyTorch parity
-
- 09 Mar, 2022 1 commit
-
-
lewtun authored
* Add ONNX support for ViT * Refactor to use generic preprocessor * Add vision dep to tests * Extend ONNX slow tests to ViT * Add dummy image generator * Use model_type to determine modality * Add deprecation warnings for tokenizer argument * Add warning when overwriting the preprocessor * Add optional args to docstrings * Add minimum PyTorch version to OnnxConfig * Refactor OnnxConfig class variables from CONSTANT_NAME to snake_case * Add reasonable value for default atol Co-authored-by:Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 02 Mar, 2022 1 commit
-
-
Michael Benayoun authored
* Add M2M100 support for ONNX export * Delete useless imports * Add M2M100 to tests * Fix protobuf issue
-
- 23 Feb, 2022 1 commit
-
-
Lysandre Debut authored
* Per-folder tests reorganization Co-authored-by:
sgugger <sylvain.gugger@gmail.com> Co-authored-by:
Stas Bekman <stas@stason.org>
-