1. 23 Dec, 2022 1 commit
  2. 22 Dec, 2022 1 commit
  3. 05 Dec, 2022 1 commit
  4. 22 Nov, 2022 1 commit
  5. 21 Nov, 2022 1 commit
    • Matthijs Hollemans's avatar
      add MobileNetV1 model (#17799) · d21c97cc
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      rename files for MobileNetV1
      
      initial implementation of MobileNetV1
      
      fix conversion script
      
      cleanup
      
      write docs
      
      tweaks
      
      fix conversion script
      
      extract hidden states
      
      fix test cases
      
      make fixup
      
      fixup it all
      
      remove main from doc link
      
      fixes
      
      fix tests
      
      fix up
      
      use google org
      
      fix weird assert
      
      * fixup
      
      * use google organization for checkpoints
      d21c97cc
  6. 14 Nov, 2022 1 commit
    • Matthijs Hollemans's avatar
      add MobileNetV2 model (#17845) · f711d683
      Matthijs Hollemans authored
      * add model files etc for MobileNetV2
      
      * rename files for MobileNetV1
      
      * initial implementation of MobileNetV1
      
      * fix conversion script
      
      * cleanup
      
      * write docs
      
      * tweaks
      
      * fix conversion script
      
      * extract hidden states
      
      * fix test cases
      
      * make fixup
      
      * fixup it all
      
      * rename V1 to V2
      
      * fix checkpoints
      
      * fixup
      
      * implement first block + weight conversion
      
      * add remaining layers
      
      * add output stride and dilation
      
      * fixup
      
      * add tests
      
      * add deeplabv3+ head
      
      * a bit of fixup
      
      * finish deeplab conversion
      
      * add link to doc
      
      * fix issue with JIT trace
      
      in_height and in_width would be Tensor objects during JIT trace, which caused Core ML conversion to fail on the remainder op. By making them ints, the result of the padding calculation becomes a constant value.
      
      * cleanup
      
      * fix order of models
      
      * fix rebase error
      
      * remove main from doc link
      
      * add image processor
      
      * remove old feature extractor
      
      * fix converter + other issues
      
      * fixup
      
      * fix unit test
      
      * add to onnx tests (but these appear broken now)
      
      * add post_process_semantic_segmentation
      
      * use google org
      
      * remove unused imports
      
      * move args
      
      * replace weird assert
      f711d683
  7. 01 Nov, 2022 1 commit
    • Mohit Sharma's avatar
      Added onnx config whisper (#19525) · c796b6de
      Mohit Sharma authored
      * Added onnx config whisper
      
      * added whisper support onnx
      
      * add audio input data
      
      * added whisper support onnx
      
      * fixed the seqlength value
      
      * Updated the whisper onnx ocnfig
      
      * restore files to old version
      
      * removed attention mask from inputs
      
      * Updated get_dummy_input_onnxruntime docstring
      
      * Updated relative imports and token generation
      
      * update docstring
      c796b6de
  8. 28 Oct, 2022 1 commit
  9. 10 Oct, 2022 1 commit
  10. 07 Oct, 2022 2 commits
  11. 31 Aug, 2022 1 commit
  12. 30 Aug, 2022 2 commits
  13. 25 Aug, 2022 2 commits
    • Craig Chan's avatar
      Determine framework automatically before ONNX export (#18615) · fbf382c8
      Craig Chan authored
      
      
      * Automatic detection for framework to use when exporting to ONNX
      
      * Log message change
      
      * Incorporating PR comments, adding unit test
      
      * Adding tf for pip install for run_tests_onnxruntime CI
      
      * Restoring past changes to circleci yaml and test_onnx_v2.py, tests moved to tests/onnx/test_features.py
      
      * Fixup
      
      * Adding test to fetcher
      
      * Updating circleci config to log more
      
      * Changing test class name
      
      * Comment typo fix in tests/onnx/test_features.py
      Co-authored-by: default avatarlewtun <lewis.c.tunstall@gmail.com>
      
      * Moving torch_str/tf_str to self.framework_pt/tf
      
      * Remove -rA flag in circleci config
      Co-authored-by: default avatarlewtun <lewis.c.tunstall@gmail.com>
      fbf382c8
    • Patrick Deutschmann's avatar
      Add ONNX support for Longformer (#17176) · 3223d493
      Patrick Deutschmann authored
      * Implement ONNX support for Longformer
      
      Fix repo consistency check complaints
      
      Fix value mismatches
      
      Add pooler output for default model
      
      Increase validation atol to accommodate multiple-choice error
      
      Fix copies
      
      Fix chunking for longer sequence lengths
      
      Add future comment
      
      * Fix issue in mask_invalid_locations
      
      * Remove torch imports in configuration_longformer
      
      * Change config access to fix LED
      
      * Push opset version to support tril
      
      * Work in review comments (mostly style)
      
      * Add Longformer to ONNX tests
      3223d493
  14. 18 Aug, 2022 1 commit
  15. 17 Aug, 2022 1 commit
  16. 12 Aug, 2022 1 commit
  17. 10 Aug, 2022 1 commit
  18. 09 Aug, 2022 2 commits
  19. 05 Aug, 2022 1 commit
  20. 03 Aug, 2022 1 commit
  21. 18 Jul, 2022 1 commit
  22. 06 Jul, 2022 1 commit
  23. 01 Jul, 2022 1 commit
  24. 30 Jun, 2022 1 commit
  25. 28 Jun, 2022 2 commits
  26. 21 Jun, 2022 1 commit
  27. 14 Jun, 2022 1 commit
  28. 13 Jun, 2022 1 commit
    • Daniel Stancl's avatar
      Add `LongT5` model (#16792) · a72f1c9f
      Daniel Stancl authored
      
      
      * Initial commit
      
      * Make some fixes
      
      * Make PT model full forward pass
      
      * Drop TF & Flax implementation, fix copies etc
      
      * Add Flax model and update some corresponding stuff
      
      * Drop some TF things
      
      * Update config and flax local attn
      
      * Add encoder_attention_type to config
      
      * .
      
      * Update docs
      
      * Do some cleansing
      
      * Fix some issues -> make style; add some docs
      
      * Fix position_bias + mask addition + Update tests
      
      * Fix repo consistency
      
      * Fix model consistency by removing flax operation over attn_mask
      
      * [WIP] Add PT TGlobal LongT5
      
      * .
      
      * [WIP] Add flax tglobal model
      
      * [WIP] Update flax model to use the right attention type in the encoder
      
      * Fix flax tglobal model forward pass
      
      * Make the use of global_relative_attention_bias
      
      * Add test suites for TGlobal model
      
      * Fix minor bugs, clean code
      
      * Fix pt-flax equivalence though not convinced with correctness
      
      * Fix LocalAttn implementation to match the original impl. + update READMEs
      
      * Few updates
      
      * Update: [Flax] improve large model init and loading #16148
      
      * Add ckpt conversion script accoring to #16853 + handle torch device placement
      
      * Minor updates to conversion script.
      
      * Typo: AutoModelForSeq2SeqLM -> FlaxAutoModelForSeq2SeqLM
      
      * gpu support + dtype fix
      
      * Apply some suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * * Remove (de)parallelize stuff
      * Edit shape comments
      * Update README.md
      * make fix-copies
      
      * Remove caching logic for local & tglobal attention
      
      * Apply another batch of suggestions from code review
      
      * Add missing checkpoints
      * Format converting scripts
      * Drop (de)parallelize links from longT5 mdx
      
      * Fix converting script + revert config file change
      
      * Revert "Remove caching logic for local & tglobal attention"
      
      This reverts commit 2a619828f6ddc3e65bd9bb1725a12b77fa883a46.
      
      * Stash caching logic in Flax model
      
      * Make side relative bias used always
      
      * Drop caching logic in PT model
      
      * Return side bias as it was
      
      * Drop all remaining model parallel logic
      
      * Remove clamp statements
      
      * Move test files to the proper place
      
      * Update docs with new version of hf-doc-builder
      
      * Fix test imports
      
      * Make some minor improvements
      
      * Add missing checkpoints to docs
      * Make TGlobal model compatible with torch.onnx.export
      * Replace some np.ndarray with jnp.ndarray
      
      * Fix TGlobal for ONNX conversion + update docs
      
      * fix _make_global_fixed_block_ids and masked neg  value
      
      * update flax model
      
      * style and quality
      
      * fix imports
      
      * remove load_tf_weights_in_longt5 from init and fix copies
      
      * add slow test for TGlobal model
      
      * typo fix
      
      * Drop obsolete is_parallelizable and one warning
      
      * Update __init__ files to fix repo-consistency
      
      * fix pipeline test
      
      * Fix some device placements
      
      * [wip]: Update tests -- need to generate summaries to update expected_summary
      
      * Fix quality
      
      * Update LongT5 model card
      
      * Update (slow) summarization tests
      
      * make style
      
      * rename checkpoitns
      
      * finish
      
      * fix flax tests
      Co-authored-by: default avatarphungvanduy <pvduy23@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarpatil-suraj <surajp815@gmail.com>
      a72f1c9f
  29. 09 Jun, 2022 2 commits
  30. 03 Jun, 2022 1 commit
  31. 01 Jun, 2022 1 commit
  32. 31 May, 2022 1 commit
    • Ritik Nandwal's avatar
      Added XLM onnx config (#17030) · 5af38953
      Ritik Nandwal authored
      * Add onnx configuration for xlm
      
      * Add supported features for xlm
      
      * Add xlm to models exportable with onnx
      
      * Add xlm architecture to test file
      
      * Modify docs
      
      * Make code quality fixes
      5af38953
  33. 18 May, 2022 1 commit
  34. 12 May, 2022 1 commit