"vscode:/vscode.git/clone" did not exist on "38e5b71abbf3f04925beb0a81af69fd17476405c"
  1. 31 Aug, 2022 1 commit
    • Ankur Goyal's avatar
      Add LayoutLMForQuestionAnswering model (#18407) · 5c4c8690
      Ankur Goyal authored
      
      
      * Add LayoutLMForQuestionAnswering model
      
      * Fix output
      
      * Remove TF TODOs
      
      * Add test cases
      
      * Add docs
      
      * TF implementation
      
      * Fix PT/TF equivalence
      
      * Fix loss
      
      * make fixup
      
      * Fix up documentation code examples
      
      * Fix up documentation examples + test them
      
      * Remove LayoutLMForQuestionAnswering from the auto mapping
      
      * Docstrings
      
      * Add better docstrings
      
      * Undo whitespace changes
      
      * Update tokenizers in comments
      
      * Fixup code and remove `from_pt=True`
      
      * Fix tests
      
      * Revert some unexpected docstring changes
      
      * Fix tests by overriding _prepare_for_class
      Co-authored-by: default avatarAnkur Goyal <ankur@impira.com>
      5c4c8690
  2. 30 Aug, 2022 5 commits
  3. 29 Aug, 2022 3 commits
  4. 26 Aug, 2022 2 commits
  5. 25 Aug, 2022 2 commits
    • Craig Chan's avatar
      Determine framework automatically before ONNX export (#18615) · fbf382c8
      Craig Chan authored
      
      
      * Automatic detection for framework to use when exporting to ONNX
      
      * Log message change
      
      * Incorporating PR comments, adding unit test
      
      * Adding tf for pip install for run_tests_onnxruntime CI
      
      * Restoring past changes to circleci yaml and test_onnx_v2.py, tests moved to tests/onnx/test_features.py
      
      * Fixup
      
      * Adding test to fetcher
      
      * Updating circleci config to log more
      
      * Changing test class name
      
      * Comment typo fix in tests/onnx/test_features.py
      Co-authored-by: default avatarlewtun <lewis.c.tunstall@gmail.com>
      
      * Moving torch_str/tf_str to self.framework_pt/tf
      
      * Remove -rA flag in circleci config
      Co-authored-by: default avatarlewtun <lewis.c.tunstall@gmail.com>
      fbf382c8
    • Patrick Deutschmann's avatar
      Add ONNX support for Longformer (#17176) · 3223d493
      Patrick Deutschmann authored
      * Implement ONNX support for Longformer
      
      Fix repo consistency check complaints
      
      Fix value mismatches
      
      Add pooler output for default model
      
      Increase validation atol to accommodate multiple-choice error
      
      Fix copies
      
      Fix chunking for longer sequence lengths
      
      Add future comment
      
      * Fix issue in mask_invalid_locations
      
      * Remove torch imports in configuration_longformer
      
      * Change config access to fix LED
      
      * Push opset version to support tril
      
      * Work in review comments (mostly style)
      
      * Add Longformer to ONNX tests
      3223d493
  6. 24 Aug, 2022 2 commits
  7. 22 Aug, 2022 1 commit
  8. 19 Aug, 2022 1 commit
  9. 18 Aug, 2022 2 commits
  10. 17 Aug, 2022 2 commits
    • amyeroberts's avatar
      Update feature extractor methods to enable type cast before normalize (#18499) · 49e44b21
      amyeroberts authored
      * Update methods to optionally rescale
      This is necessary to allow for casting our images / videos to numpy arrays within the feature extractors' call. We want to do this to make sure the behaviour is as expected when flags like  are False. If some transformations aren't applied, then the output type can't be unexpected e.g. a list of PIL images instead of numpy arrays.
      
      * Cast images to numpy arrays in call to enable consistent behaviour with different configs
      
      * Remove accidental clip changes
      
      * Update tests to reflect the scaling logic
      We write a generic  function to handle rescaling of our arrays. In order for the API to be intuitive, we take some factor c and rescale the image values by that. This means, the rescaling done in normalize and to_numpy_array are now done with array * (1/255) instead of array / 255. This leads to small differences in the resulting image. When testing, this was in the order of 1e-8, and so deemed OK
      49e44b21
    • Yih-Dar's avatar
      Fix Yolos ONNX export test (#18606) · c99e9846
      Yih-Dar authored
      
      Co-authored-by: default avatarlewtun <lewis.c.tunstall@gmail.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      c99e9846
  11. 16 Aug, 2022 1 commit
  12. 12 Aug, 2022 6 commits
    • Younes Belkada's avatar
      small change (#18584) · 1ccd2515
      Younes Belkada authored
      1ccd2515
    • Niklas Muennighoff's avatar
      Update BLOOM parameter counts (#18531) · 56ef0ba4
      Niklas Muennighoff authored
      * Update BLOOM parameter counts
      
      * Update BLOOM parameter counts
      56ef0ba4
    • NielsRogge's avatar
      Add Donut (#18488) · 2ab790e8
      NielsRogge authored
      
      
      * First draft
      
      * Improve script
      
      * Update script
      
      * Make conversion work
      
      * Add final_layer_norm attribute to Swin's config
      
      * Add DonutProcessor
      
      * Convert more models
      
      * Improve feature extractor and convert base models
      
      * Fix bug
      
      * Improve integration tests
      
      * Improve integration tests and add model to README
      
      * Add doc test
      
      * Add feature extractor to docs
      
      * Fix integration tests
      
      * Remove register_buffer
      
      * Fix toctree and add missing attribute
      
      * Add DonutSwin
      
      * Make conversion script work
      
      * Improve conversion script
      
      * Address comment
      
      * Fix bug
      
      * Fix another bug
      
      * Remove deprecated method from docs
      
      * Make Swin and Swinv2 untouched
      
      * Fix code examples
      
      * Fix processor
      
      * Update model_type to donut-swin
      
      * Add feature extractor tests, add token2json method, improve feature extractor
      
      * Fix failing tests, remove integration test
      
      * Add do_thumbnail for consistency
      
      * Improve code examples
      
      * Add code example for document parsing
      
      * Add DonutSwin to MODEL_NAMES_MAPPING
      
      * Add model to appropriate place in toctree
      
      * Update namespace to appropriate organization
      Co-authored-by: default avatarNiels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
      2ab790e8
    • Younes Belkada's avatar
      Supporting seq2seq models for `bitsandbytes` integration (#18579) · a5ca56ff
      Younes Belkada authored
      * Supporting seq2seq models for `bitsandbytes` integration
      
      - `bitsandbytes` integration supports now seq2seq models
      - check if a model has tied weights as an additional check
      
      * small modification
      
      - tie the weights before looking at tied weights!
      a5ca56ff
    • Joao Gante's avatar
      Generate: validate `model_kwargs` (and catch typos in generate arguments) (#18261) · ed1924e8
      Joao Gante authored
      * validate generate model_kwargs
      
      * generate tests -- not all models have an attn mask
      ed1924e8
    • Arthur's avatar
      Load sharded pt to flax (#18419) · bce36ee0
      Arthur authored
      
      
      * initial commit
      
      * add small test
      
      * add cross pt tf flag to test
      
      * fix quality
      
      * style
      
      * update test with new repo
      
      * fix failing test
      
      * update
      
      * fix wrong param ordering
      
      * style
      
      * update based on review
      
      * update related to recent new caching mechanism
      
      * quality
      
      * Update based on review
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      
      * quality and style
      
      * Update src/transformers/modeling_flax_utils.py
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      bce36ee0
  13. 11 Aug, 2022 2 commits
  14. 10 Aug, 2022 3 commits
    • Dhruv Karan's avatar
      Adds CLIP to models exportable with ONNX (#18515) · f62cb831
      Dhruv Karan authored
      
      
      * onnx config for clip
      
      * default opset as 14
      
      * changes from the original repo
      
      * input values order fix
      
      * outputs fix
      
      * remove unused import
      
      * ran make fix-copies
      
      * black format
      
      * review comments: forward ref, import fix, model change revert, .to cleanup
      
      * make style
      
      * formatting fixes
      
      * revert groupvit
      
      * comment for cast to int32
      
      * comment fix
      
      * make .T as .t() for onnx conversion
      
      * ran make fix-copies
      
      * remove unneeded comment
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * fix copies
      
      * remove comment
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      f62cb831
    • Sylvain Gugger's avatar
      Use commit hash to look in cache instead of calling head (#18534) · 0d0aada5
      Sylvain Gugger authored
      
      
      * Use commit hash to look in cache instead of calling head
      
      * Add tests
      
      * Add attr for local configs too
      
      * Stupid typos
      
      * Fix tests
      
      * Update src/transformers/utils/hub.py
      Co-authored-by: default avatarJulien Chaumond <julien@huggingface.co>
      
      * Address Julien's comments
      Co-authored-by: default avatarJulien Chaumond <julien@huggingface.co>
      0d0aada5
    • Younes Belkada's avatar
      `bitsandbytes` - `Linear8bitLt` integration into `transformers` models (#17901) · 4a51075a
      Younes Belkada authored
      
      
      * first commit
      
      * correct replace function
      
      * add final changes
      
      - works like charm!
      - cannot implement tests yet
      - tested
      
      * clean up a bit
      
      * add bitsandbytes dependencies
      
      * working version
      
      - added import function
      - added bitsandbytes utils file
      
      * small fix
      
      * small fix
      
      - fix import issue
      
      * fix import issues
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * refactor a bit
      
      - move bitsandbytes utils to utils
      - change comments on functions
      
      * reformat docstring
      
      - reformat docstring on init_empty_weights_8bit
      
      * Update src/transformers/__init__.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * revert bad formatting
      
      * change to bitsandbytes
      
      * refactor a bit
      
      - remove init8bit since it is useless
      
      * more refactoring
      
      - fixed init empty weights issue
      - added threshold param
      
      * small hack to make it work
      
      * Update src/transformers/modeling_utils.py
      
      * Update src/transformers/modeling_utils.py
      
      * revmoe the small hack
      
      * modify utils file
      
      * make style + refactor a bit
      
      * create correctly device map
      
      * add correct dtype for device map creation
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * apply suggestions
      
      - remove with torch.grad
      - do not rely on Python bool magic!
      
      * add docstring
      
       - add docstring for new kwargs
      
      * add docstring
      
      - comment `replace_8bit_linear` function
      - fix weird formatting
      
      * - added more documentation
      - added new utility function for memory footprint tracking
      - colab demo to add
      
      * few modifs
      
      - typo doc
      - force cast into float16 when load_in_8bit is enabled
      
      * added colab link
      
      * add test architecture + docstring a bit
      
      * refactor a bit testing class
      
      * make style + refactor a bit
      
      * enhance checks
      
      - add more checks
      - start writing saving test
      
      * clean up a bit
      
      * male style
      
      * add more details on doc
      
      * add more tests
      
      - still needs to fix 2 tests
      
      * replace by "or"
      
      - could not fix it from GitHub GUI
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * refactor a bit testing code + add readme
      
      * make style
      
      * fix import issue
      
      * Update src/transformers/modeling_utils.py
      Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
      
      * add few comments
      
      * add more doctring + make style
      
      * more docstring
      
      * raise error when loaded in 8bit
      
      * make style
      
      * add warning if loaded on CPU
      
      * add small sanity check
      
      * fix small comment
      
      * add bitsandbytes on dockerfile
      
      * Improve documentation
      
      - improve documentation from comments
      
      * add few comments
      
      * slow tests pass on the VM but not on the CI VM
      
      * Fix merge conflict
      
      * make style
      
      * another test should pass on a multi gpu setup
      
      * fix bad import in testing file
      
      * Fix slow tests
      
      - remove dummy batches
      - no more CUDA illegal memory errors
      
      * odify dockerfile
      
      * Update docs/source/en/main_classes/model.mdx
      
      * Update Dockerfile
      
      * Update model.mdx
      
      * Update Dockerfile
      
      * Apply suggestions from code review
      
      * few modifications
      
      - lm head can stay on disk/cpu
      - change model name so that test pass
      
      * change test value
      
      - change test value to the correct output
      - torch bmm changed to baddmm in bloom modeling when merging
      
      * modify installation guidelines
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * replace `n`by `name`
      
      * merge `load_in_8bit` and `low_cpu_mem_usage`
      
      * first try - keep the lm head in full precision
      
      * better check
      
      - check the attribute `base_model_prefix` instead of computing the number of parameters
      
      * added more tests
      
      * Update src/transformers/utils/bitsandbytes.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Merge branch 'integration-8bit' of https://github.com/younesbelkada/transformers
      
       into integration-8bit
      
      * improve documentation
      
      - fix typos for installation
      - change title in the documentation
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
      4a51075a
  15. 09 Aug, 2022 3 commits
  16. 08 Aug, 2022 2 commits
    • Sylvain Gugger's avatar
      Clean up hub (#18497) · 377cdded
      Sylvain Gugger authored
      * Clean up utils.hub
      
      * Remove imports
      
      * More fixes
      
      * Last fix
      377cdded
    • Nicolas Patry's avatar
      [DX fix] Fixing QA pipeline streaming a dataset. (#18516) · a4562552
      Nicolas Patry authored
      * [DX fix] Fixing QA pipeline streaming a dataset.
      
      QuestionAnsweringArgumentHandler would iterate over the whole dataset
      effectively killing all properties of the pipeline.
      This restores nice properties when using `Dataset` or `Generator` since
      those are meant to be consumed lazily.
      
      * Handling TF better.
      a4562552
  17. 06 Aug, 2022 1 commit
  18. 05 Aug, 2022 1 commit