".github/vscode:/vscode.git/clone" did not exist on "4ef85fee718969f1703d7dffa134deb72f4de828"
  1. 30 May, 2023 1 commit
  2. 26 May, 2023 7 commits
  3. 25 May, 2023 9 commits
  4. 24 May, 2023 17 commits
    • Matt's avatar
      Remove the last few TF serving sigs (#23738) · e45e756d
      Matt authored
      Remove some more serving methods that (I think?) turned up while this PR was open
      e45e756d
    • Sylvain Gugger's avatar
      Enable prompts on the Hub (#23662) · 9850e6dd
      Sylvain Gugger authored
      
      
      * Enable prompts on the Hub
      
      * Update src/transformers/tools/prompts.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Address review comments
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      9850e6dd
    • Zachary Mueller's avatar
      Fix sagemaker DP/MP (#23681) · 75bbf20b
      Zachary Mueller authored
      * Check for use_sagemaker_dp
      
      * Add a check for is_sagemaker_mp when setting _n_gpu again. Should be last broken thing
      
      * Try explicit check?
      
      * Quality
      75bbf20b
    • Daniel King's avatar
      Fix the regex in `get_imports` to support multiline try blocks and excepts... · 89159651
      Daniel King authored
      Fix the regex in `get_imports` to support multiline try blocks and excepts with specific exception types (#23725)
      
      * fix and test get_imports for multiline try blocks, and excepts with specific errors
      
      * fixup
      
      * add some more tests
      
      * add license
      89159651
    • Sanchit Gandhi's avatar
      d8222be5
    • Matt's avatar
      Overhaul TF serving signatures + dummy inputs (#23234) · 814de8fa
      Matt authored
      * Let's try autodetecting serving sigs
      
      * Don't clobber existing sigs
      
      * Change shapes for multiplechoice models
      
      * Make default dummy inputs smarter too
      
      * Fix missing f-string
      
      * Let's YOLO a serving output too
      
      * Read __class__.__name__ properly
      
      * Don't just pass naked lists in there and expect it to be okay
      
      * Code cleanup
      
      * Update default serving sig
      
      * Clearer error messages
      
      * Further updates to the default serving output
      
      * make fixup
      
      * Update the serving output a bit more
      
      * Cleanups and renames, raise errors appropriately when we can't infer inputs
      
      * More renames
      
      * we're building in a functional context again, yolo
      
      * import DUMMY_INPUTS from the right place
      
      * import DUMMY_INPUTS from the right place
      
      * Support cross-attention in the dummies
      
      * Support cross-attention in the dummies
      
      * Complete removal of dummy/serving overrides in BERT
      
      * Complete removal of dummy/serving overrides in RoBERTa
      
      * Obliterate lots and lots of serving sig and dummy overrides
      
      * merge type hint changes
      
      * Fix for token_type_ids with vocab_size 1
      
      * Add missing property decorator
      
      * Fix T5 and hopefully some models that take conv inputs
      
      * More signature pruning
      
      * Fix T5's signature
      
      * Fix Wav2Vec2 signature
      
      * Fix LongformerForMultipleChoice input signature
      
      * Fix BLIP and LED
      
      * Better default serving output error handling
      
      * Fix BART dummies
      
      * Fix dummies for cross-attention, esp encoder-decoder models
      
      * Fix visionencoderdecoder signature
      
      * Fix BLIP serving output
      
      * Small tweak to BART dummies
      
      * Cleanup the ugly parameter inspection line that I used in a few places
      
      * committed a breakpoint again
      
      * Move the text_dims check
      
      * Remove blip_text serving_output
      
      * Add decoder_input_ids to the default input sig
      
      * Remove all the manual overrides for encoder-decoder model signatures
      
      * Tweak longformer/led input sigs
      
      * Tweak default serving output
      
      * output.keys() -> output
      
      * make fixup
      814de8fa
    • Connor Henderson's avatar
      fix: Whisper generate, move text_prompt_ids trim up for max_new_tokens calculation (#23724) · 3d7baef1
      Connor Henderson authored
      move text_prompt_ids trimming to top
      3d7baef1
    • Jungnerd's avatar
      fix: delete duplicate sentences in `document_question_answering.mdx` (#23735) · 50a56bed
      Jungnerd authored
      fix: delete duplicate sentence
      50a56bed
    • Matt's avatar
      TF SAM memory reduction (#23732) · d2d88226
      Matt authored
      * Extremely small change to TF SAM dummies to reduce memory usage on build
      
      * remove debug breakpoint
      
      * Debug print statement to track array sizes
      
      * More debug shape printing
      
      * More debug shape printing
      
      * Now remove the debug shape printing
      
      * make fixup
      
      * make fixup
      d2d88226
    • pagarsky's avatar
      Minor awesome-transformers.md fixes (#23453) · 28aa438c
      pagarsky authored
      Minor docs fixes
      28aa438c
    • Matt's avatar
      Better TF docstring types (#23477) · f8b25744
      Matt authored
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Don't forget the imports
      
      * Add the imports to tests too
      
      * make fixup
      
      * Refactor tests that depended on get_type_hints
      
      * Better test refactor
      
      * Fix an old hidden bug in the test_keras_fit input creation code
      
      * Fix for the Deit tests
      f8b25744
    • Wang, Yi's avatar
      fix gptj could not jit.trace in GPU (#23317) · 767e6b53
      Wang, Yi authored
      
      Signed-off-by: default avatarWang, Yi A <yi.a.wang@intel.com>
      767e6b53
    • uchuhimo's avatar
      fix: use bool instead of uint8/byte in Deberta/DebertaV2/SEW-D to make it... · b4698b7e
      uchuhimo authored
      
      fix: use bool instead of uint8/byte in Deberta/DebertaV2/SEW-D to make it compatible with TensorRT (#23683)
      
      * Use bool instead of uint8/byte in DebertaV2 to make it compatible with TensorRT
      
      TensorRT cannot accept onnx graph with uint8/byte intermediate tensors. This PR uses bool tensors instead of unit8/byte tensors to make the exported onnx file can work with TensorRT.
      
      * fix: use bool instead of uint8/byte in Deberta and SEW-D
      
      ---------
      Co-authored-by: default avatarYuxian Qiu <yuxianq@nvidia.com>
      b4698b7e
    • Maria Khalusova's avatar
      Export to ONNX doc refocused on using optimum, added tflite (#23434) · 2eaaf17a
      Maria Khalusova authored
      
      
      * doc refocused on using optimum, tflite
      
      * minor updates to fix checks
      
      * Apply suggestions from code review
      Co-authored-by: default avatarregisss <15324346+regisss@users.noreply.github.com>
      
      * TFLite to separate page, added links
      
      * Removed the onnx list builder
      
      * make style
      
      * Update docs/source/en/serialization.mdx
      Co-authored-by: default avatarregisss <15324346+regisss@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatarregisss <15324346+regisss@users.noreply.github.com>
      2eaaf17a
    • Tim Dettmers's avatar
      Paged Optimizer + Lion Optimizer for Trainer (#23217) · 796162c5
      Tim Dettmers authored
      
      
      * Added lion and paged optimizers and made original tests pass.
      
      * Added tests for paged and lion optimizers.
      
      * Added and fixed optimizer tests.
      
      * Style and quality checks.
      
      ---------
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      796162c5
    • Tim Dettmers's avatar
      4-bit QLoRA via bitsandbytes (4-bit base model + LoRA) (#23479) · 9d73b922
      Tim Dettmers authored
      
      
      * Added lion and paged optimizers and made original tests pass.
      
      * Added tests for paged and lion optimizers.
      
      * Added and fixed optimizer tests.
      
      * Style and quality checks.
      
      * Initial draft. Some tests fail.
      
      * Fixed dtype bug.
      
      * Fixed bug caused by torch_dtype='auto'.
      
      * All test green for 8-bit and 4-bit layers.
      
      * Added fix for fp32 layer norms and bf16 compute in LLaMA.
      
      * Initial draft. Some tests fail.
      
      * Fixed dtype bug.
      
      * Fixed bug caused by torch_dtype='auto'.
      
      * All test green for 8-bit and 4-bit layers.
      
      * Added lion and paged optimizers and made original tests pass.
      
      * Added tests for paged and lion optimizers.
      
      * Added and fixed optimizer tests.
      
      * Style and quality checks.
      
      * Fixing issues for PR #23479.
      
      * Added fix for fp32 layer norms and bf16 compute in LLaMA.
      
      * Reverted variable name change.
      
      * Initial draft. Some tests fail.
      
      * Fixed dtype bug.
      
      * Fixed bug caused by torch_dtype='auto'.
      
      * All test green for 8-bit and 4-bit layers.
      
      * Added lion and paged optimizers and made original tests pass.
      
      * Added tests for paged and lion optimizers.
      
      * Added and fixed optimizer tests.
      
      * Style and quality checks.
      
      * Added missing tests.
      
      * Fixup changes.
      
      * Added fixup changes.
      
      * Missed some variables to rename.
      
      * revert trainer tests
      
      * revert test trainer
      
      * another revert
      
      * fix tests and safety checkers
      
      * protect import
      
      * simplify a bit
      
      * Update src/transformers/trainer.py
      
      * few fixes
      
      * add warning
      
      * replace with `load_in_kbit = load_in_4bit or load_in_8bit`
      
      * fix test
      
      * fix tests
      
      * this time fix tests
      
      * safety checker
      
      * add docs
      
      * revert torch_dtype
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * multiple fixes
      
      * update docs
      
      * version checks and multiple fixes
      
      * replace `is_loaded_in_kbit`
      
      * replace `load_in_kbit`
      
      * change methods names
      
      * better checks
      
      * oops
      
      * oops
      
      * address final comments
      
      ---------
      Co-authored-by: default avataryounesbelkada <younesbelkada@gmail.com>
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      9d73b922
    • Wang, Yi's avatar
  5. 23 May, 2023 6 commits