1. 05 Oct, 2023 1 commit
  2. 04 Oct, 2023 1 commit
  3. 03 Oct, 2023 1 commit
  4. 18 Sep, 2023 1 commit
    • Arthur's avatar
      🚨🚨 🚨🚨 [`Tokenizer`] attemp to fix add_token issues🚨🚨 🚨🚨 (#23909) · 2da88537
      Arthur authored
      
      
      * fix test for bart. Order is correct now let's skip BPEs
      
      * ouf
      
      * styling
      
      * fix bert....
      
      * slow refactoring
      
      * current updates
      
      * massive refactoring
      
      * update
      
      * NICE!
      
      * update to see where I am at
      
      * updates
      
      * update
      
      * update
      
      * revert
      
      * updates
      
      * updates
      
      * start supporting legacy_save
      
      * styling
      
      * big update
      
      * revert some changes
      
      * nits
      
      * nniiiiiice
      
      * small fixes
      
      * kinda fix t5 with new behaviour
      
      * major update
      
      * fixup
      
      * fix copies
      
      * today's updates
      
      * fix byt5
      
      * upfate
      
      * update
      
      * update
      
      * updates
      
      * update vocab size test
      
      * Barthez does not use not need the fairseq offset ids
      
      * super calll must be after
      
      * calll super
      
      * move all super init
      
      * move other super init
      
      * fixup
      
      * nits
      
      * more fixes
      
      * nits
      
      * more fixes
      
      * nits
      
      * more fix
      
      * remove useless files
      
      * ouch all of them are affected
      
      * and more!
      
      * small imporvements
      
      * no more sanitize token
      
      * more changes around unique no split tokens
      
      * partially fix more things
      
      * keep legacy save but add warning
      
      * so... more fixes
      
      * updates
      
      * guess deberta tokenizer could be nuked
      
      * fixup
      
      * fixup did some bad things
      
      * nuke it if it breaks
      
      * remove prints and pretrain fast from slow with new format.
      
      * fixups
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * fiou
      
      * nit
      
      * by default specials should not be normalized?
      
      * update
      
      * remove brakpoint
      
      * updates
      
      * a lot of updates
      
      * fixup
      
      * fixes revert some changes to match fast
      
      * small nits
      
      * that makes it cleaner
      
      * fix camembert accordingly
      
      * update
      
      * some lest breaking changes
      
      * update
      
      * fixup
      
      * fix byt5 and whisper mostly
      
      * some more fixes, canine's byte vocab
      
      * fix gpt2
      
      * fix most of the perceiver tests (4 left)
      
      * fix layout lmv3
      
      * fixup
      
      * fix copies for gpt2 style
      
      * make sure to only warn once
      
      * fix perciever and gpt2 tests
      
      * some more backward compatibility: also read special tokens map because some ppl use it........////.....
      
      * fixup
      
      * add else when reading
      
      * nits
      
      * fresh updates
      
      * fix copies
      
      * will this make everything faster?
      
      * fixes
      
      * more fixes
      
      * update
      
      * more fixes
      
      * fixup
      
      * is the source of truth right?
      
      * sorry camembert for the troubles
      
      * current updates
      
      * fixup
      
      * update led
      
      * update
      
      * fix regression
      
      * fix single word
      
      * more model specific fixes
      
      * fix t5 tests
      
      * fixup
      
      * more comments
      
      * update
      
      * fix nllb
      
      * rstrip removed
      
      * small fixes
      
      * better handle additional_special_tokens and vocab sizes
      
      * fixing
      
      * styling
      
      * fix 4 / 21
      
      * fixup
      
      * fix nlbb's tests
      
      * some fixes
      
      * fix t5
      
      * fixes
      
      * style
      
      * fix canine tests
      
      * damn this is nice
      
      * nits
      
      * m2m100 nit
      
      * fixups
      
      * fixes!
      
      * fixup
      
      * stash
      
      * fix merge
      
      * revert bad change
      
      * fixup
      
      * correct order for code Llama
      
      * fix speecht5 post merge
      
      * styling
      
      * revert source of 11 fails
      
      * small nits
      
      * all changes in one go
      
      * fnet hack
      
      * fix 2 more tests
      
      * update based on main branch of tokenizers
      
      * fixup
      
      * fix VITS issues
      
      * more fixes
      
      * fix mgp test
      
      * fix camembert issues
      
      * oups camembert still has 2 failing tests
      
      * mluke fixes
      
      * decode fixes
      
      * small nits
      
      * nits
      
      * fix llama and vits
      
      * fix camembert
      
      * smal nits
      
      * more fixes when initialising a fast from a slow and etc
      
      * fix one of the last test
      
      * fix CPM tokenizer test
      
      * fixups
      
      * fix pop2piano
      
      * fixup
      
      * ️ Change tokenizers required version ️
      
      * ️ Change tokenizers required version ️
      
      * "tokenizers>=0.14,<0.15", don't forget smaller than
      
      * fix musicgen tests and pretraiendtokenizerfast
      
      * fix owlvit and all
      
      * update t5
      
      * fix 800 red
      
      * fix tests
      
      * fix the fix of the fix of t5
      
      * styling
      
      * documentation nits
      
      * cache _added_tokens_encoder
      
      * fixups
      
      * Nit
      
      * fix red tests
      
      * one last nit!
      
      * make eveything a lot simpler
      
      * Now it's over 😉
      
      
      
      * few small nits
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * updates that work for now
      
      * tests that should no be skipped / changed and fixed next
      
      * fixup
      
      * i am ashamed
      
      * pushe the fix
      
      * update
      
      * fixups
      
      * nits
      
      * fix added_tokens_encoder
      
      * fix canine test
      
      * fix pegasus vocab
      
      * fix transfoXL
      
      * fixup
      
      * whisper needs to be fixed for train new
      
      * pegasus nits
      
      * more pegasus fixes
      
      * minor update
      
      * better error message in failed test
      
      * fix whisper failing test
      
      * fix whisper failing test
      
      * fix pegasus
      
      * fixup
      
      * fix **** pegasus
      
      * reset things
      
      * remove another file
      
      * attempts to fix the strange custome encoder and offset
      
      * nits here and there
      
      * update
      
      * fixup
      
      * nit
      
      * fix the whisper test
      
      * nits nits
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * updates based on review
      
      * some small update to potentially remove
      
      * nits
      
      * import rlu cache
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarLysandre Debut <hi@lysand.re>
      
      * move warning to `from_pretrained`
      
      * update tests results now that the special tokens are always added
      
      ---------
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      Co-authored-by: default avatarLysandre Debut <hi@lysand.re>
      2da88537
  5. 05 Sep, 2023 1 commit
  6. 11 Aug, 2023 1 commit
  7. 02 Aug, 2023 2 commits
  8. 30 Jun, 2023 1 commit
    • Matt's avatar
      Speed up TF tests by reducing hidden layer counts (#24595) · 134caef3
      Matt authored
      * hidden layers, huh, what are they good for (absolutely nothing)
      
      * Some tests break with 1 hidden layer, use 2
      
      * Use 1 hidden layer in a few slow models
      
      * Use num_hidden_layers=2 everywhere
      
      * Slightly higher tol for groupvit
      
      * Slightly higher tol for groupvit
      134caef3
  9. 27 Jun, 2023 1 commit
  10. 20 Jun, 2023 2 commits
  11. 12 Jun, 2023 1 commit
  12. 07 Jun, 2023 1 commit
  13. 02 Jun, 2023 1 commit
  14. 24 May, 2023 1 commit
    • Matt's avatar
      Better TF docstring types (#23477) · f8b25744
      Matt authored
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Rework TF type hints to use | None instead of Optional[] for tf.Tensor
      
      * Don't forget the imports
      
      * Add the imports to tests too
      
      * make fixup
      
      * Refactor tests that depended on get_type_hints
      
      * Better test refactor
      
      * Fix an old hidden bug in the test_keras_fit input creation code
      
      * Fix for the Deit tests
      f8b25744
  15. 22 May, 2023 1 commit
  16. 26 Apr, 2023 1 commit
    • Ritik Nandwal's avatar
      Add TensorFlow Wav2Vec2 for sequence classification (#22073) · 20ac86c6
      Ritik Nandwal authored
      * Add initial changes for TF wav2vec2 for sequence classification
      
      * Add suggested changes
      
      * Add serving and serving output methods
      
      * Add serving_output implementation and fix layer_weights
      
      * Add fixes
      
      * Fixed test cases
      
      * Fixing test and adding suggested changes
      20ac86c6
  17. 04 Apr, 2023 1 commit
    • Matt's avatar
      Fix inverted conditional in TF common test! (#22540) · edb704b2
      Matt authored
      * Fix inverted conditional in TF common test!
      
      * Make the same change in the PT tests file
      
      * Make sure hidden states for GPT2 have the same output shape in PT/TF
      
      * Minor fix to PT implementation of token classification loss
      
      * Skip loss equivalence test for TFHubert because it keeps overflowing to inf
      
      * Compute LM loss for TF the (weird) way it's computed in PT
      
      * Skip loss equivalence test for Wav2Vec2 for the same reason as Hubert
      
      * Fix - don't try to access the hidden states property when output is a tuple
      edb704b2
  18. 21 Mar, 2023 1 commit
  19. 01 Mar, 2023 1 commit
    • Arthur's avatar
      prepare for "__floordiv__ is deprecated and its behavior will change in a... · 44e3e3fb
      Arthur authored
      prepare for "__floordiv__ is deprecated  and its behavior will change in a future version of pytorch" (#20211)
      
      * rounding_mode = "floor"  instead of // to prevent behavioral change
      
      * add other TODO
      
      * use `torch_int_div` from pytrch_utils
      
      * same for tests
      
      * fix copies
      
      * style
      
      * use relative imports when needed
      
      * Co-authored-by: sgugger <sylvain.gugger@gmail.com>
      44e3e3fb
  20. 28 Feb, 2023 1 commit
    • Yih-Dar's avatar
      🔥Rework pipeline testing by removing `PipelineTestCaseMeta` 🚀 (#21516) · 871c31a6
      Yih-Dar authored
      
      
      * Add PipelineTesterMixin
      
      * remove class PipelineTestCaseMeta
      
      * move validate_test_components
      
      * Add for ViT
      
      * Add to SPECIAL_MODULE_TO_TEST_MAP
      
      * style and quality
      
      * Add feature-extraction
      
      * update
      
      * raise instead of skip
      
      * add tiny_model_summary.json
      
      * more explicit
      
      * skip tasks not in mapping
      
      * add availability check
      
      * Add Copyright
      
      * A way to diable irrelevant tests
      
      * update with main
      
      * remove disable_irrelevant_tests
      
      * skip tests
      
      * better skip message
      
      * better skip message
      
      * Add all pipeline task tests
      
      * revert
      
      * Import PipelineTesterMixin
      
      * subclass test classes with PipelineTesterMixin
      
      * Add pipieline_model_mapping
      
      * Fix import after adding pipieline_model_mapping
      
      * Fix style and quality after adding pipieline_model_mapping
      
      * Fix one more import after adding pipieline_model_mapping
      
      * Fix style and quality after adding pipieline_model_mapping
      
      * Fix test issues
      
      * Fix import requirements
      
      * Fix mapping for MobileViTModelTest
      
      * Update
      
      * Better skip message
      
      * pipieline_model_mapping could not be None
      
      * Remove some PipelineTesterMixin
      
      * Fix typo
      
      * revert tests_fetcher.py
      
      * update
      
      * rename
      
      * revert
      
      * Remove PipelineTestCaseMeta from ZeroShotAudioClassificationPipelineTests
      
      * style and quality
      
      * test fetcher for all pipeline/model tests
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      871c31a6
  21. 22 Feb, 2023 1 commit
  22. 15 Feb, 2023 2 commits
  23. 14 Feb, 2023 2 commits
  24. 13 Feb, 2023 2 commits
  25. 10 Feb, 2023 1 commit
  26. 06 Feb, 2023 1 commit
    • Sylvain Gugger's avatar
      Update quality tooling for formatting (#21480) · 6f79d264
      Sylvain Gugger authored
      * Result of black 23.1
      
      * Update target to Python 3.7
      
      * Switch flake8 to ruff
      
      * Configure isort
      
      * Configure isort
      
      * Apply isort with line limit
      
      * Put the right black version
      
      * adapt black in check copies
      
      * Fix copies
      6f79d264
  27. 27 Dec, 2022 1 commit
  28. 10 Nov, 2022 1 commit
    • Sanchit Gandhi's avatar
      [processor] Add 'model input names' property (#20117) · 905e5773
      Sanchit Gandhi authored
      * [processor] Add 'model input names' property
      
      * add test
      
      * no f string
      
      * add generic property method to mixin
      
      * copy to multimodal
      
      * copy to vision
      
      * tests for all audio
      
      * remove ad-hoc tests
      
      * style
      
      * fix flava test
      
      * fix test
      
      * fix processor code
      905e5773
  29. 27 Oct, 2022 1 commit
  30. 18 Oct, 2022 2 commits
  31. 12 Oct, 2022 1 commit
  32. 16 Sep, 2022 1 commit
  33. 12 Sep, 2022 1 commit
  34. 09 Sep, 2022 1 commit
    • Matt's avatar
      Fix train_step, test_step and tests for CLIP (#18684) · 660e0b97
      Matt authored
      
      
      * Fix train_step and test_step, correctly enable CLIP fit test
      
      * Stop using get_args on older Python versions
      
      * Don't use get_origin either
      
      * UnionType is actually even newer, don't use that either
      
      * Apply the same fix to test_loss_computation
      
      * Just realized I was accidentally skipping a bunch of tests!
      
      * Fix test_loss_computation for models without separable labels
      
      * Fix scalar losses in test_step and train_step
      
      * Stop committing your breakpoints
      
      * Fix Swin loss shape
      
      * Fix Tapas loss shape
      
      * Shape fixes for TAPAS, DeIT, HuBERT and ViTMAE
      
      * Add loss computation to TFMobileBertForPreTraining
      
      * make fixup and move copied from statement
      
      * make fixup and move copied from statement
      
      * Correct copied from
      
      * Add labels and next_sentence_label inputs to TFMobileBERT
      
      * Make sure total_loss is always defined
      
      * Update tests/test_modeling_tf_common.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Fix copied from
      
      * Ensure CTC models get labels in tests
      
      * Ensure CTC models get labels in tests
      
      * Fix tests for vit_mae
      
      * Fix tests for vit_mae
      
      * Fix tests for vit_mae
      
      * Reduce batch size for wav2vec2 testing because it was causing OOM
      
      * Skip some TAPAS tests that are failing
      
      * Skip a failing HuBERT test
      
      * make style
      
      * Fix mobilebertforpretraining test
      
      * Skip Wav2Vec2 tests that use huge amounts of mem
      
      * Skip keras_fit for Wav2Vec2 as well
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      660e0b97