1. 18 May, 2023 1 commit
  2. 04 May, 2023 1 commit
  3. 24 Apr, 2023 1 commit
  4. 21 Apr, 2023 1 commit
  5. 20 Apr, 2023 1 commit
  6. 17 Apr, 2023 1 commit
  7. 04 Apr, 2023 1 commit
  8. 23 Mar, 2023 2 commits
  9. 22 Mar, 2023 1 commit
    • Luc CAILLIAU's avatar
      Chunkable token classification pipeline (#21771) · d62e7d88
      Luc CAILLIAU authored
      
      
      * Chunkable classification pipeline 
      
      The TokenClassificationPipeline is now able to process sequences longer than 512. No matter the framework, the model, the tokenizer. We just have to pass process_all=True and a stride number (optional). The behavior remains the same if you don't pass these optional parameters. For overlapping parts when using stride above 0, we consider only the max scores for each overlapped token in all chunks where the token is.
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * update with latest black format
      
      * update black format
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * format correction
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update comments
      
      * Update src/transformers/pipelines/token_classification.py
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      
      * Update token_classification.py
      
      Correct spaces, remove process_all and keep only stride. If stride is provided, the pipeline is applied to the whole text.
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update chunk aggregation
      
      Update the chunk aggregation strategy based on entities aggregation.
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      Remove unnecessary pop from outputs dict
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update token_classification.py
      
      * Update src/transformers/pipelines/token_classification.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * add chunking tests
      
      * correct formating
      
      * correct formatting
      
      * correct model id for test chunking
      
      * update scores with nested simplify
      
      * Update test_pipelines_token_classification.py
      
      * Update test_pipelines_token_classification.py
      
      * update model to a tiny one
      
      * Update test_pipelines_token_classification.py
      
      * Adding smaller test for chunking.
      
      * Fixup
      
      * Update token_classification.py
      
      * Update src/transformers/pipelines/token_classification.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/pipelines/token_classification.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      d62e7d88
  10. 16 Mar, 2023 1 commit
  11. 15 Mar, 2023 1 commit
  12. 09 Mar, 2023 1 commit
  13. 08 Mar, 2023 1 commit
  14. 02 Mar, 2023 2 commits
  15. 28 Feb, 2023 1 commit
    • Yih-Dar's avatar
      馃敟Rework pipeline testing by removing `PipelineTestCaseMeta` 馃殌 (#21516) · 871c31a6
      Yih-Dar authored
      
      
      * Add PipelineTesterMixin
      
      * remove class PipelineTestCaseMeta
      
      * move validate_test_components
      
      * Add for ViT
      
      * Add to SPECIAL_MODULE_TO_TEST_MAP
      
      * style and quality
      
      * Add feature-extraction
      
      * update
      
      * raise instead of skip
      
      * add tiny_model_summary.json
      
      * more explicit
      
      * skip tasks not in mapping
      
      * add availability check
      
      * Add Copyright
      
      * A way to diable irrelevant tests
      
      * update with main
      
      * remove disable_irrelevant_tests
      
      * skip tests
      
      * better skip message
      
      * better skip message
      
      * Add all pipeline task tests
      
      * revert
      
      * Import PipelineTesterMixin
      
      * subclass test classes with PipelineTesterMixin
      
      * Add pipieline_model_mapping
      
      * Fix import after adding pipieline_model_mapping
      
      * Fix style and quality after adding pipieline_model_mapping
      
      * Fix one more import after adding pipieline_model_mapping
      
      * Fix style and quality after adding pipieline_model_mapping
      
      * Fix test issues
      
      * Fix import requirements
      
      * Fix mapping for MobileViTModelTest
      
      * Update
      
      * Better skip message
      
      * pipieline_model_mapping could not be None
      
      * Remove some PipelineTesterMixin
      
      * Fix typo
      
      * revert tests_fetcher.py
      
      * update
      
      * rename
      
      * revert
      
      * Remove PipelineTestCaseMeta from ZeroShotAudioClassificationPipelineTests
      
      * style and quality
      
      * test fetcher for all pipeline/model tests
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      871c31a6
  16. 27 Feb, 2023 1 commit
  17. 24 Feb, 2023 1 commit
  18. 22 Feb, 2023 1 commit
  19. 21 Feb, 2023 1 commit
  20. 16 Feb, 2023 1 commit
  21. 15 Feb, 2023 1 commit
  22. 10 Feb, 2023 1 commit
  23. 06 Feb, 2023 1 commit
    • Sylvain Gugger's avatar
      Update quality tooling for formatting (#21480) · 6f79d264
      Sylvain Gugger authored
      * Result of black 23.1
      
      * Update target to Python 3.7
      
      * Switch flake8 to ruff
      
      * Configure isort
      
      * Configure isort
      
      * Apply isort with line limit
      
      * Put the right black version
      
      * adapt black in check copies
      
      * Fix copies
      6f79d264
  24. 02 Feb, 2023 1 commit
  25. 30 Jan, 2023 1 commit
  26. 25 Jan, 2023 3 commits
  27. 23 Jan, 2023 1 commit
  28. 20 Jan, 2023 1 commit
  29. 19 Jan, 2023 1 commit
    • Arthur's avatar
      [Whisper] Fix timestamp processor (#21187) · e9b4800d
      Arthur authored
      
      
      * add draft logit processor
      
      * add template functions
      
      * update timesapmt processor parameters
      
      * draft script
      
      * simplify code
      
      * cleanup
      
      * fixup and clean
      
      * update pipeline
      
      * style
      
      * clean up previous idea
      
      * add tokenization utils
      
      * update tokenizer and asr output
      
      * fit whisper type
      
      * style and update test
      
      * clean test
      
      * style test
      
      * update tests
      
      * update error test
      
      * udpate code (not based on review yet)
      
      * update tokenization
      
      * update asr pipeline
      
      * update code
      
      * cleanup and update test
      
      * fmt
      
      * remove text verificatino
      
      * cleanup
      
      * cleanup
      
      * add model test
      
      * update tests
      
      * update code add docstring
      
      * update code and add docstring
      
      * fix pipeline tests
      
      * add draft logit processor
      
      add template functions
      
      update timesapmt processor parameters
      
      draft script
      
      simplify code
      
      cleanup
      
      fixup and clean
      
      update pipeline
      
      style
      
      clean up previous idea
      
      add tokenization utils
      
      update tokenizer and asr output
      
      fit whisper type
      
      style and update test
      
      clean test
      
      style test
      
      update tests
      
      update error test
      
      udpate code (not based on review yet)
      
      update tokenization
      
      update asr pipeline
      
      update code
      
      cleanup and update test
      
      fmt
      
      remove text verificatino
      
      cleanup
      
      cleanup
      
      add model test
      
      update tests
      
      update code add docstring
      
      update code and add docstring
      
      fix pipeline tests
      
      * Small update.
      
      * Fixup.
      
      * Tmp.
      
      * More support.
      
      * Making `forced_decoder_ids` non mandatory for users to set.
      
      * update and fix first bug
      
      * properly process sequence right after merge if last
      
      * tofo
      
      * allow list inputs + compute begin index better
      
      * start adding tests
      
      * add the 3 edge cases
      
      * style
      
      * format sequences
      
      * fixup
      
      * update
      
      * update
      
      * style
      
      * test passes, edge cases should be good
      
      * update last value
      
      * remove Trie
      
      * update tests and expec ted values
      
      * handle bigger chunk_length
      
      * clean tests a bit
      
      * refactor chunk iter and clean pipeline
      
      * update tests
      
      * style
      
      * refactor chunk iter and clean pipeline
      
      * upade
      
      * resolve comments
      
      * Apply suggestions from code review
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      
      * take stride right into account
      
      * update test expected values
      
      * Update code based on review
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      
      * major refactor
      
      * add correct strides for tests
      
      * Update src/transformers/pipelines/automatic_speech_recognition.py
      
      * fix whisper timestamp test
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      e9b4800d
  30. 18 Jan, 2023 1 commit
  31. 17 Jan, 2023 1 commit
    • Arthur's avatar
      Whisper Timestamp processor and prediction (#20620) · bb300ac6
      Arthur authored
      
      
      * add draft logit processor
      
      * add template functions
      
      * update timesapmt processor parameters
      
      * draft script
      
      * simplify code
      
      * cleanup
      
      * fixup and clean
      
      * update pipeline
      
      * style
      
      * clean up previous idea
      
      * add tokenization utils
      
      * update tokenizer and asr output
      
      * fit whisper type
      
      * style and update test
      
      * clean test
      
      * style test
      
      * update tests
      
      * update error test
      
      * udpate code (not based on review yet)
      
      * update tokenization
      
      * update asr pipeline
      
      * update code
      
      * cleanup and update test
      
      * fmt
      
      * remove text verificatino
      
      * cleanup
      
      * cleanup
      
      * add model test
      
      * update tests
      
      * update code add docstring
      
      * update code and add docstring
      
      * fix pipeline tests
      
      * add draft logit processor
      
      add template functions
      
      update timesapmt processor parameters
      
      draft script
      
      simplify code
      
      cleanup
      
      fixup and clean
      
      update pipeline
      
      style
      
      clean up previous idea
      
      add tokenization utils
      
      update tokenizer and asr output
      
      fit whisper type
      
      style and update test
      
      clean test
      
      style test
      
      update tests
      
      update error test
      
      udpate code (not based on review yet)
      
      update tokenization
      
      update asr pipeline
      
      update code
      
      cleanup and update test
      
      fmt
      
      remove text verificatino
      
      cleanup
      
      cleanup
      
      add model test
      
      update tests
      
      update code add docstring
      
      update code and add docstring
      
      fix pipeline tests
      
      * Small update.
      
      * Fixup.
      
      * Tmp.
      
      * More support.
      
      * Making `forced_decoder_ids` non mandatory for users to set.
      
      * update and fix first bug
      
      * properly process sequence right after merge if last
      
      * tofo
      
      * allow list inputs + compute begin index better
      
      * start adding tests
      
      * add the 3 edge cases
      
      * style
      
      * format sequences
      
      * fixup
      
      * update
      
      * update
      
      * style
      
      * test passes, edge cases should be good
      
      * update last value
      
      * remove Trie
      
      * update tests and expec ted values
      
      * handle bigger chunk_length
      
      * clean tests a bit
      
      * refactor chunk iter and clean pipeline
      
      * update tests
      
      * style
      
      * refactor chunk iter and clean pipeline
      
      * upade
      
      * resolve comments
      
      * Apply suggestions from code review
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      
      * take stride right into account
      
      * update test expected values
      
      * Update code based on review
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      Co-authored-by: default avatarNicolas Patry <patry.nicolas@protonmail.com>
      Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
      bb300ac6
  32. 16 Jan, 2023 1 commit
    • Nicolas Patry's avatar
      Fixing batching pipelines on single items for ChunkPipeline (#21132) · 488a179c
      Nicolas Patry authored
      * Fixing #20783
      
      * Update src/transformers/pipelines/base.py
      
      * Fixing some tests.
      
      * Fixup.
      
      * Remove ffmpeg dep + a bit more relaxed for bigbird QA precision.
      
      * Better dataset.
      
      * Prevent failing on TF.
      
      * Better condition. We can't use `can_use_iterator` since we cannot use it
      directly.
      488a179c
  33. 12 Jan, 2023 1 commit
  34. 09 Jan, 2023 1 commit
  35. 08 Jan, 2023 1 commit
  36. 03 Jan, 2023 1 commit