1. 30 Sep, 2022 1 commit
  2. 02 Sep, 2022 1 commit
  3. 19 Aug, 2022 1 commit
  4. 12 Aug, 2022 1 commit
  5. 23 Jul, 2022 1 commit
  6. 28 Jun, 2022 1 commit
  7. 21 Jun, 2022 1 commit
  8. 10 Jun, 2022 1 commit
  9. 19 May, 2022 1 commit
  10. 12 May, 2022 1 commit
  11. 11 Apr, 2022 1 commit
  12. 16 Mar, 2022 1 commit
  13. 07 Mar, 2022 1 commit
  14. 04 Mar, 2022 1 commit
    • Chan Woo Kim's avatar
      Constrained Beam Search [*With* Disjunctive Decoding] (#15761) · 5c6f57ee
      Chan Woo Kim authored
      
      
      * added classes to get started with constrained beam search
      
      * in progress, think i can directly force tokens now but not yet with the round robin
      
      * think now i have total control, now need to code the bank selection
      
      * technically works as desired, need to optimize and fix design choices leading to undersirable outputs
      
      * complete PR #1 without disjunctive decoding
      
      * removed incorrect tests
      
      * Delete k.txt
      
      * Delete test.py
      
      * Delete test.sh
      
      * revert changes to test scripts
      
      * genutils
      
      * full implementation with testing, no disjunctive yet
      
      * shifted docs
      
      * passing all tests realistically ran locally
      
      * removing accidentally included print statements
      
      * fixed source of error in initial PR test
      
      * fixing the get_device() vs device trap
      
      * fixed documentation docstrings about constrained_beam_search
      
      * fixed tests having failing for Speech2TextModel's floating point inputs
      
      * fix cuda long tensor
      
      * added examples and testing for them and founx & fixed a bug in beam_search and constrained_beam_search
      
      * deleted accidentally added test halting code with assert False
      
      * code reformat
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      
      * fixing based on comments on PR
      
      * took out the testing code that should but work fails without the beam search moditification ; style changes
      
      * fixing comments issues
      
      * docstrings for ConstraintListState
      
      * typo in PhrsalConstraint docstring
      
      * docstrings improvements
      
      * finished adding what is sort of an opinionated implementation of disjunctive generation, but it revealed errors in inner beam search logic during testing.
      
      * fixed bug found in constrained beam search that used beam_idx that were not global across all the batches
      
      * disjunctive constraint working 100% correctly
      
      * passing all tests
      
      * Accidentally included mlruns
      
      * Update src/transformers/generation_beam_constraints.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update src/transformers/generation_beam_constraints.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * complete overhaul of type complexities and other nits
      
      * strict type checks in generate()
      
      * fixing second round of feedback by narsil
      
      * fixed failing generation test because of type check overhaul
      
      * generation test fail fix
      
      * fixing test fails
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      5c6f57ee
  15. 23 Feb, 2022 1 commit
  16. 09 Feb, 2022 1 commit
    • Chan Woo Kim's avatar
      Constrained Beam Search [without disjunctive decoding] (#15416) · 2b5603f6
      Chan Woo Kim authored
      
      
      * added classes to get started with constrained beam search
      
      * in progress, think i can directly force tokens now but not yet with the round robin
      
      * think now i have total control, now need to code the bank selection
      
      * technically works as desired, need to optimize and fix design choices leading to undersirable outputs
      
      * complete PR #1 without disjunctive decoding
      
      * removed incorrect tests
      
      * Delete k.txt
      
      * Delete test.py
      
      * Delete test.sh
      
      * revert changes to test scripts
      
      * genutils
      
      * full implementation with testing, no disjunctive yet
      
      * shifted docs
      
      * passing all tests realistically ran locally
      
      * removing accidentally included print statements
      
      * fixed source of error in initial PR test
      
      * fixing the get_device() vs device trap
      
      * fixed documentation docstrings about constrained_beam_search
      
      * fixed tests having failing for Speech2TextModel's floating point inputs
      
      * fix cuda long tensor
      
      * added examples and testing for them and founx & fixed a bug in beam_search and constrained_beam_search
      
      * deleted accidentally added test halting code with assert False
      
      * code reformat
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update tests/test_generation_utils.py
      
      * fixing based on comments on PR
      
      * took out the testing code that should but work fails without the beam search moditification ; style changes
      
      * fixing comments issues
      
      * docstrings for ConstraintListState
      
      * typo in PhrsalConstraint docstring
      
      * docstrings improvements
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      2b5603f6
  17. 24 Jan, 2022 1 commit
    • Patrick von Platen's avatar
      [Beam Search] Correct returned beam scores (#14654) · 8d6acc6c
      Patrick von Platen authored
      * better
      
      * save intermediate
      
      * finish code
      
      * up
      
      * docs
      
      * Apply suggestions from code review
      
      * up
      
      * add compute transition  beam scores function to model and make sure scores are correct with eos
      
      * apply nicos comments
      
      * Apply suggestions from code review
      
      * another fix
      8d6acc6c
  18. 30 Dec, 2021 1 commit
  19. 23 Dec, 2021 1 commit
  20. 21 Dec, 2021 1 commit
  21. 17 Dec, 2021 2 commits
  22. 16 Dec, 2021 1 commit
  23. 19 Nov, 2021 1 commit
  24. 08 Oct, 2021 1 commit
  25. 11 Jun, 2021 1 commit
  26. 10 Jun, 2021 1 commit
  27. 27 May, 2021 1 commit
    • Nicolas Patry's avatar
      Adding new argument `max_new_tokens` for generate. (#11476) · 80d712fa
      Nicolas Patry authored
      * Adding new argument `max_new_tokens` for generate.
      
      This is a proposal to add a new argument `max_new_tokens` to `generate`.
      This include a `MaxNewTokensCriteria` that enables callers that don't
      know about the token length ahead (like pipelines callers) to manage
      more easily the length of their generated output.
      
      * Adding a test for the user warning when both`max_length` and
      `max_new_tokens` are used together.
      
      * Removed redundant `no_grad`.
      80d712fa
  28. 19 May, 2021 1 commit
  29. 18 May, 2021 1 commit
    • Daniel Stancl's avatar
      Fix usage of head masks by PT encoder-decoder models' `generate()` function (#11621) · 680d181c
      Daniel Stancl authored
      * Add missing head masking for generate() function
      
      * Add head_mask, decoder_head_mask and cross_attn_head_mask
      into prepare_inputs_for_generation for generate() function
      for multiple encoder-decoder models.
      
      * Add test_genereate_with_head_masking
      
      * [WIP] Update the new test and handle special cases
      
      * make style
      
      * Omit ProphetNet test so far
      
      * make fix-copies
      680d181c
  30. 07 May, 2021 1 commit
    • Vasudev Gupta's avatar
      Add BigBirdPegasus (#10991) · dc3f6758
      Vasudev Gupta authored
      
      
      * init bigbird pegasus
      
      * add debugging nb ; update config
      
      * init conversion
      
      * update conversion script
      
      * complete conversion script
      
      * init forward()
      
      * complete forward()
      
      * add tokenizer
      
      * add some slow tests
      
      * commit current
      
      * fix copies
      
      * add docs
      
      * add conversion script for bigbird-roberta-summarization
      
      * remove TODO
      
      * small fixups
      
      * correct tokenizer
      
      * add bigbird core for now
      
      * fix config
      
      * fix more
      
      * revert pegasus-tokenizer back
      
      * make style
      
      * everything working for pubmed; yayygit status
      
      * complete tests finally
      
      * remove bigbird pegasus tok
      
      * correct tokenizer
      
      * correct tests
      
      * add tokenizer files
      
      * finish make style
      
      * fix test
      
      * update
      
      * make style
      
      * fix tok utils base file
      
      * make fix-copies
      
      * clean a bit
      
      * small update
      
      * fix some suggestions
      
      * add to readme
      
      * fix a bit, clean tests
      
      * fix more tests
      
      * Update src/transformers/__init__.py
      
      * Update src/transformers/__init__.py
      
      * make fix-copies
      
      * complete attn switching, auto-padding left
      
      * make style
      
      * fix auto-padding test
      
      * make style
      
      * fix batched attention tests
      
      * put tolerance at 1e-1 for stand-alone decoder test
      
      * fix docs
      
      * fix tests
      
      * correct slow tokenizer conversion
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * complete remaining suggestions
      
      * fix test
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      dc3f6758
  31. 26 Apr, 2021 1 commit
  32. 21 Apr, 2021 1 commit
    • Nicolas Patry's avatar
      Removed `max_length` from being mandatory within `generate`. (#11314) · aad95c7c
      Nicolas Patry authored
      * Removed `max_length` from being mandatory within `generate`.
      
      - Moving on to fully using `StoppingCriteria` for `greedy` and `sample`
      modes.
      - `max_length` still used for `beam_search` and `group_beam_search`
      (Follow up PR)
      - Fixes a bug with MaxLengthStoppingCriteria (we should stop as soon a
      we hit the max_length, the comparison needs to be or equal, that affects
      the tests).
      - Added options to use `logits_processor` and `stopping_criteria`
      directly within `generate` function (so some users can define their own
      `logits_processor` and `stopping_criteria`).
      - Modified the backward compat tests to make sure we issue a warning.
      
      * Fix `max_length` argument in `generate`.
      
      * Moving validate to being functional.
      
      - Renamed `smax_length` to `stoppping_max_length`.
      
      * Removing `logits_processor` and `stopping_criteria` from `generate`
      arguments.
      
      * Deepcopy.
      
      * Fix global variable name.
      aad95c7c
  33. 22 Mar, 2021 1 commit
  34. 12 Mar, 2021 1 commit
    • Nicolas Patry's avatar
      Adding new parameter to `generate`: `max_time`. (#9846) · 543d0549
      Nicolas Patry authored
      * [WIP] Adding new parameter to `generate`:  `max_time`.
      
      Generation by tokens number is sometimes a bit clunky because we don't
      know how many tokens are good enough or even how many tokens are in
      the payload (for pipelines users for instance). This leads to hard
      to understand behavior.
      
      This PR proposes a new argument `max_time` which is a float of seconds
      for the allowed time for `generate` to run on.
      Ideally combinations of `max_tokens=None`, `max_time=2` could be used to
      generate as many tokens as possible within time budget.
      
      NB: Another possible approach consists of passing a callback to `generate`
        putting the caller in charge of the actual decision of when to stop
        generating tokens. It opens the door to 'which args should we pass'
        to this callback. It's hard to imagine other use-cases for this
        early stopping behavior than time (that are not already covered by
        parameters of generate)
      
      * Revamp with StoppingCriteria
      
      * Removing deprecated mentions.
      
      * Forgot arguments to stopping criteria.
      
      * Readding max_length it's not just used as a stopping criteria.
      
      * Default value for `stopping_criteria`.
      
      * Address @patrickvonplaten comments.
      
      - More docstrings
      - Actual doc
      - Include in global namespace
      - Remove TF work.
      
      * Put back `max_length` (deprecation different PR).
      
      * Doc quality.
      
      * Fixing old behavior without `stopping_criteria` but with `max_length`.
      
      Making sure we don't break that in the future.
      
      * Adding more tests for possible inconsistencies between
      
      `max_length` and `stopping_criteria`.
      
      * Fixing the torch imports.
      543d0549
  35. 10 Mar, 2021 1 commit
    • Suraj Patil's avatar
      Speech2TextTransformer (#10175) · d26b37e7
      Suraj Patil authored
      
      
      * s2t
      
      * fix config
      
      * conversion script
      
      * fix import
      
      * add tokenizer
      
      * fix tok init
      
      * fix tokenizer
      
      * first version working
      
      * fix embeds
      
      * fix lm head
      
      * remove extra heads
      
      * fix convert script
      
      * handle encoder attn mask
      
      * style
      
      * better enc attn mask
      
      * override _prepare_attention_mask_for_generation
      
      * handle attn_maks in encoder and decoder
      
      * input_ids => input_features
      
      * enable use_cache
      
      * remove old code
      
      * expand embeddings if needed
      
      * remove logits bias
      
      * masked_lm_loss => loss
      
      * hack tokenizer to support feature processing
      
      * fix model_input_names
      
      * style
      
      * fix error message
      
      * doc
      
      * remove inputs_embeds
      
      * remove input_embeds
      
      * remove unnecessary docstring
      
      * quality
      
      * SpeechToText => Speech2Text
      
      * style
      
      * remove shared_embeds
      
      * subsample => conv
      
      * remove Speech2TextTransformerDecoderWrapper
      
      * update output_lengths formula
      
      * fix table
      
      * remove max_position_embeddings
      
      * update conversion scripts
      
      * add possibility to do upper case for now
      
      * add FeatureExtractor and Processor
      
      * add tests for extractor
      
      * require_torch_audio => require_torchaudio
      
      * add processor test
      
      * update import
      
      * remove classification head
      
      * attention mask is now 1D
      
      * update docstrings
      
      * attention mask should be of type long
      
      * handle attention mask from generate
      
      * alwyas return attention_mask
      
      * fix test
      
      * style
      
      * doc
      
      * Speech2TextTransformer => Speech2Text
      
      * Speech2TextTransformerConfig => Speech2TextConfig
      
      * remove dummy_inputs
      
      * nit
      
      * style
      
      * multilinguial tok
      
      * fix tokenizer
      
      * add tgt_lang setter
      
      * save lang_codes
      
      * fix tokenizer
      
      * add forced_bos_token_id to tokenizer
      
      * apply review suggestions
      
      * add torchaudio to extra deps
      
      * add speech deps to CI
      
      * fix dep
      
      * add libsndfile to ci
      
      * libsndfile1
      
      * add speech to extras all
      
      * libsndfile1 -> libsndfile1
      
      * libsndfile
      
      * libsndfile1-dev
      
      * apt update
      
      * add sudo to install
      
      * update deps table
      
      * install libsndfile1-dev on CI
      
      * tuple to list
      
      * init conv layer
      
      * add model tests
      
      * quality
      
      * add integration tests
      
      * skip_special_tokens
      
      * add speech_to_text_transformer in toctree
      
      * fix tokenizer
      
      * fix fp16 tests
      
      * add tokenizer tests
      
      * fix copyright
      
      * input_values => input_features
      
      * doc
      
      * add model in readme
      
      * doc
      
      * change checkpoint names
      
      * fix copyright
      
      * fix code example
      
      * add max_model_input_sizes in tokenizer
      
      * fix integration tests
      
      * add do_lower_case to tokenizer
      
      * remove clamp trick
      
      * fix "Add modeling imports here"
      
      * fix copyrights
      
      * fix tests
      
      * SpeechToTextTransformer => SpeechToText
      
      * fix naming
      
      * fix table formatting
      
      * fix typo
      
      * style
      
      * fix typos
      
      * remove speech dep from extras[testing]
      
      * fix copies
      
      * rename doc file,
      
      * put imports under is_torch_available
      
      * run feat extract tests when torch is available
      
      * dummy objects for processor and extractor
      
      * fix imports in tests
      
      * fix import in modeling test
      
      * fxi imports
      
      * fix torch import
      
      * fix imports again
      
      * fix positional embeddings
      
      * fix typo in import
      
      * adapt new extractor refactor
      
      * style
      
      * fix torchscript test
      
      * doc
      
      * doc
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * fix docs, copied from, style
      
      * fix docstring
      
      * handle imports
      
      * remove speech from all extra deps
      
      * remove s2t from seq2seq lm mapping
      
      * better names
      
      * skip training tests
      
      * add install instructions
      
      * List => Tuple
      
      * doc
      
      * fix conversion script
      
      * fix urls
      
      * add instruction for libsndfile
      
      * fix fp16 test
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      d26b37e7
  36. 03 Mar, 2021 1 commit
  37. 10 Feb, 2021 1 commit
    • Suraj Patil's avatar
      remove adjust_logits_during_generation method (#10087) · c130e67d
      Suraj Patil authored
      * add forced logits processors
      
      * delete adjust_logits method
      
      * add forced_eos_token_id argument in config
      
      * add tests for forced logits processors
      
      * update gen utils tests
      
      * add forced option to tf generate
      
      * remove adjust_logits method from tf models
      
      * update adjust_logits for marian
      
      * delete _force_token_id_to_be_generated method
      
      * style
      
      * import warnings
      
      * pass max_length to _get_logits_processor
      
      * set forced_eos_token_id to None
      
      * set forced attributes in conf utils
      
      * typo
      
      * fix rag generate
      
      * add forced_eos_token_id in rag config
      
      * remove force_bos_token_to_be_generated from BartConfig
      
      * remove _force_token_ids_generation from FSMT
      
      * nit
      
      * fix negative constant
      
      * apply suggestions from code review
      c130e67d
  38. 26 Jan, 2021 1 commit
  39. 06 Jan, 2021 1 commit