- 29 Oct, 2021 10 commits
-
-
Lysandre authored
-
Lysandre Debut authored
* Torch 1.10 * torch scatter for 1.10 * style * Skip tests ok
-
Haram Lee authored
-
Nicolas Patry authored
* Fixing image segmentation for inference mode. * Update src/transformers/pipelines/base.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
* Generalize problem_type to all classification models * Missing import * Deberta BC and fix tests * Fix template * Missing imports * Revert change to reformer test * Fix style
-
Sylvain Gugger authored
* Fix pipeline tests env and fetch * Fix quality
-
Nicolas Patry authored
* Adding `handle_long_generation` paramters for `text-generation` pipeline. * More error handling * Fixing tests by dropping tf support on this functionality, it needs `max_new_tokens` to make it possible to understand user's intent. Otherwise, `max_length` == `tokenizer.model_max_length` < input_ids.shape[0]. * Fixing doc ? * Doc ? * Remove link from doc. * Catched an issue on roberta. * Damn doc. * Non BC proposal ? * Cleaning the fix ? * Finally using only a test override. * Don't need to modify this. * Bad print.
-
Daniel Stancl authored
* Add the support for the fast (rust) implementation of BlenbderbotTokenizer * Fix a converter and a typo in a doc * Apply the patil-suraj's suggestion * (Nitpick) Fast tokenization -> Fast Tokenization in doc * Apply the SaulLu's suggestion * Apply Narsil's suggestion to fix test pipelines * Add encoder_no_repeat_ngram_size according to the Narsil's suggestion * Revert the last (unnecessary) commit * Override pipeline config for Blenderbot to allow for larger pos. emb. * make fix-copies
-
Thomas Wang authored
* Remove n_ctx from configs * Fix GPTJ and OpenAIGPT, both are acceptable breaking changes as there are no configs such that it breaks * Remove unecessary n_positions from TFOpenAIGPT
-
Nicolas Patry authored
* Tentative enabling of `batch_size` for pipelines. * Add systematic test for pipeline batching. * Enabling batch_size on almost all pipelines - Not `zero-shot` (it's already passing stuff as batched so trickier) - Not `QA` (preprocess uses squad features, we need to switch to real tensors at this boundary. * Adding `min_length_for_response` for conversational. * Making CTC, speech mappings avaiable regardless of framework. * Attempt at fixing automatic tests (ffmpeg not enabled for fast tests) * Removing ffmpeg dependency in tests. * Small fixes. * Slight cleanup. * Adding docs and adressing comments. * Quality. * Update docs/source/main_classes/pipelines.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/pipelines/question_answering.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/pipelines/zero_shot_classification.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Improving docs. * Update docs/source/main_classes/pipelines.rst Co-authored-by:
Philipp Schmid <32632186+philschmid@users.noreply.github.com> * N -> oberved_batch_size softmax trick. * Follow `padding_side`. * Supporting image pipeline batching (and padding). * Rename `unbatch` -> `loader_batch`. * unbatch_size forgot. * Custom padding for offset mappings. * Attempt to remove librosa. * Adding require_audio. * torchaudio. * Back to using datasets librosa. * Adding help to set a pad_token on the tokenizer. * Update src/transformers/pipelines/base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/pipelines/base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/pipelines/base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Quality. Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Philipp Schmid <32632186+philschmid@users.noreply.github.com>
-
- 28 Oct, 2021 11 commits
-
-
David del R铆o Medina authored
-
Patrick von Platen authored
-
Lysandre authored
-
Lysandre authored
-
NielsRogge authored
* Fix docs * Apply suggestions from review + fix bug
-
NielsRogge authored
* First draft * Make tuple output more readable * Replace assertions by value errors * Make it possible to predict_with_generate for vision and speech models * Adapt Seq2SeqTrainer to work with VisionEncoderDecoder/SpeechEncoderDecoder * Add deprecation warning * Add copied from statements to vision and speech encoder decoders * Fix failing test * Apply @patrickvonplaten's suggestion * Use reshape instead of view for consistency
-
Anton Lozhkov authored
* Fix SEW-D * Update tests * isort
-
Anton Lozhkov authored
-
NielsRogge authored
* First draft * Make style & quality * Improve conversion script * Add print statement to see actual slice * Make absolute tolerance smaller * Fix image classification models * Add post_process_semantic method * Disable padding * Improve conversion script * Rename to ForSemanticSegmentation, add integration test, remove post_process methods * Improve docs * Fix code quality * Fix feature extractor tests * Fix tests for image classification model * Delete file * Add is_torch_available to feature extractor * Improve documentation of feature extractor methods * Apply suggestions from @sgugger's code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Apply some more suggestions of code review * Rebase with master * Fix rebase issues * Make sure model only outputs hidden states when the user wants to * Apply suggestions from code review * Add pad method * Support padding of 2d images * Add print statement * Add print statement * Move padding method to SegformerFeatureExtractor * Fix issue * Add casting of segmentation maps * Add test for padding * Add small note about padding Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Stas Bekman authored
* respect dtype in _get_resized_lm_head * Update src/transformers/modeling_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * consistency Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Patrick von Platen authored
-
- 27 Oct, 2021 8 commits
-
-
Patrick von Platen authored
-
Patrick von Platen authored
* up * up * fix * up * Update examples/pytorch/test_xla_examples.py * correct labels * up * up * up * up * up * up
-
Anton Lozhkov authored
* Add conversion * Rename * Add an integration test and remove layer_norm * Remove layer_norm from the converter * wording * Fix imports
-
Lahfa Samy authored
* Replace assert of data_collator.py by ValueError * Replace assert of data_collator.py by ValueError
-
Anton Lozhkov authored
-
Patrick von Platen authored
* up * up * finish * up * final changes
-
Anton Lozhkov authored
* Add SEW CTC models * Update paths * Update paths
-
Lysandre Debut authored
-
- 26 Oct, 2021 11 commits
-
-
Kamal Raj authored
* switch to inference_mode from no_gard faster inference * added switch to support older version of pytorch
-
Emanuel Huber authored
Updated masked-language modeling examples in pytorch with convention defined by #12789
-
Matthew Goldey authored
* specify the text column name in the error message * pluralize the word fields
-
Jangwon Park authored
-
Lysandre authored
-
Jay Zhang authored
* Add symbolic function for XSoftmax op for exporting to ONNX. * Fix format issues. * Fix a CI issue relative to copies.
-
Patrick von Platen authored
* unispeech * add copy from * remove hubert copy from * finish for today * add unispeech-sat * adapt more * up * up * up * up * add modeling * add tests * up * up * finish * up * Apply suggestions from code review * up * up * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * up * up Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Patrick von Platen authored
-
Stas Bekman authored
* [megatron_gpt2] dynamic gelu, add tokenizer, save config * cleanup * Update src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * apply suggestions Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sergio Valcarcel Macua authored
* Include KerasTensor in allowed types - This allows propagating symbolic tensors through TFBert models and layers' call(), which allows converting the subclass models to functional models. * Style pass Co-authored-by:
Sergio Valcarcel Macua <sergiov@graphcore.ai> Co-authored-by:
matt <rocketknight1@gmail.com>
-
Patrick von Platen authored
[Speech Recognition] - Distributed training: Make sure vocab file removal and creation don't interfer (#14161) * up * better
-