- 17 Jul, 2021 1 commit
-
-
Tomohiro Endo authored
* Detect mismatch by analyzing config * Fix comment * Fix import * Update src/transformers/tokenization_utils_base.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * Revise based on reviews * remove kwargs * Fix exception * Fix handling exception again * Disable mismatch test in PreTrainedTokenizerFast Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com>
-
- 16 Jul, 2021 3 commits
-
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * finish * correct script * correct script
-
SaulLu authored
* preserve type of `additional_special_tokens` in `special_token_map` * format * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Funtowicz Morgan authored
* Set model in eval mode when exporting to ONNX. * Disable t5 for now. * Disable T5 with past too. * Style.
-
- 15 Jul, 2021 9 commits
-
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * start adding tests * correct wav2vec2 pretraining * up * up Co-authored-by:Patrick von Platen <patrick@huggingface.co>
-
Lysandre Debut authored
* Don't test LED on torchscript * Typo
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
- 14 Jul, 2021 3 commits
-
-
Stas Bekman authored
* split the test into 4 sub-tests to avoid timeout * fix decorator order
-
Sylvain Gugger authored
* Base test * More test * Fix mistake * Add a docstring change * Add doc ignore * Add changes * Add recursive dep search * Add recursive dep search * save * Finalize test mapping * Fix bug * Print prettier * Ignore comments and empty lines * Make script runnable from anywhere * Need dev install * Like that * Adapt * Add as artifact * Try on torch tests * Fix yaml error * Install GitPython * Apply everywhere * Be more defensive * Revert to all tests if something is wrong * Install GitPython * Test if there are tests before launching. * Fixes * Fixes * Fixes * Fixes * Bash syntax is horrible * Be less stupid * Try differently * Typo * Typo * Typo * Style * Better name * Escape quotes * Ignore black unhelpful re-formatting * Not a docstring * Deal with inits in dependency map * Run all tests once PR is merged. * Add last job * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Stronger dependencies gather * Ignore empty lines too! * Clean up * Fix quality Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Stas Bekman authored
-
- 13 Jul, 2021 5 commits
-
-
Stas Bekman authored
* zero_to_fp32 tests * args change * remove unnecessary work * use transformers.trainer_utils.get_last_checkpoint * document the new features * cleanup * wip * fix fsmt * add bert * cleanup * add xlm-roberta * electra works * cleanup * sync * split off the model zoo tests * cleanup * cleanup * cleanup * cleanup * reformat * cleanup * casing * deepspeed>=0.4.3 * adjust distilbert * Update docs/source/main_classes/deepspeed.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * correct greedy search * save intertmed * add final logits bias * correct * up * add more tests * fix another bug * finish tests * finish marian tests * up Co-authored-by:Patrick von Platen <patrick@huggingface.co>
-
Sylvain Gugger authored
* Add option to load a pretrained model with mismatched shapes * Fail at loading when mismatched shapes in Flax * Fix tests * Update src/transformers/modeling_flax_utils.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Address review comments Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Lysandre Debut authored
* **encode_plus() shouldn't run for W2V2CTC * Typo
-
Lysandre Debut authored
-
- 12 Jul, 2021 4 commits
-
-
Lysandre Debut authored
* Cleanup test * Skip TF TransfoXL test
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
* Skip TestMarian_MT_EN * Skip EN_ZH and EN_ROMANCE * Skip EN_ROMANCE pipeline
-
- 09 Jul, 2021 4 commits
-
-
Will Rice authored
* TFHubert * Update with TFWav2Vec Bug Fixes * Add OOV Error * Feedback changes * Fix kwargs call
-
Alex Hedges authored
* Pass model_kwargs when loading a model in pipeline * Add test for model_kwargs parameter of pipeline() * Rewrite test to not download model * Fix failing style checks
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * add marian * finish make style * add model * add docs * add test * add integration tests * up * solve bug * correct tests * correct some tests * Apply suggestions from code review Co-authored-by:
Suraj Patil <surajp815@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * correct adapt marian * finish Co-authored-by:
Patrick von Platen <patrick@huggingface.co> Co-authored-by:
Suraj Patil <surajp815@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Nicolas Patry authored
* This will reduce "Already borrowed error": Original issue https://github.com/huggingface/tokenizers/issues/537 The original issue is caused by transformers calling many times mutable functions on the rust tokenizers. Rust needs to guarantee that only 1 agent has a mutable reference to memory at a given time (for many reasons which don't need explaining here). Usually, the rust compiler can guarantee that this property is true at compile time. Unfortunately, this is impossible for Python to do that, so PyO3, the bridge between rust and python used by `tokenizers`, will change the compile guarantee for a dynamic guarantee, so if multiple agents try to have multiple mutable borrows at the same time, then the runtime will yell with "Already borrowed". The proposed fix here in transformers, is simply to reduce the actual number of calls that really need mutable borrows. By reducing them, we reduce the risk of running into "Already borrowed" error. The caveat is now we add a call to read the current configuration of the `_tokenizer`, so worst case we have 2 calls instead of 1, and best case we simply have 1 + a Python comparison of a dict (should be negligible). * Adding a test. * trivial error :(. * Update tests/test_tokenization_fast.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * Adding reference to original issues in the tests. * Update the tests with fast tokenizer. Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com>
-
- 08 Jul, 2021 2 commits
-
-
Nicolas Patry authored
* Fixing the pipeline optimization by rescaling the logits first. * Add test for target equivalence Co-authored-by:Lysandre <lysandre.debut@reseau.eseo.fr>
-
Funtowicz Morgan authored
* Laying down building stone for more flexible ONNX export capabilities * Ability to provide a map of config key to override before exporting. * Makes it possible to export BART with/without past keys. * Supports simple mathematical syntax for OnnxVariable.repeated * Effectively apply value override from onnx config for model * Supports export with additional features such as with-past for seq2seq * Store the output path directly in the args for uniform usage across. * Make BART_ONNX_CONFIG_* constants and fix imports. * Support BERT model. * Use tokenizer for more flexibility in defining the inputs of a model. * Add TODO as remainder to provide the batch/sequence_length as CLI args * Enable optimizations to be done on the model. * Enable GPT2 + past * Improve model validation with outputs containing nested structures * Enable Roberta * Enable Albert * Albert requires opset >= 12 * BERT-like models requires opset >= 12 * Remove double printing. * Enable XLM-Roberta * Enable DistilBERT * Disable optimization by default * Fix missing setattr when applying optimizer_features * Add value field to OnnxVariable to define constant input (not from tokenizers) * Add T5 support. * Simplify model type retrieval * Example exporting token_classification pipeline for DistilBERT. * Refactoring to package `transformers.onnx` * Solve circular dependency & __main__ * Remove unnecessary imports in `__init__` * Licences * Use @Narsil's suggestion to forward the model's configuration to the ONNXConfig to avoid interpolation. * Onnx export v2 fixes (#12388) * Tiny fixes Remove `convert_pytorch` from onnxruntime-less runtimes Correct reference to model * Style * Fix Copied from * LongFormer ONNX config. * Removed optimizations * Remvoe bad merge relicas. * Remove unused constants. * Remove some deleted constants from imports. * Fix unittest to remove usage of PyTorch model for onnx.utils. * Fix distilbert export * Enable ONNX export test for supported model. * Style. * Fix lint. * Enable all supported default models. * GPT2 only has one output * Fix bad property name when overriding config. * Added unittests and docstrings. * Disable with_past tests for now. * Enable outputs validation for default export. * Remove graph opt lvls. * Last commit with on-going past commented. * Style. * Disabled `with_past` for now * Remove unused imports. * Remove framework argument * Remove TFPreTrainedModel reference * Add documentation * Add onnxruntime tests to CircleCI * Add test * Rename `convert_pytorch` to `export` * Use OrderedDict for dummy inputs * WIP Wav2Vec2 * Revert "WIP Wav2Vec2" This reverts commit f665efb04c92525c3530e589029f0ae7afdf603e. * Style * Use OrderedDict for I/O * Style. * Specify OrderedDict documentation. * Style :) Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 07 Jul, 2021 2 commits
-
-
Nicolas Patry authored
* Adding support for `pipeline("automatic-speech-recognition")`. - Ugly `"config"` choice for AutoModel. It would be great to have the possibility to have something like `AutoModelFor` that would implement the same logic (Load the config, check Architectures and load the first one) * Remove `model_id` was not needed in the end. * Rebased ! * Remove old code. * Rename `nlp`. -
Daniel Stancl authored
* Copy BART to MBart and rename some stuff * Add copy statements pointing to FlaxBart * Update/add some common files * Update shift_tokens_rigth + fix imports * Fix shift_tokens_right method according to MBart implementation * Update shift_tokens_right in tests accordingly * Fix the import issue and update docs file * make style quality * Do some minor changes according to patil-suraj suggestions * Change the order of normalization layer and attention * Add some copu statementes * Update generate method and add integration test for mBart * Make a few updates after a review Besides, add `lang_code_to_id` to MBartTokenizeFast * fix-copies; make style quality * Apply suggestions from code review * Apply suggestions from code review * Apply suggestions from code review * fix output type, style * add copied from * resolve conflicts Co-authored-by:Suraj Patil <surajp815@gmail.com>
-
- 06 Jul, 2021 3 commits
-
-
sadakmed authored
* implementing tflxmertmodel integration test * move import * revert and fix
-
Suraj Patil authored
* flax gpt neo * fix query scaling * update generation test * use flax model for test
-
yujun authored
* add RoFormerTokenizerFast into AutoTokenizer * fix typo in roformer docs * make onnx export happy * update RoFormerConfig embedding_size * use jieba not rjieba * fix 12244 and make test_alignement passed * update ARCHIVE_MAP * make style & quality & fixup * update * make style & quality & fixup * make style quality fixup * update * suggestion from LysandreJik Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * make style * use rjieba Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 05 Jul, 2021 1 commit
-
-
sadakmed authored
* create LxmertModelIntegrationTest * implementation using numpy seeding to fix inputs params. * fix code quality * isort check
-
- 02 Jul, 2021 1 commit
-
-
Lysandre Debut authored
-
- 01 Jul, 2021 2 commits
-
-
Stas Bekman authored
* fix lm_head.decoder.weight ignore_key handling * fix the mutable class variable * Update src/transformers/models/roberta/modeling_roberta.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * replicate the comment * make deterministic Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * fix test
-