- 28 Aug, 2021 1 commit
-
-
Stefan Schweter authored
* examples: only use keep_linebreaks when reading TXT files for all CLM examples * examples: only use keep_linebreaks when reading TXT files for all CLM examples * examples: only use keep_linebreaks when reading TXT files for all CLM examples
-
- 27 Aug, 2021 8 commits
-
-
Anton Lozhkov authored
* Add hubert classifier + tests * Add hubert classifier + tests * Dummies for all classification tests * Wav2Vec2 classifier + ER test * Fix hubert integration tests * Add hubert IC * Pass tests for all classification tasks on Hubert * Pass all tests + copies * Move models to the SUPERB org
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * finish find and replace
-
Nicolas Patry authored
* Fixing mbart50 with `return_tensors` argument too. * Adding mbart50 tokenization tests.
-
Nicolas Patry authored
* Moving `zero-shot-classification` pipeline to new testing. * Cleaning up old mixins. * Fixing tests `sshleifer/tiny-distilbert-base-uncased-finetuned-sst-2-english` is corrupted in PT. * Adding warning.
-
NielsRogge authored
* First pass * Fix docs of bool_masked_pos * Add integration script * Fix docstring * Add integration test for BeitForMaskedImageModeling * Remove file * Fix docs
-
Nicolas Patry authored
* Moving `translation` pipeline to new testing scheme. * Update tokenization mbart tests.
-
Stefan Schweter authored
* examples: add keep_linebreaks option to text dataset loader for all CLM examples * examples: introduce new keep_linebreaks option as data argument in CLM examples
-
Nicolas Patry authored
* Moving `token-classification` pipeline to new testing. * Fix tests.
-
- 26 Aug, 2021 13 commits
-
-
Nicolas Patry authored
* Moving `text-generation` pipeline to new testing framework. * Keep check_model_type but log instead of raise Exception. * warning -> error.
-
NielsRogge authored
* First commit * Add interpolation of patch embeddings * Comment out code * Fix bug * Fix another bug * Fix bug * Fix another bug * Remove print statements * Update conversion script * Use the official vit implementation * Add support for converting dino_vits8 * Add DINO to docs of ViT * Remove assertion * Add interpolation of position encodings * Fix bug * Add align_corners * Add interpolate_pos_encoding option to forward pass of ViTModel * Improve interpolate_pos_encoding method * Add docstring
-
Nicolas Patry authored
-
Nicolas Patry authored
-
Nicolas Patry authored
-
Nicolas Patry authored
-
Nicolas Patry authored
* Moving `summarization` pipeline to new testing format. * Remove generate_kwargs from __init__ args.
-
Nicolas Patry authored
Moving question_answering tests to the new testing scheme. Had to tweak a little some ModelTesterConfig for pipelines. (#13277) * Moving question_answering tests to the new testing scheme. Had to tweak a little some ModelTesterConfig for pipelines. * Removing commented code.
-
Nicolas Patry authored
-
Nicolas Patry authored
- Enforce `test_small_models_{tf,pt}` methods to exist (enforce checking actual values in small tests) - Add support for non RGB image for the pipeline. -
Bram Vanroy authored
* add error message concerning revision * Update src/transformers/configuration_utils.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * re-add double line endings * is not None instead of implicit bool casting Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Stas Bekman authored
* fix tokenizer_class_from_name * Update src/transformers/models/auto/tokenization_auto.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * add test Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Nicolas Patry authored
* New test format for conversational. * Putting back old mixin. * Re-enabling auto tests with LazyLoading. * Feature extraction tests. * Remove feature-extraction. * Feature extraction with feature_extractor (No pun intended). * Update check_model_type for fill-mask.
-
- 25 Aug, 2021 8 commits
-
-
Lysandre Debut authored
-
Lysandre Debut authored
* Some tokenizers cannot be in the mapping * Style
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Lysandre Debut authored
-
Nishant Prabhu authored
-
Lysandre authored
-
Lysandre Debut authored
-
- 24 Aug, 2021 6 commits
-
-
Will Frey authored
If you're using type hints, then passing an `int` where a `float` is annotated is acceptable as per [PEP 484](https://www.python.org/dev/peps/pep-0484/#the-numeric-tower). This makes life a little nicer.
-
dependabot[bot] authored
Bumps [notebook](http://jupyter.org ) from 6.1.5 to 6.4.1. --- updated-dependencies: - dependency-name: notebook dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Ambesh Shekhar authored
* Adding custom errors and BatchSizeError for GPT2 * Adding custom errors and BatchSizeError for GPT2 * Changing Exception to BaseException * Exception * Adding args to Custom Exception * Adding args to Custom Exception * Changing from BaseException to Exception * Changing Conditional loop syntax * Adding Copyright info * Handling check_code_quality * Handling check_code_quality pt2 * Handling check_code_quality pt3 * Handling check_code_quality pt4 * Handling check_code_quality pt5 * Handling check_code_quality pt6 * Handling check_code_quality pt6 * Using black for check_code_quality * sorting import style * Changing * Changing * verified through style_doc.py * verified through style_doc.py * applying isort * Removing indentation * Changing * Changing * Changing * Used ValueError * Using ValueError * Reformatted Style doc * Using style doc on modeling_gp2.py * Adding indentation * Changing
-
Ori Ram authored
-
Stas Bekman authored
* fix AutoModel.from_pretrained(..., torch_dtype=...) * fix to_diff_dict * add better test * torch is not always available when a model has self.torch_dtype
-
Bram Vanroy authored
* allow local_files_only for fast pretrained tokenizers * make style
-
- 23 Aug, 2021 4 commits
-
-
Lysandre Debut authored
-
Allan Lin authored
-
Kamal Raj authored
-
Yih-Dar authored
* make flax gpt2 working with cross attention * Remove encoder->decoder projection layer * A draft (incomplete) for FlaxEncoderDecoderModel * Add the method from_encoder_decoder_pretrained + the docstrings * Fix the mistakes of using EncoderDecoderModel * Fix style * Add FlaxEncoderDecoderModel to the library * Fix cyclic imports * Add FlaxEncoderDecoderModel to modeling_flax_auto.py * Remove question comments * add tests for FlaxEncoderDecoderModel * add flax_encoder_decoder to the lists of ignored entries in check_repo.py * fix missing required positional arguments * Remove **kwargs when creating FlaxEncoderDecoderModel in from_encoder_decoder_pretrained() Also fix generation eos/pad tokens issue * Fix: Use sequences from the generated_output * Change a check from assert to raise ValueError * Fix examples and token ids issues * Fix missing all_cross_attentions when outputting tuple in modeling_gpt2 * Remove the changes in configuration docstrings. * allow for bert 2 gpt2 * make fix-copies * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Change remaining examples to bert2gpt2 * Change the test to Bert2GPT2 * Fix examples * Fix import * Fix unpack bug * Rename to FlaxEncoderDecoderModelTest and change the test to bert2gpt2 * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Fix: NotImplentedError -> NotImplementedError * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * up * finalize Co-authored-by:
ydshieh <ydshieh@user.noreply> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-