- 15 Nov, 2022 14 commits
-
-
Kendall authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Matt authored
* Add missing ESM autoclass * Correct ESMFold checkpoint
-
Arthur authored
-
Yong woo Song authored
* chore: typo on docstring in tokenization_electra * chore: typo on docstring in tokenization_electra * update for check copies
-
amyeroberts authored
* Add transforms for object detection * Update src/transformers/image_transforms.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Better var names & docstring * Remove unused var desc in docstring * Update src/transformers/image_transforms.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Younes Belkada authored
* first commit * add more comments * add router v1 * clean up - remove `tf` modeling files * clean up - remove `tf` modeling files * clean up * v0 routers * added more router - Implemented `ExpertsChooseMaskedRouter` - added tests - 2 more routers to implement * last router * improved docstring - completed the docstring in `router.py` - added more args in the config * v0 sparse mlp * replace wrong naming * forward pass run * update MOE layer * small router update * fixup * consistency * remove scatter router * remove abstract layer * update test and model for integration testing * v1 conversion * update * hardcode hack * all keys match * add gin conversion, without additional libraries * update conversion sctipy * delete router file * update tests wrt router deletion * fix router issues * update expert code * update, logits match, code needsREFACTORING * Refactor code Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * add generate tests Co-authored-by:
younesbelkada <younesbelkada@gmail.com> * add support for router loss Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * fix forward error * refactor a bit * remove `FlaxSwitchTransformers` modules * more tests pass * Update code Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * fixup * fix tests * fix doc * fix doc + tokenization * fix tokenizer test * fix test * fix loss output * update code for backward pass * add loss support * update documentation * fix documentation, clean tokenizer * more doc fix, cleanup example_switch * fix failing test * fix test * fix test * fix loss issue * move layer * update doc and fix router capacity usage * fixup * add sparse mlp index for documentation on hub * fixup * test sparse mix architecture * Apply suggestions from code review * Update docs/source/en/model_doc/switch_transformers.mdx * fixup on update * fix tests * fix another test * attempt fix * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/switch_transformers/convert_switch_transformers_original_flax_checkpoint_to_pytorch.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * try * all tests pass * fix jitter noise * Apply suggestions from code review * doc tests pass * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * remove assert * change config order * fix readme japanese * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * remove parallelizable tests + add one liners * remove ONNX config * fix nits - add `T5Tokenizer` in auto mapping - remove `Switch Transformers` from ONNX supported models * remove `_get_router` * remove asserts * add check in test for `router_dtype` * add `SwitchTransformersConfig` in `run_pipeline_test` * Update tests/pipelines/test_pipelines_summarization.py * add huge model conversion script * fix slow tests - add better casting for `Linear8bitLt` - remove `torchscript` tests * add make dir * style on new script * fix nits - doctest - remove `_keys_to_ignore_on_load_unexpected` * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py * add google as authors * fix year * remove last `assert` statements * standardize vertical spaces * fix failing import * fix another failing test * Remove strange 脿uthorized_keys` * removing todo and padding that is never used Co-authored-by:
Arthur Zucker <arthur.zucker@gmail.com> Co-authored-by:
ybelkada <younes@huggingface.co> Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Arthur Zucker <arthur@huggingface.co>
-
amyeroberts authored
-
Younes Belkada authored
* add `accelerate` support for `ViT` family - add `_no_split_modules` - manually cast to the right `dtype`: to change * enable `float16` for `deit` * fix `make fixup` * add `slow` test for `fp16` inference * another safety check * Update src/transformers/models/deit/modeling_deit.py
-
Arthur authored
* Update modeling tests * update tokenization test * typo * nit * fix expected attention outputs * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update tests from review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> * remove problematics kwargs passed to the padding function Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Arthur authored
* update relative positional embedding * make fix copies * add `use_cache` to list of arguments * fixup * 1line fucntion * add `test_decoder_model_past_with_large_inputs_relative_pos_emb` * add relative pos embedding test for more models * style
-
Yih-Dar authored
* Fix ImageSegmentationPipelineTests * Use 0.9 * no zip * links to show images * links to show images * rebase Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
bofeng huang authored
-
Wonhyeong Seo authored
* docs: fix: set overflowing image width to auto-scale * docs: fix: new language Korean is also affected * docs: fix: unnecessary line break in index page
-
- 14 Nov, 2022 16 commits
-
-
Nicolas Patry authored
* Very crude matching algorithm. * Fixing tests. * Removing comments * Adding warning + fix short matches. * Cleanup tests. * Quality. * Less noisy. * Fixup.
-
Joao Gante authored
-
amyeroberts authored
-
Ming Liu authored
There is typo in the original hyperlink. Below is the original version: Based on the script [`run_translation_no_trainer.py`](https://github.com/huggingface/transformers/blob/main/examples/pytorch/translation/**run_translationn_no_trainer.py**).
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sanchit Gandhi authored
* merge conflicts * bos and eos in datacollator * (temp) hardcode removal of attention mask * freeze encoder * actually freeze encoder * set max length / num beams according to gen kwargs * (temp) fix tests * don't pop attn mask * override return attention mask config from Hub * Hub configs updated
馃 * final fixes * update type annotations * backward comp -
Wonhyeong Seo authored
Part of #20183 docs: add relevant labels to i18n issue template fix: typo on completion count
-
Wonhyeong Seo authored
docs: i18n: first draft of index page docs: fix: first revision of index page docs: i18n: missed section - supported frameworks docs: fix: second revision of index page review by @ArthurZucker refactor: remove untranslated files from korean docs: fix: remove untranslated references from toctree.yml feat: enable korean docs in gh actions docs: feat: add in_translation page as placeholder docs: bug: testing if internal toc need alphabet chars docs: fix: custom english anchor for non-alphanumeric headings review by @sgugger docs: i18n: translate comments on install methods in _config.py docs: refactor: more concise wording for translations
-
Arthur authored
-
Younes Belkada authored
* fix slow test * Update tests/models/roc_bert/test_modeling_roc_bert.py
-
Joao Gante authored
-
IMvision12 authored
* initial commit * Update modeling_whisper.py * Fixing Tests * modeling_vision_text_dual_encoder * modeling_vision_encoder_decoder * Update modeling_vit.py * Update modeling_vit_msn.py * Update modeling_trajectory_transformer.py * style * Update modeling_time_series_transformer.py * Update modeling_time_series_transformer.py * Update modeling_segformer.py * Update modeling_plbart.py * Update modeling_dpt.py * Update modeling_deit.py * Update modeling_dpt.py * Update modeling_esm.py * Update modeling_fnet.py * Update modeling_fnet.py * Update modeling_fnet.py * Update modeling_flava.py * Update modeling_flava.py * Update modeling_layoutlmv3.py * Update modeling_levit.py
-
Nicolas Patry authored
Proposal Remove the weird `inspect` in ASR pipeline and make WhisperEncoder just nice to use. (#19571) * Proposal Remove the weird `inspect` in ASR pipeline and make WhisperEncoder just nice to use. It seems that accepting `attention_mask` is kind of an invariant of our models. For Seq2Seq ASR models, we had a special comment on how it actually was important to send it. `inspecting` seems pretty brittle way to handle this case. My suggestion is to simply add it as an kwarg that and just ignoring it with the docstring explaining why it's ignored. * Fixup. * Update src/transformers/models/whisper/modeling_whisper.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Doc fixing . Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
code-with-rajeev authored
Fixed a grammatical error.
-
Bartosz Szmelczynski authored
* First draft * Remove scatter dependency * Add require_torch * update vectorized sum test, add clone call * remove artifacts * fix style * fix style v2 * remove "scatter" mentions from the code base * fix isort error Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Matthijs Hollemans authored
* add model files etc for MobileNetV2 * rename files for MobileNetV1 * initial implementation of MobileNetV1 * fix conversion script * cleanup * write docs * tweaks * fix conversion script * extract hidden states * fix test cases * make fixup * fixup it all * rename V1 to V2 * fix checkpoints * fixup * implement first block + weight conversion * add remaining layers * add output stride and dilation * fixup * add tests * add deeplabv3+ head * a bit of fixup * finish deeplab conversion * add link to doc * fix issue with JIT trace in_height and in_width would be Tensor objects during JIT trace, which caused Core ML conversion to fail on the remainder op. By making them ints, the result of the padding calculation becomes a constant value. * cleanup * fix order of models * fix rebase error * remove main from doc link * add image processor * remove old feature extractor * fix converter + other issues * fixup * fix unit test * add to onnx tests (but these appear broken now) * add post_process_semantic_segmentation * use google org * remove unused imports * move args * replace weird assert
-
- 11 Nov, 2022 3 commits
-
-
amyeroberts authored
-
NielsRogge authored
* Apply fix * Fix test * Remove another argument which is not used * Fix pipeline test * Add argument back, add deprecation warning * Add warning add other location * Use warnings instead * Add num_channels to config Co-authored-by:Niels Rogge <nielsrogge@Nielss-MBP.localdomain>
-
Nicolas Patry authored
-
- 10 Nov, 2022 7 commits
-
-
Arthur authored
-
Sylvain Gugger authored
-
Sanchit Gandhi authored
* [processor] Add 'model input names' property * add test * no f string * add generic property method to mixin * copy to multimodal * copy to vision * tests for all audio * remove ad-hoc tests * style * fix flava test * fix test * fix processor code
-
Matt authored
* Fix arg names for our models * Clean out the other uses of "residx" in infer() * make fixup
-
Joao Gante authored
-
Yih-Dar authored
* Check ImageProcessingMixin in OnnxConfig.generate_dummy_inputs * Check ImageProcessingMixin in OnnxConfig.generate_dummy_inputs * Add back Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Matthijs Hollemans authored
-