- 15 Nov, 2022 25 commits
-
-
Yih-Dar authored
* Allow trainer to return loss for CLIP-like models * Apply suggestions * update * update * update Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Zachary Mueller authored
* Update reqs to include min gather_for_metrics Accelerate version * Other reqs
-
Ambuj Pawar authored
* WIP: Added CLIP resources from HuggingFace blog * ADD: Notebooks documentation to clip * Add link straight to notebook Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Change notebook links to colab Co-authored-by:
Ambuj Pawar <your_email@abc.example> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Saad Mahmud authored
* Add to DeBERTa resources * Fix mistakes with chapter number * Add fill-mask pipeline * Add sequence, token and QA pipeline * Change token classification pipeline order * Remove flax script and notebook links
-
Matt authored
* Slightly alter Keras dummy loss * Slightly alter Keras dummy loss * Add sample weight to test_keras_fit * Fix test_keras_fit for datasets * Skip the sample_weight stuff for models where the model tester has no batch_size
-
Suraj Patil authored
* allow loading projection in text and vision model * begin tests * finish test for CLIPTextModelTest * style * add slow tests * add new classes for projection heads * remove with_projection * add in init * add in doc * fix tests * fix some more tests * fix copies * fix docs * remove leftover from fix-copies * add the head models in IGNORE_NON_AUTO_CONFIGURED * fix docstr * fix tests * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * add docstr for models Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
* Try PT1.13 by removing torch scatter * Skip failing tests * Style * Remvoe testing extras for repo utils * Try with all decorators * Try to wipe the cache * Fix all tests? * Try this way * Fix comma * Update to main * Try with less deps * Quality
-
Muhammad Sakib Khan Inan authored
* Init Update * ClearML Callbacks integration * update corrections * args reporting updated * {'tensorboard': False, 'pytorch': False} * ClearML Tests added * add clearml * output_uri=True in Task.init * reformatted integrations.py * reformatted and fixed * IF-ELSE statement issue on "has_clearml" resolved * Add clearml in main callback docs * Add additional clearml documentation * Update src/transformers/integrations.py Co-authored-by:Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Accept suggestion Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Accept suggestion Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Small change in comments * Make style clearml * Accept suggestion Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Victor Sonck <victor.sonck@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
NielsRogge authored
* Fix bug * Add another fix * Add print statement * Apply fix * Fix feature extractor * Fix feature extractor * Add print statements * Add print statements * Remove print statements * Add instance segmentation integration test * Add integration test for semantic segmentation * Add draft for panoptic segmentation integration test * Fix integration test for panoptic segmentation * Remove slow annotator Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
TilmannR authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Kendall authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Matt authored
* Add missing ESM autoclass * Correct ESMFold checkpoint
-
Arthur authored
-
Yong woo Song authored
* chore: typo on docstring in tokenization_electra * chore: typo on docstring in tokenization_electra * update for check copies
-
amyeroberts authored
* Add transforms for object detection * Update src/transformers/image_transforms.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Better var names & docstring * Remove unused var desc in docstring * Update src/transformers/image_transforms.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Younes Belkada authored
* first commit * add more comments * add router v1 * clean up - remove `tf` modeling files * clean up - remove `tf` modeling files * clean up * v0 routers * added more router - Implemented `ExpertsChooseMaskedRouter` - added tests - 2 more routers to implement * last router * improved docstring - completed the docstring in `router.py` - added more args in the config * v0 sparse mlp * replace wrong naming * forward pass run * update MOE layer * small router update * fixup * consistency * remove scatter router * remove abstract layer * update test and model for integration testing * v1 conversion * update * hardcode hack * all keys match * add gin conversion, without additional libraries * update conversion sctipy * delete router file * update tests wrt router deletion * fix router issues * update expert code * update, logits match, code needsREFACTORING * Refactor code Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * add generate tests Co-authored-by:
younesbelkada <younesbelkada@gmail.com> * add support for router loss Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * fix forward error * refactor a bit * remove `FlaxSwitchTransformers` modules * more tests pass * Update code Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> * fixup * fix tests * fix doc * fix doc + tokenization * fix tokenizer test * fix test * fix loss output * update code for backward pass * add loss support * update documentation * fix documentation, clean tokenizer * more doc fix, cleanup example_switch * fix failing test * fix test * fix test * fix loss issue * move layer * update doc and fix router capacity usage * fixup * add sparse mlp index for documentation on hub * fixup * test sparse mix architecture * Apply suggestions from code review * Update docs/source/en/model_doc/switch_transformers.mdx * fixup on update * fix tests * fix another test * attempt fix * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/switch_transformers/convert_switch_transformers_original_flax_checkpoint_to_pytorch.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * try * all tests pass * fix jitter noise * Apply suggestions from code review * doc tests pass * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/switch_transformers/modeling_switch_transformers.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * remove assert * change config order * fix readme japanese * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * remove parallelizable tests + add one liners * remove ONNX config * fix nits - add `T5Tokenizer` in auto mapping - remove `Switch Transformers` from ONNX supported models * remove `_get_router` * remove asserts * add check in test for `router_dtype` * add `SwitchTransformersConfig` in `run_pipeline_test` * Update tests/pipelines/test_pipelines_summarization.py * add huge model conversion script * fix slow tests - add better casting for `Linear8bitLt` - remove `torchscript` tests * add make dir * style on new script * fix nits - doctest - remove `_keys_to_ignore_on_load_unexpected` * Update src/transformers/models/switch_transformers/configuration_switch_transformers.py * add google as authors * fix year * remove last `assert` statements * standardize vertical spaces * fix failing import * fix another failing test * Remove strange 脿uthorized_keys` * removing todo and padding that is never used Co-authored-by:
Arthur Zucker <arthur.zucker@gmail.com> Co-authored-by:
ybelkada <younes@huggingface.co> Co-authored-by:
Younes Belkada <younesbelkada@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Arthur Zucker <arthur@huggingface.co>
-
amyeroberts authored
-
Younes Belkada authored
* add `accelerate` support for `ViT` family - add `_no_split_modules` - manually cast to the right `dtype`: to change * enable `float16` for `deit` * fix `make fixup` * add `slow` test for `fp16` inference * another safety check * Update src/transformers/models/deit/modeling_deit.py
-
Arthur authored
* Update modeling tests * update tokenization test * typo * nit * fix expected attention outputs * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update tests from review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> * remove problematics kwargs passed to the padding function Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Arthur authored
* update relative positional embedding * make fix copies * add `use_cache` to list of arguments * fixup * 1line fucntion * add `test_decoder_model_past_with_large_inputs_relative_pos_emb` * add relative pos embedding test for more models * style
-
Yih-Dar authored
* Fix ImageSegmentationPipelineTests * Use 0.9 * no zip * links to show images * links to show images * rebase Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
bofeng huang authored
-
Wonhyeong Seo authored
* docs: fix: set overflowing image width to auto-scale * docs: fix: new language Korean is also affected * docs: fix: unnecessary line break in index page
-
- 14 Nov, 2022 15 commits
-
-
Nicolas Patry authored
* Very crude matching algorithm. * Fixing tests. * Removing comments * Adding warning + fix short matches. * Cleanup tests. * Quality. * Less noisy. * Fixup.
-
Joao Gante authored
-
amyeroberts authored
-
Ming Liu authored
There is typo in the original hyperlink. Below is the original version: Based on the script [`run_translation_no_trainer.py`](https://github.com/huggingface/transformers/blob/main/examples/pytorch/translation/**run_translationn_no_trainer.py**).
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sanchit Gandhi authored
* merge conflicts * bos and eos in datacollator * (temp) hardcode removal of attention mask * freeze encoder * actually freeze encoder * set max length / num beams according to gen kwargs * (temp) fix tests * don't pop attn mask * override return attention mask config from Hub * Hub configs updated
馃 * final fixes * update type annotations * backward comp -
Wonhyeong Seo authored
Part of #20183 docs: add relevant labels to i18n issue template fix: typo on completion count
-
Wonhyeong Seo authored
docs: i18n: first draft of index page docs: fix: first revision of index page docs: i18n: missed section - supported frameworks docs: fix: second revision of index page review by @ArthurZucker refactor: remove untranslated files from korean docs: fix: remove untranslated references from toctree.yml feat: enable korean docs in gh actions docs: feat: add in_translation page as placeholder docs: bug: testing if internal toc need alphabet chars docs: fix: custom english anchor for non-alphanumeric headings review by @sgugger docs: i18n: translate comments on install methods in _config.py docs: refactor: more concise wording for translations
-
Arthur authored
-
Younes Belkada authored
* fix slow test * Update tests/models/roc_bert/test_modeling_roc_bert.py
-
Joao Gante authored
-
IMvision12 authored
* initial commit * Update modeling_whisper.py * Fixing Tests * modeling_vision_text_dual_encoder * modeling_vision_encoder_decoder * Update modeling_vit.py * Update modeling_vit_msn.py * Update modeling_trajectory_transformer.py * style * Update modeling_time_series_transformer.py * Update modeling_time_series_transformer.py * Update modeling_segformer.py * Update modeling_plbart.py * Update modeling_dpt.py * Update modeling_deit.py * Update modeling_dpt.py * Update modeling_esm.py * Update modeling_fnet.py * Update modeling_fnet.py * Update modeling_fnet.py * Update modeling_flava.py * Update modeling_flava.py * Update modeling_layoutlmv3.py * Update modeling_levit.py
-
Nicolas Patry authored
Proposal Remove the weird `inspect` in ASR pipeline and make WhisperEncoder just nice to use. (#19571) * Proposal Remove the weird `inspect` in ASR pipeline and make WhisperEncoder just nice to use. It seems that accepting `attention_mask` is kind of an invariant of our models. For Seq2Seq ASR models, we had a special comment on how it actually was important to send it. `inspecting` seems pretty brittle way to handle this case. My suggestion is to simply add it as an kwarg that and just ignoring it with the docstring explaining why it's ignored. * Fixup. * Update src/transformers/models/whisper/modeling_whisper.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Doc fixing . Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
code-with-rajeev authored
Fixed a grammatical error.
-
Bartosz Szmelczynski authored
* First draft * Remove scatter dependency * Add require_torch * update vectorized sum test, add clone call * remove artifacts * fix style * fix style v2 * remove "scatter" mentions from the code base * fix isort error Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-