- 16 May, 2023 10 commits
-
-
Joao Gante authored
-
Younes Belkada authored
add conditional generation on docstring
-
Lucain authored
* Few fixes in new Tools implementation * code quality
-
Sohyun Sim authored
* docs: ko: task/asr.mdx * feat: manual draft * fix: resolve suggestions Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> --------- Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com>
-
Ivan Sedykh authored
fix chat prompts
-
Joao Gante authored
-
Yih-Dar authored
* fix * fix * fix test + add more docs --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> Co-authored-by:
younesbelkada <younesbelkada@gmail.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
ready-research authored
* Use `mkstemp` to replace deprecated `mktemp` The `tempfile.mktemp` function is [deprecated](https://docs.python.org/3/library/tempfile.html#tempfile.mktemp) due to [security issues](https://cwe.mitre.org/data/definitions/377.html ). * Update src/transformers/utils/hub.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Koki Tanaka authored
* Replace numpy operations with jax.numpy for JIT compatibility Replaced numpy operations with their jax.numpy equivalents in the transformer library. This change was necessary to prevent errors during JIT compilation. Specifically, the modifications involve changing numpy's in-place assignments to jax.numpy's immutable update methods. * rm numpy import * rm numpy import and fix np->jnp * fixed slices bug * fixed decoder_start_tokens -> decoder_start_token_id * fixed jnp in modleing mt5 * doc fix * rm numpy import * make
-
- 15 May, 2023 11 commits
-
-
dewa authored
* Added type hints for `Graphormer` pytorch version added type hints for graphormers pytorch , checked formating issues . * made the code less bloated
-
LWprogramming authored
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
AinL authored
[Bugfix] `OPTDecoderLayer` does not return attentions when `gradient_checkpointing` and `training` is enabled. (#23367) Update modeling_opt.py
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
richardachen authored
Update graphormer.mdx Typo suggestion
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Gregory authored
Removing twice defined position_embeddings The self.position_embeddings in LongFormerEmbeddings is defined twice. Removing the first with padding_idx
-
- 12 May, 2023 10 commits
-
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Susnato Dhar authored
replaced assert with raise ValueError for t5, switch_transformers, pix2struct, mt5, longt5, gptsan_japanese. (#23273) * replaced assert with raise ValueError * one liner * reverse one liner and cache-decoder check
-
Alisamar Husain authored
* Handle padding warning in generation when using `inputs_embeds` * Simpler condition * Black formatter * Changed warning logic
-
hwuebben authored
or or and
-
Steven Liu authored
fix kwargs
-
Yih-Dar authored
* min. version for pytest * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Mario Lezcano Casado authored
When working on TorchInductor, I realised that there was a part from `XLNetLMHeadModel` that was being compiled to CPU code. This PR should allow to fuse this operation with other CUDA operations in `torch.compile`. It also should be faster on eager mode, as it has a this implementation has a lower foot-print. If in-place operations are not allowed even in non-grad context, I still believe that doing ones + tril rather than a ones + tril + zeros + cat should be faster simply due to the number of memory reads/writes. I tested that this code produces the same results for `0 <= qlen,mlen < 10` and `same_length in (True, False)`.
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Shehan Munasinghe authored
* Commit the automatically generated code using add-new-model-like * Update description at swiftformer.mdx file * remove autogenerated code for MaskedImageModeling * update weight conversion scripts * Update modeling_swiftformer.py * update configuration_swiftformer.py * Update test_modeling_swiftformer.py * update modeling code - remove einops dependency * Update _toctree.yml * update modeling code - remove copied from comments * update docs * Revert "update docs" This reverts commit c2e05e2998fe2cd6eaee8b8cc31aca5222bac9fb. * update docs * remove unused reference SwiftFormerImageProcessor * update dependency_versions_table.py * update swiftformer.mdx * update swiftformer.mdx * change model output type - no attentions * update model org name * Fix typo * fix copies * Update tests/models/swiftformer/test_modeling_swiftformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/auto/image_processing_auto.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/auto/feature_extraction_auto.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update docs/source/en/model_doc/swiftformer.mdx Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/swiftformer/configuration_swiftformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update modeling_swiftformer.py fix-copies * make style, make quality, fix-copies * Apply suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * make style Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Add suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Add suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * make fix-copies * Update modeling_swiftformer.py * Update modeling_swiftformer.py * Add suggestions from code review Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Yih-Dar authored
remove LanguageIdentificationTool Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 11 May, 2023 9 commits
-
-
Sylvain Gugger authored
Revert "search buffers for dtype (#23159)" This reverts commit ef42c2c4.
-
Yih-Dar authored
* unpin tf prob --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Freddy Boulton authored
Fix typo
-
Sylvain Gugger authored
-
Lysandre Debut authored
* Agents extras * Add to docs
-
raghavanone authored
Add gradient_checkpointing parameter
-
Alessandro Pietro Bardelli authored
* Better check for packages availability * amend _optimumneuron_available * amend torch_version * amend PIL detection and lint * lint * amend _faiss_available * remove overloaded signatures of _is_package_available * fix sklearn and decord detection * remove unused checks * revert
-