"docs/source/vscode:/vscode.git/clone" did not exist on "ae710425d2d8edf4d197bf893b90ed0546054701"
- 04 Aug, 2023 7 commits
-
-
Sanchit Gandhi authored
-
Yih-Dar authored
* temp * update * update * update * small dim * small dim * small dim * fix * update * fix * fix * fix * fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* Document check copies better and add tests * Include header in check for copies * Manual fixes * Try autofix * Fixes * Clean tests * Finalize doc * Remove debug print * More fixes
-
Sylvain Gugger authored
* Deal better with nested configs * Fixes * More fixes * Fix last test * Clean up existing configs * Remove hack in MPT Config * Update src/transformers/configuration_utils.py Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * Fix setting a nested config via dict in the kwargs * Adapt common test * Add test for nested config load with dict --------- Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
-
Sylvain Gugger authored
* Add offline mode for agents * Disable second check too
-
Joao Gante authored
-
Peter Law authored
The former spelling is deprecated and has been discouraged for a while. The latter spelling seems to be more common in this project anyway, so this change ought to be safe. Fixes https://github.com/huggingface/transformers/issues/25283
-
- 03 Aug, 2023 5 commits
-
-
Sanchit Gandhi authored
* [JAX] Bump min version * make fixup
-
Roland Szabo authored
* Add timeout parameter to load_image function. * Remove line. * Reformat code Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Add parameter to docs. --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Yoach Lacombe authored
* add generate method to SpeechT5ForTextToSpeech * update speecht5forTTS docstrings * Remove defaults to None in generate docstrings Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
amyeroberts authored
* Update InstructBLIP values Note: the tests are not independent. Running the test independentely produces different logits compared to running all the integration tests * Update test values after rescale update * Remove left over commented out code * Revert to previous rescaling logic * Update rescale tests
-
Tom Aarsen authored
* Update list of logging integrations in docstring Also update type hint * Also add 'flyte' to report_to callback list * Revert 'report_to' type hint update Due to CLI breaking
-
- 02 Aug, 2023 8 commits
-
-
Patrick von Platen authored
* [MMS] Fix mms * [MMS] Fix mms * fix mms loading * Apply suggestions from code review * make style * Update tests/models/wav2vec2/test_modeling_wav2vec2.py
-
Euan Ong authored
Fix bug in InstructBlip generate function Previously, the postprocessing conducted on generated sequences in InstructBlip's generate function assumed these sequences were tensors (i.e. that `return_dict_in_generate == False`). This commit checks whether the result of the call to the wrapped language model `generate()` is a tensor, and if not attempts to postprocess the sequence attribute of the returned results object.
-
Ashish Thomas Chempolil authored
* added example and modified docs for EtaLogitsWarper * make style * fixed styling issue on 544 * removed error info and added set_seed * Update src/transformers/generation/logits_process.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/generation/logits_process.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * updated the results --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Yupeng Jia authored
* Update modeling_deformable_detr.py Fix bugs for two stage training * Update modeling_deformable_detr.py * Add test_two_stage_training to DeformableDetrModelTest --------- Co-authored-by:yupeng.jia <yupeng.jia@momenta.ai>
-
Sourab Mangrulkar authored
* resolving zero3 init when using accelerate config with Trainer * refactor * fix * fix import
-
YQ authored
* add pathname and lineno to logging formatter in debug mode * use TRANSFORMERS_VERBOSITY="detail" to print pathname and lineno
-
YQ authored
* add test for `get_keys_to_not_convert` * add minimum patch to keep mpt lm_head from 8bit quantization * add reivsion to
-
Sylvain Gugger authored
-
- 01 Aug, 2023 5 commits
-
-
amyeroberts authored
Move dtype recasting to match torchvision ToTensor
-
Younes Belkada authored
* fix detr weird issue * Update src/transformers/models/conditional_detr/modeling_conditional_detr.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * fix copies * fix copies --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Younes Belkada authored
[`Docs`/`quantization`] Clearer explanation on how things works under the hood. + remove outdated info (#25216) * clearer explanation on how things works under the hood. * Update docs/source/en/main_classes/quantization.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/main_classes/quantization.md Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add `load_in_4bit` in `from_pretrained` --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Younes Belkada authored
* fix pix2struct cross attention * fix torchscript slow test
-
Wang, Yi authored
make build_mpt_alibi_tensor a method of MptModel so that deepspeed could override it to make autoTP work Signed-off-by:Wang, Yi A <yi.a.wang@intel.com>
-
- 31 Jul, 2023 4 commits
-
-
Younes Belkada authored
* revisit device check for pipeline * let's raise an error.
-
Younes Belkada authored
wrap `cuda` and `to` method correctly
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
- 28 Jul, 2023 7 commits
-
-
amyeroberts authored
* Fix rescaling bug * Add tests * Update integration tests * Fix up * Update src/transformers/image_transforms.py * Update test - new possible order in list
-
Sanchit Gandhi authored
* move to device * update with cuda values * fix fp16 * more rigorous
-
Yoni Gottesman authored
-
Alexander Markov authored
* added compiled model support for inference * linter * Fix tests * linter * linter * remove inference mode from pipelines * Linter --------- Co-authored-by:amarkov <alexander@inworld.ai>
-
jiqing-feng authored
Fix jit trace
-
YQ authored
-
Lucain authored
* Fix .push_to_hub and cleanup get_full_repo_name usage * Do not rely on Python bool conversion magic * request changes
-
- 27 Jul, 2023 4 commits
-
-
Sylvain Gugger authored
-
Sanchit Gandhi authored
* First commit * step 1 working * add alibi * placeholder for `scan` * add matrix mult alibi * beta scaling factor for bmm * working v1 - simple forward pass * move layer_number from attribute to arg in call * partial functioning scan * hacky working scan * add more modifs * add test * update scan for new kwarg order * fix position_ids problem * fix bug in attention layer * small fix - do the alibi broadcasting only once * prelim refactor * finish refactor * alibi shifting * incorporate dropout_add to attention module * make style * make padding work again * update * remove bogus file * up * get generation to work * clean code a bit * added small tests * adding albii test * make CI tests pass: - change init weight - add correct tuple for output attention - add scan test - make CI tests work * fix few nits * fix nit onnx * fix onnx nit * add missing dtype args to nn.Modules * remove debugging statements * fix scan generate * Update modeling_flax_bloom.py * Update test_modeling_flax_bloom.py * Update test_modeling_flax_bloom.py * Update test_modeling_flax_bloom.py * fix small test issue + make style * clean up * Update tests/models/bloom/test_modeling_flax_bloom.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * fix function name * small fix test * forward contrib credits from PR17761 * Fix failing test * fix small typo documentation * fix non passing test - remove device from build alibi * refactor call - refactor `FlaxBloomBlockCollection` module * make style * upcast to fp32 * cleaner way to upcast * remove unused args * remove layer number * fix scan test * make style * fix i4 casting * fix slow test * Update src/transformers/models/bloom/modeling_flax_bloom.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove `layer_past` * refactor a bit * fix `scan` slow test * remove useless import * major changes - remove unused code - refactor a bit - revert import `torch` * major refactoring - change build alibi * remove scan * fix tests * make style * clean-up alibi * add integration tests * up * fix batch norm conversion * style * style * update pt-fx cross tests * update copyright * Update src/transformers/modeling_flax_pytorch_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * per-weight check * style * line formats --------- Co-authored-by:
younesbelkada <younesbelkada@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com> Co-authored-by:
haileyschoelkopf <haileyschoelkopf@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
* fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yoach Lacombe authored
* initial Bark offload proposal * use hooks instead of manually offloading * add test of bark offload to cpu feature * Apply nit suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docstrings of offload Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove unecessary set_seed in Bark tests --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
-