"examples/vscode:/vscode.git/clone" did not exist on "725a56329d304bb997fc3cf3013c908a20ae50a0"
- 31 May, 2023 26 commits
-
-
Sylvain Gugger authored
* Re-enable squad test * [all-test] * [all-test] Fix all test command * Fix the all-test
-
Sourab Mangrulkar authored
remove the extra `accelerator.prepare` that slipped in with multiple update from main
😅 -
amyeroberts authored
Bug fix - flip_channel_order for channels_first
-
Sylvain Gugger authored
* Try easy first * Add an empty job * Fix name * Fix method
-
amyeroberts authored
Raise error if loss can't be calculated
-
Hari authored
* add conditional statement for auxiliary loss calculation * fix style and copies
-
Younes Belkada authored
fix RWKV 4bit
-
Zachary Mueller authored
* Upgrade safetensors * Second table
-
Connor Henderson authored
fix: Replace `add_prefix_space` in `get_prompt_ids` with manual space for FastTokenizer compatibility (#23796) * add ' ' replacement for add_prefix_space * add fast tokenizer test
-
Zachary Mueller authored
* Move import check to before state reset * Guard better
-
Younes Belkada authored
* add warning for gpt2-like models * more details * adapt from suggestions
-
Sanchit Gandhi authored
* fix for ragged list * unpin numba * make style * np.object -> object * propagate changes to tokenizer as well * np.long -> "long" * revert tokenization changes * check with tokenization changes * list/tuple logic * catch numpy * catch else case * clean up * up * better check * trigger ci * Empty commit to trigger CI
-
Xinyu Yang authored
* ensure banned_mask and indices in same device * ensure banned_mask and indices in same device switch the order in which indices and banned_mask are created and create banned_mask on the proper device
-
Thomas Wang authored
* Suport shared storage * Really be sure we have the same storage * Make style * - Refactor storage identifier mechanism - Group everything into a single for loop * Make style * PR * make style * Update src/transformers/pytorch_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
-
Calico authored
-
Sylvain Gugger authored
-
Sourab Mangrulkar authored
* mixed precision support via accelerate * fix issues * fix for the sharded ddp case * fix flax and tf failing tests * `refactor the place to create `Accelerator` object * move ddp prep to accelerate * fix
😅 * resolving comments * move fsdp handling to accelerate * fixex * fix saving * shift torch dynamo handling to accelerate * shift deepspeed integration and save & load utils to accelerate * fix accelerate launcher support * oops * fix🐛 * save ckpt fix * Trigger CI * nasty🐛 😅 * as deepspeed needs grad_acc fixes, transfer grad_acc to accelerate * make tests happy * quality✨ * loss tracked needs to account for grad_acc * fixing the deepspeed tests * quality✨ *😅 😅 😅 * tests😡 * quality✨ * Trigger CI * resolve comments and fix the issue with the previous merge from branch * Trigger CI * accelerate took over deepspeed integration --------- Co-authored-by:Stas Bekman <stas@stason.org>
-
Denisa Roberts authored
* Add tf code for efficientformer * Fix return dict bug - return last hidden state after last stage * Fix corresponding return dict bug * Override test tol * Change default values of training to False * Set training to default False X3 * Rm axis from ln * Set init in dense projection * Rm debug stuff * Make style; all tests pass. * Modify year to 2023 * Fix attention biases codes * Update the shape list logic * Add a batch norm eps config * Remove extract comments in test files * Add conditional attn and hidden states return for serving output * Change channel dim checking logic * Add exception for withteacher model in training mode * Revert layer count for now * Add layer count for conditional layer naming * Transpose for conv happens only in main layer * Make tests smaller * Make style * Update doc * Rm from_pt * Change to actual expect image class label * Remove stray print in tests * Update image processor test * Remove the old serving output logic * Make style * Make style * Complete test
-
Sylvain Gugger authored
-
Sam Passaglia authored
* add \n * removed copied from header
-
Sourab Mangrulkar authored
* mixed precision support via accelerate * fix issues * fix for the sharded ddp case * fix flax and tf failing tests * `refactor the place to create `Accelerator` object * move ddp prep to accelerate * fix
😅 * resolving comments * move fsdp handling to accelerate * fixex * fix saving * shift torch dynamo handling to accelerate -
Sourab Mangrulkar authored
* mixed precision support via accelerate * fix issues * fix for the sharded ddp case * fix flax and tf failing tests * `refactor the place to create `Accelerator` object * move ddp prep to accelerate * fix
😅 * resolving comments * move fsdp handling to accelerate * fixex * fix saving -
Sohyun Sim authored
* docs: ko: pad_truncation.mdx * feat: manual draft * fix: resolve suggestions Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> --------- Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com>
-
Sourab Mangrulkar authored
* mixed precision support via accelerate * fix issues * fix for the sharded ddp case * fix flax and tf failing tests * `refactor the place to create `Accelerator` object * move ddp prep to accelerate * fix
😅 * resolving comments -
Sourab Mangrulkar authored
* mixed precision support via accelerate * fix issues * fix for the sharded ddp case * fix flax and tf failing tests * `refactor the place to create `Accelerator` object * address comments by removing debugging print statements
-
- 30 May, 2023 14 commits
-
-
Abhinav Patil authored
Adds support for AutoProcessor.from_pretrained to MCTCTProcessor models
-
George authored
* Editing issue with pickle def with lambda function * fix type * Made helper function private * delete tab --------- Co-authored-by:georgebredis <9454-georgebredis@users.noreply.gitlab.aicrowd.com>
-
Arthur authored
* Better warning * Update src/transformers/modeling_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * format line --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Vijeth Moudgalya authored
-
Arthur authored
* Update the processor when changing add_eos and add_bos * fixup * update * add a test * fix failing tests * fixup
-
Clémentine Fourrier authored
-
peridotml authored
* initial flyte callback * lint * logs should still be saved to Flyte even if pandas isn't install (unlikely) * cr - flyte team * add docs for Flytecallback * fix doc string - cr sgugger * Apply suggestions from code review cr - sgugger fix doc strings Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Hyeonseo Yun authored
* docs: ko: troubleshooting.mdx * revised: fix _toctree.yml #23112 * feat: nmt draft `troubleshooting.mdx` * fix: manual edits `troubleshooting.mdx` * revised: resolve suggestions troubleshooting.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> --------- Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com>
-
Kihoon Son authored
* task/video_classification translated Co-Authored-By:
Hyeonseo Yun <0525_hhgus@naver.com> Co-Authored-By:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-Authored-By:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-Authored-By:
Nayeon Han <nayeon2.han@gmail.com> Co-Authored-By:
Wonhyeong Seo <wonhseo@kakao.com> Co-Authored-By:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/tasks/video_classification.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> * Update video_classification.mdx * Update _toctree.yml * Update _toctree.yml * Update _toctree.yml * Update _toctree.yml --------- Co-authored-by:
Hyeonseo Yun <0525_hhgus@naver.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Nayeon Han <nayeon2.han@gmail.com> Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com>
-
Kihoon Son authored
* docs: ko: fast_tokenizer.mdx content - translated Co-Authored-By:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-Authored-By:
Nayeon Han <nayeon2.han@gmail.com> Co-Authored-By:
Hyeonseo Yun <0525_hhgus@naver.com> Co-Authored-By:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-Authored-By:
Jungnerd <46880056+jungnerd@users.noreply.github.com> Co-Authored-By:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * Update docs/source/ko/fast_tokenizers.mdx Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> * Update fast_tokenizers.mdx * Update fast_tokenizers.mdx * Update fast_tokenizers.mdx * Update fast_tokenizers.mdx * Update _toctree.yml --------- Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Nayeon Han <nayeon2.han@gmail.com> Co-authored-by:
Hyeonseo Yun <0525_hhgus@naver.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com>
-
Matthijs Hollemans authored
* move input features to GPU * skip these tests because undefined behavior * unskip tests
-
Matt authored
SAM shape flexibility fixes for compilation
-
Samin Yasar authored
* add type hint in pipeline model argument * add pretrainedmodel and tfpretainedmodel type hint * make type hints string
-
Eli Simhayev authored
* ran `transformers-cli add-new-model-like` * added `AutoformerLayernorm` and `AutoformerSeriesDecomposition` * added `decomposition_layer` in `init` and `moving_avg` to config * added `AutoformerAutoCorrelation` to encoder & decoder * removed caninical self attention `AutoformerAttention` * added arguments in config and model tester. Init works!
😁 * WIP autoformer attention with autocorrlation * fixed `attn_weights` size * wip time_delay_agg_training * fixing sizes and debug time_delay_agg_training * aggregation in training works!😁 * `top_k_delays` -> `top_k_delays_index` and added `contiguous()` * wip time_delay_agg_inference * finish time_delay_agg_inference😎 * added resize to autocorrelation * bug fix: added the length of the output signal to `irfft` * `attention_mask = None` in the decoder * fixed test: changed attention expected size, `test_attention_outputs` works! * removed unnecessary code * apply AutoformerLayernorm in final norm in enc & dec * added series decomposition to the encoder * added series decomp to decoder, with inputs * added trend todos * added autoformer to README * added to index * added autoformer.mdx * remove scaling and init attention_mask in the decoder * make style * fix copies * make fix-copies * inital fix-copies * fix from https://github.com/huggingface/transformers/pull/22076 * make style * fix class names * added trend * added d_model and projection layers * added `trend_projection` source, and decomp layer init * added trend & seasonal init for decoder input * AutoformerModel cannot be copied as it has the decomp layer too * encoder can be copied from time series transformer * fixed generation and made distrb. out more robust * use context window to calculate decomposition * use the context_window for decomposition * use output_params helper * clean up AutoformerAttention * subsequences_length off by 1 * make fix copies * fix test * added init for nn.Conv1d * fix IGNORE_NON_TESTED * added model_doc * fix ruff * ignore tests * remove dup * fix SPECIAL_CASES_TO_ALLOW * do not copy due to conv1d weight init * remove unused imports * added short summary * added label_length and made the model non-autoregressive * added params docs * better doc for `factor` * fix tests * renamed `moving_avg` to `moving_average` * renamed `factor` to `autocorrelation_factor` * make style * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * fix configurations * fix integration tests * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * fixing `lags_sequence` doc * Revert "fixing `lags_sequence` doc" This reverts commit 21e34911e36a6f8f45f25cbf43584a49e5316c55. * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * model layers now take the config * added `layer_norm_eps` to the config * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * added `config.layer_norm_eps` to AutoformerLayernorm * added `config.layer_norm_eps` to all layernorm layers * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * fix variable names * added inital pretrained model * added use_cache docstring * doc strings for trend and use_cache * fix order of args * imports on one line * fixed get_lagged_subsequences docs * add docstring for create_network_inputs * get rid of layer_norm_eps config * add back layernorm * update fixture location * fix signature * use AutoformerModelOutput dataclass * fix pretrain config * no need as default exists * subclass ModelOutput * remove layer_norm_eps config * fix test_model_outputs_equivalence test * test hidden_states_output * make fix-copies * Update src/transformers/models/autoformer/configuration_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * removed unused attr * Update tests/models/autoformer/test_modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/autoformer/modeling_autoformer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * use AutoFormerDecoderOutput * fix formatting * fix formatting --------- Co-authored-by:
Kashif Rasul <kashif.rasul@gmail.com> Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-