- 17 Apr, 2023 5 commits
-
-
Matt authored
-
fpgaminer authored
-
Jungnerd authored
fix: docs: ko: sagemaker anchors and `_toctree.yml` Co-authored-by:
Hyeonseo Yun <0525_hhgus@naver.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Na Yeon Han <nayeon2.han@gmail.com> Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
Na Yeon Han authored
docs: ko: translated `custom_models.mdx` Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 15 Apr, 2023 1 commit
-
-
bcol authored
-
- 14 Apr, 2023 12 commits
-
-
oscar-garzon authored
-
amyeroberts authored
* Indexing fix - CLIP checkpoint conversion * Fix up
-
Joao Gante authored
-
Mayank Agarwal authored
* Fix word_ids hyperlink * Add suggested fix
-
Matt authored
* If EOS is None, don't add it to sequences * If EOS is None, don't add it to sequences
-
Sohyun Sim authored
* add ko preprocessing * translate preprocessing.mdx to korean * translate preprocessing.mdx * Update preprocessing.mdx Fixed the line 273 as below: 또한, 특징 추출기에 `sampling_rate` 인자를 추가하여 발생할 수 있는 조용한 오류(silent errors)를 더 잘 디버깅하는 것을 권장합니다. * translate Image part * translated preprocess.mdx * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * fixed translation --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Alexander Ljungberg authored
Fixed string format; better tokenizer message. Before: `Saving a {tokenizer_class} to {tokenizer_path}` After: `Saving a LlamaTokenizerFast to outdir.` -
Joao Gante authored
-
Joao Gante authored
-
Sayak Paul authored
* add: tokenizer training script for TF TPU LM training. * add: script for preparing the TFRecord shards. * add: sequence of execution to readme. * remove limit from the tfrecord shard name. * Add initial train_model.py * Add basic training arguments and model init * Get up to the point of writing the data collator * Pushing progress so far! * Complete first draft of model training code * feat: grouping of texts efficiently. Co-authored-by:
Matt <rocketknight1@gmail.com> * Add proper masking collator and get training loop working * fix: things. * Read sample counts from filenames * Read sample counts from filenames * Draft README * Improve TPU warning * Use distribute instead of distribute.experimental * Apply suggestions from code review Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com> * Modularize loading and add MLM probability as arg * minor refactoring to better use the cli args. * readme fillup. * include tpu and inference sections in the readme. * table of contents. * parallelize maps. * polish readme. * change script name to run_mlm.py * address PR feedback (round I). --------- Co-authored-by:
Matt <rocketknight1@gmail.com> Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com>
-
Hyeonseo Yun authored
* docs: ko: init: tasks/sequence_classification.mdx * docs: ko: revised: change voca in tasks/sequence_classification.mdx * docs: ko: revised: [RE] change voca in tasks/sequence_classification.mdx * docs: ko: revised: spell check and sentence naturally in tasks/sequence_classification.mdx * docs: ko: revised: spell check and consistent vocabulary in tasks/sequence_classification.mdx * docs: ko: revised: Add full stop and change voca in tasks/sequence_classification.mdx * docs: ko: revised: sync first section templates in tasks/sequence_classification.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * fix: revert use of full-stops to colons * colons are used to emphasize the code block that follows * @0525hhgus @wonhyeongseo docs: ko: revised: sync second section templates in tasks/sequence_classification.mdx Co-Authored-By:
Wonhyeong Seo <wonhseo@kakao.com> * docs: ko: revised: change 'train', 'finetuning' in tasks/sequence_classification.mdx --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
- 13 Apr, 2023 15 commits
-
-
Yih-Dar authored
* fix * style * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Ruiyang Sun authored
Bug in LlamaTokenizer when #22742
-
Stas Bekman authored
* [trainer] update url * style
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Gabriel Yang authored
translate training doc to Korean
-
Yih-Dar authored
* fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
NielsRogge authored
* Add model to doc tests * Remove generate and replace by prepare_inputs_for_generation * More fixes * Remove print statements * Update integration tests * Fix generate * Remove model from auto mapping * Use auto processor * Fix integration tests * Fix test * Add inference code snippet * Remove is_encoder_decoder * Update docs * Remove notebook link
-
Rinat authored
* Update modeling_vilt.py Vilt compatible with model parallelism * Update modeling_switch_transformers.py switch_transformers compatible with model parallelism
-
Joel Lamy-Poirier authored
Fix indexing
-
Elabonga Atuo authored
* added configuration file for mvp model * added configuration_mvp.py line to file
-
Elabonga Atuo authored
m2m-100-config for doctest
-
Sylvain Gugger authored
-
- 12 Apr, 2023 7 commits
-
-
Matt authored
* Fix docstrings for TFBLIP * Fix missing line in TF port! * Use values from torch tests now other bugs fixed * Use values from torch tests now other bugs fixed * Fix doctest string
-
NielsRogge authored
* Use different level * Remove futurewarning * Use warning_once * Update copies
-
Arthur authored
* add fast support and option * update based on review * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/llama/convert_llama_weights_to_hf.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * nit * add print * fixup --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Michael Benayoun authored
`torch.distributed` group initialization for `torch_neuron` disabled when `optimum-neuron` is installed (#22728) * Make the process group initialization not happen if optimum_neuron is installed * Add warning * Remove list and added warning
-
Stas Bekman authored
-
ARKA1112 authored
generator(model="openai/whisper-large") always returns error. As the error says the generator expects an input, just like the .flac file above. Even the generator object has no parameters called model. While there are parameters which can be passed to generator like 'batch_size' but to pass a model i believe the the parameter has to be passed while instantiating the pipeline and not as a parameter to the instance. I believe the correct term should be: generator = pipeline(model="openai/whisper-large", device=0)
-
Younes Belkada authored
* make serialization of int8 models possible * make fixup * add docs * add ability to push to hub and save pretrained * fixes * more addition * more tests * fix issues * change variable * clearer message * adapt from suggestions * few fixes * remove unused function * Update src/transformers/utils/quantization_config.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * address last comments * last warning * clarify doc * protect import * Update src/transformers/modeling_utils.py * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-