"docs/source/en/training.md" did not exist on "77321481247787c97568c3b9f64b19e22351bab8"
- 24 Aug, 2022 3 commits
-
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Constantin H眉tterer authored
* Add minor doc-string change to include hp_name * fix: missing type-information for kwargs * fix: missing white-space in hyperparameter_search doc-strings
-
Mishig Davaadorj authored
-
- 23 Aug, 2022 4 commits
-
-
Joao Gante authored
-
SaulLu authored
* improve add_tokens documentation * format
-
Nicolas Patry authored
and friends.
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 22 Aug, 2022 3 commits
-
-
Atharva Ingle authored
-
tgadeliya authored
-
Yih-Dar authored
Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
- 19 Aug, 2022 3 commits
-
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Patrick von Platen authored
* add first generation tutorial * [Circle CI] Temporary fix for broken detectron2 import * remove generation
-
Joao Gante authored
-
- 18 Aug, 2022 12 commits
-
-
Atharva Ingle authored
* `model.tie_weights()` should be applied after `accelerator.prepare` Weight tying should be done after the model has been moved to XLA device as mentioned on PyTorch/XLA Troubleshooting guide [here](https://github.com/pytorch/xla/blob/master/TROUBLESHOOTING.md#xla-tensor-quirks) * format code
-
Loubna Ben Allal authored
* add examples subfolder * mention examples in codeparrot readme * use Trainer optimizer and scheduler type and add output_dir as argument * add example of text-to-python and python-to-text models * mention the downstream examples in the readme * fix typo
-
Younes Belkada authored
* fix bnb documentation - move bnb documentation to `infer_gpu_many` * small refactoring - added text on infer_gpu_one - added a small note on infer_gpu_many - added customized multi gpu example on infer_gpu_many * Update docs/source/en/perf_infer_gpu_many.mdx Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * apply suggestions Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Zachary Mueller authored
-
Severin Simmler authored
* Fix quantization * Save model * Remove unused comments * Fix formatting
-
lewtun authored
-
regisss authored
-
amyeroberts authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Patrick von Platen authored
* add first generation tutorial * [LongT5 Docs] Correct docs * correct expected string * remove incorrect file
-
Matt authored
* Allow users to force TF availability * Correctly name the envvar!
-
- 17 Aug, 2022 4 commits
-
-
amyeroberts authored
* Update methods to optionally rescale This is necessary to allow for casting our images / videos to numpy arrays within the feature extractors' call. We want to do this to make sure the behaviour is as expected when flags like are False. If some transformations aren't applied, then the output type can't be unexpected e.g. a list of PIL images instead of numpy arrays. * Cast images to numpy arrays in call to enable consistent behaviour with different configs * Remove accidental clip changes * Update tests to reflect the scaling logic We write a generic function to handle rescaling of our arrays. In order for the API to be intuitive, we take some factor c and rescale the image values by that. This means, the rescaling done in normalize and to_numpy_array are now done with array * (1/255) instead of array / 255. This leads to small differences in the resulting image. When testing, this was in the order of 1e-8, and so deemed OK
-
Jingya HUANG authored
-
Yih-Dar authored
Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Stefan Schweter authored
* examples: add Bloom support for token classification (FLAX, PyTorch and TensorFlow) * examples: remove support for Bloom in token classication (FLAX and TensorFlow currently have no support for it)
-
- 16 Aug, 2022 7 commits
-
-
Younes Belkada authored
* bnb minor modifications - refactor documentation - add troubleshooting README - add PyPi library on DockerFile * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Apply suggestions from code review * Apply suggestions from code review * Apply suggestions from code review * put in one block - put bash instructions in one block * update readme - refactor a bit hardware requirements * change text a bit * Apply suggestions from code review Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * apply suggestions Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * add link to paper * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Update tests/mixed_int8/README.md * Apply suggestions from code review * refactor a bit * add instructions Turing & Amperer Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * add A6000 * clarify a bit * remove small part * Update tests/mixed_int8/README.md Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com>
-
zhoutang776 authored
* Update run_translation_no_trainer.py found an error in selecting `no_decay` parameters and some small modifications when the user continues to train from a checkpoint * fixs `no_decay` and `resume_step` issue 1. change `no_decay` list 2. if use continue to train their model from provided checkpoint, the `resume_step` will not be initialized properly if `args.gradient_accumulation_steps != 1`
-
flozi00 authored
-
Joao Gante authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sourab Mangrulkar authored
* mac m1 `mps` integration * Update docs/source/en/main_classes/trainer.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * addressing comments * Apply suggestions from code review Co-authored-by:
Dan Saattrup Nielsen <47701536+saattrupdan@users.noreply.github.com> * resolve comment Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Dan Saattrup Nielsen <47701536+saattrupdan@users.noreply.github.com>
-
- 14 Aug, 2022 1 commit
-
-
Karim Foda authored
* [Flax] Add remat (gradient checkpointing) * fix variable naming in test * flip: checkpoint using a method * fix naming * fix class naming * apply PVP's suggestions from code review * add gradient_checkpointing to examples * Add gradient_checkpointing to run_mlm_flax * Add remat to longt5 * Add gradient checkpointing test longt5 * Fix args errors * Fix remaining tests * Make fixup & quality fixes * replace kwargs * remove unecessary kwargs * Make fixup changes * revert long_t5_flax changes * Remove return_dict and copy to LongT5 * Remove test_gradient_checkpointing Co-authored-by:sanchit-gandhi <sanchit@huggingface.co>
-
- 12 Aug, 2022 3 commits
-
-
Younes Belkada authored
-
Stas Bekman authored
* [fsmt] deal with -100 indices in decoder ids Fixes: https://github.com/huggingface/transformers/issues/17945 decoder ids get the default index -100, which breaks the model - like t5 and many other models add a fix to replace -100 with the correct pad index. For some reason this use case hasn't been used with this model until recently - so this issue was there since the beginning it seems. Any suggestions to how to add a simple test here? or perhaps we have something similar already? user's script is quite massive. * style
-
Stas Bekman authored
the manual anchors end up being duplicated with automatically added anchors and no longer work.
-