"vscode:/vscode.git/clone" did not exist on "6af0854efa3693e0b38c936707966685ec3d0ae8"
- 11 Apr, 2022 17 commits
-
-
Yih-Dar authored
* add error message * Use names in the error message * allow ModelOutput * rename to check_pt_tf_outputs and move outside * fix style * skip past_key_values in a better way * Add comments * improve code for label/loss * make the logic clear by moving the ignore keys out * fix _postprocessing_to_ignore * fix _postprocessing_to_ignore: create new outputs from the remaining fields * ignore past_key_values in TFGPT2 models for now * make check_pt_tf_outputs better regarding names * move check_pt_tf_models outside * rename methods * remove test_pt_tf_model_equivalence in TFCLIPModelTest * Reduce TFViTMAEModelTest.test_pt_tf_model_equivalence * move prepare_pt_inputs_from_tf_inputs outside check_pt_tf_models * Fix quality * Clean-up TFLxmertModelTester.test_pt_tf_model_equivalence * Fix quality * fix * fix style * Clean-up TFLEDModelTest.test_pt_tf_model_equivalence * Fix quality * add docstring * improve comment Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* update * batch_size -> text_batch_size Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Nicholas Broad authored
* private repo argument to trainer * format Co-authored-by:Nicholas Broad <nicholas@nmbroad.com>
-
Zachary Mueller authored
Adds checkpoint prefixes to the gitignore if `push_to_hub` is used along with `checkpointint_steps`
-
Yih-Dar authored
* update _create_and_check_torchscript * Enable test_torchscript * clear_class_registry Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Suraj Patil authored
-
Ahmed Elnaggar authored
* Fix t5 shard on TPU Pods The current script doesn't work properly on a TPU pod because the global batch is not divided correctly per host. This pull request fixes this issue by dividing the global batch to each host before it is shared on each host. * fix style Co-authored-by:ahmed-elnaggar <ahmed.elnaggar@allianz.com>
-
Minh Chien Vu authored
* Add doctest BERT * make fixup * fix typo * change checkpoints * make fixup * define doctest output value, update doctest for mobilebert * solve fix-copies * update QA target start index and end index * change checkpoint for docs and reuse defined variable * Update src/transformers/models/bert/modeling_tf_bert.py Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * make fixup Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com>
-
Patrick von Platen authored
-
Sadra authored
I create an archive of older checkpoints during training the checkpoint has a name with `f"{checkpoint_prefix}-*.zip/.tar ` previously `glob(f"{checkpoint_prefix}-*")` takes all files/folders starting with the name checkpoint, and later `shutil.rmtree(checkpoint)` takes a folder name; since at some point it my get a zip file; it crashes training; adding this `if os.path.isdir(x)` allows only folders on `glob_checkpoints` -
Joao Gante authored
* min length must be smaller than max length * Update min_length in tests
-
Jia LI authored
* add simple multi gpu complet * add human_eval_multi_gpu * use copy strategy to distribute across gpu, to avoid padding * add doc string * update code style * use task id to arrange output * truncate input to avoid zero pad * Stop the copy mechanism * update style * restore copies to scale better in distributed mode * update style * replace human eval * Apply suggestions from code review 1. Tokenize all input at the same time 2. use attention_mask to get the input length 3. other small fixes Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * correct typo and update docstring * update code style * remove num sample division constraint * remove max len calculation * use accelerator.gather once to speed up * use accelerate set_seed; update accelerate version * correct gather bug Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com>
-
Yih-Dar authored
* Fix some doc examples Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
SaulLu authored
* update proto sentencepiece model * Revert "update proto sentencepiece model" This reverts commit b07f671747fec35773d0b3d4788b8b15aefa0229. * add check * add test * Revert "Revert "update proto sentencepiece model"" This reverts commit 46108257b8927b73627ec8f4f3eed53a95fc700d. * test for log level * test for log level 2 * warning at the warning level * clean * format * add explanation in docstring
-
- 08 Apr, 2022 9 commits
-
-
Steven Liu authored
*
✨ update audio examples with minds dataset *🖍 make style *🖍 minor fixes for doctests -
Stas Bekman authored
* [Trainer] tf32 arg doc * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Laura Hanu authored
-
Zachary Mueller authored
* Fixed some bugs involving saving during epochs * Added tests mimicking the existing examples tests * Added in json exporting to all `no_trainer` examples for consistency
-
Yih-Dar authored
* fix QA sample * For TF_QUESTION_ANSWERING_SAMPLE Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Alan Lee authored
The defalut value of `padding` in `DataCollatorWithPadding` is `True`, not `False`.
-
Johannes Kolbe authored
* add vit tf doctest with @add_code_sample_docstrings * add labels string back in Co-authored-by:Johannes Kolbe <johannes.kolbe@tech.better.team>
-
NielsRogge authored
* Add TapexTokenizer * Improve docstrings and provide option to provide answer * Remove option for pretokenized inputs * Add TAPEX to README * Fix copies * Remove option for pretokenized inputs * Initial commit: add tapex fine-tuning examples on both table-based question answering and table-based fact verification. * - Draft a README file for running the script and introducing some background. - Remove unused code lines in tabfact script. - Disable the deafult `pad_to_max_length` option which is memory-consuming. * * Support `as_target_tokenizer` function for TapexTokenizer. * Fix the do_lower_case behaviour of TapexTokenizer. * Add unit tests for target scenarios and cased/uncased scenarios for both source and target. * * Replace the label BartTokenizer with TapexTokenizer's as_target_tokenizer function. * Fix typos in tapex example README. * * fix the evaluation script - remove the property `task_name` * * Make the label space more clear for tabfact tasks * * Using a new fine-tuning script for tapex-base on tabfact. * * Remove the lowercase code outside the tokenizer - we use the tokenizer to control whether do_lower_case * Guarantee the hyper-parameter can be run without out-of-memory on 16GB card and report the new reproduced number on wikisql * * Remove the default tokenizer_name option. * Provide evaluation command. * * Support for WikiTableQuestion dataset. * Fix a typo in README. * * Fix the datasets's key name in WikiTableQuestions * Run make fixup and move test to folder * Fix quality * Apply suggestions from code review * Apply suggestions from code review Co-authored-by:
Suraj Patil <surajp815@gmail.com> * Apply suggestions from code review * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Apply some more suggestions from code review * Improve docstrings * Overwrite failing test * Improve comment in example scripts * Fix rebase * Add TAPEX to Auto mapping * Add TAPEX to auto config mappings * Put TAPEX higher than BART in auto mapping * Add TAPEX to doc tests Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MBP.localdomain> Co-authored-by:
SivilTaram <qianlxc@outlook.com> Co-authored-by:
Niels Rogge <nielsrogge@nielss-mbp.home> Co-authored-by:
Suraj Patil <surajp815@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
- 07 Apr, 2022 8 commits
-
-
Stefan Schweter authored
-
Francesco Saverio Zuppichini authored
* base model done * make style * done * added files * Apply suggestions from code review Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Trigger doc build * resolved conversations * resolved conversations * seer models * minor changes * minor changes * make fixup * glob variables * minor changes * fix copies * config when possibile * resolved conflicts * resolved conflicts * resolved conflicts * CI * conversion script for 10b param * fixed for 10b model * minor updates in the doc + make style * removed unused code * Apply suggestions from code review Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * removed unused code * removed unused code * updated modeling_utils from main Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <Sylvain.gugger@gmail.com>
-
Britney Muller authored
* Update README.md Support Image Updates the Support image linking to our EAP page (to give it a refresh + help avoid image fatigue). Slack thread checking in with #open-source-internal on this update (https://huggingface.slack.com/archives/C021H1P1HKR/p1648838903316709) * Compressed Updated Support image * Improves Support Image Logo + Height Updated the image based on logo + size feedback. Big thanks to Bibi for making quick edits to this image.
-
Francesco Saverio Zuppichini authored
* Updated _load_pretrained_model_low_mem to check if keys are in the stored state_dict * update after conversions
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Laura Vasquez-Rodriguez authored
* Add inputs vector to calculate metric method * Include inputs for evaluation metrics with backwards compatibility * Prevent inputs create OOM issue and documentation details * Update style and code documentation * Fix style formatting issues * Update files format with make style
-
NielsRogge authored
* Fix doc * Make fixup Co-authored-by:Niels Rogge <nielsrogge@nielss-mbp.home>
-
- 06 Apr, 2022 6 commits
-
-
Zachary Mueller authored
Adds logging and save/loading to the Accelerate scripts Co-authored-by:Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
-
Anmol Joshi authored
* Update modeling_mpnet.py * Update modeling_ctrl.py * formatting * Formatting * Formatting * annotated FSMT * Added annotations for LED * Added Annotations for M2M * Added annotations for nystromformer * Added annotations for OpenAI * Added annotations for RAG * Removed unused imports * fix isort errors * Removed inputs_embeds docstring, corrected original * flake8 fixes * doc-builder fixes
-
Joao Gante authored
* refactor TF beam search * refactored generate can now properly use attention masks * add force bos/eos logit processors
-
Stas Bekman authored
-
Lysandre Debut authored
-