- 05 Jan, 2021 5 commits
-
-
Stas Bekman authored
* --model_parallel hasn't been implemented for most models * make the help clear as well * implement is_parallelizable; use it * oops * remove property
-
Julien Plu authored
-
Stas Bekman authored
This PR proposes to: * auto-flush `transformers` logging When using logging for tracing signals from different parts of the code and which could be mixed with print debug this aids to get all the logging events synchronized. I don't think this change will introduce any performance impacts. If it helps someone here is the code I used to sync `transformers` logging with various other debug prints. I was porting bart to MP and I needed to trace that the device switching happens correctly and I added a bunch of logger.info calls inside `modeling_bart.py` and also had some other helpers `print` debug messages which weren't logger based: ``` # auto flush std streams from sys import stdout, stderr def stdout_write_flush(args, w=stderr.write): w(args); stderr.flush() def stderr_write_flush(args, w=stderr.write): w(args); stderr.flush() stdout.write = stdout_write_flush stderr.write = stderr_write_flush from transformers import BartTokenizer, BartForConditionalGeneration, BartConfig import logging import transformers.utils.logging import transformers.models.bart.modeling_bart # I wanted a shorter simpler format handlers = transformers.utils.logging._get_library_root_logger().handlers for handler in handlers: formatter = logging.Formatter("[%(funcName)s] %(message)s") handler.setFormatter(formatter) transformers.models.bart.modeling_bart.logger.setLevel(transformers.logging.INFO) ``` @LysandreJik, @sgugger, @patrickvonplaten -
Julien Plu authored
* Fix longformer * Apply style * Remove serving content * Forgot a condition * Apply style * Address Patrick's comments * Fix dtype
-
Boris Dayma authored
* feat(wandb): log artifacts * fix: typo * feat(wandb): ensure name is allowed * feat(wandb): log artifact * feat(wandb): saving logic * style: improve formatting * fix: unrelated typo * feat:聽use a fake trainer * fix:聽simplify * feat(wandb): log model files as artifact * style: fix style * docs(wandb): correct description * feat: unpack model + allow env Truethy values * feat: TrainerCallback can access tokenizer * style:聽fix style * feat(wandb): log more interesting metadata * feat: unpack tokenizer * feat(wandb): metadata with load_best_model_at_end * feat(wandb): more robust metadata * style(wandb): fix formatting
-
- 04 Jan, 2021 13 commits
-
-
Stas Bekman authored
-
Qbiwan authored
* bertweet docs coverage * style doc max len 119 * maxlen style rst * run main() from style_doc * changed according to comments
-
Stas Bekman authored
-
Patrick von Platen authored
-
Julien Plu authored
* Fix DPR * Keep usual models * Apply style * Address Sylvain's comments
-
Julien Plu authored
-
Stas Bekman authored
This PR: * fixes trainer to have the logger agree with the actual default `output_dir`, but setting it one place and passing it as an argument to both places @sgugger
-
Julien Plu authored
-
dependabot[bot] authored
Bumps [notebook](https://github.com/jupyter/jupyterhub) from 6.1.4 to 6.1.5. - [Release notes](https://github.com/jupyter/jupyterhub/releases) - [Changelog](https://github.com/jupyterhub/jupyterhub/blob/master/CHECKLIST-Release.md) - [Commits](https://github.com/jupyter/jupyterhub/commits ) Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Sylvain Gugger authored
-
Julien Plu authored
-
Charles authored
* add get_cached_models function * add List type to import * fix code quality * Update src/transformers/file_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/file_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/file_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/file_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/file_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Fix style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sam Shleifer authored
-
- 03 Jan, 2021 1 commit
-
-
Yoshitomo Matsubara authored
* fix a bug in eval_batch_retrieval * should return parser as well as other staticmethod * remove duplicate argument * these kwargs are no longer accepted (cause TypeError in self.generator.generate of modeling_rag.py) * fixed file paths in README * moved an arg to add_ray_specific_args
-
- 02 Jan, 2021 3 commits
-
-
Chris Kennedy authored
-
Patrick von Platen authored
* push * make style
-
Derrick Blakely authored
-
- 30 Dec, 2020 1 commit
-
-
Stas Bekman authored
-
- 29 Dec, 2020 3 commits
-
-
Stas Bekman authored
``` python -c "from apex.normalization import FusedProphetNetLayerNorm" Traceback (most recent call last): File "<string>", line 1, in <module> ImportError: cannot import name 'FusedProphetNetLayerNorm' from 'apex.normalization' (/home/stas/anaconda3/envs/main-38/lib/python3.8/site-packages/apex/normalization/__init__.py) ``` It looks like this code has never been tested, so it silently fails inside try/except. Discovered this by accident in https://github.com/huggingface/transformers/issues/9338#issuecomment-752217708
-
Patrick von Platen authored
-
Patrick von Platen authored
-
- 28 Dec, 2020 3 commits
-
-
Julien Plu authored
-
Julien Plu authored
* Fix T5 * Fix test * Fix test
-
Patrick von Platen authored
[Seq2Seq Templates] Correct some TF-serving errors and add gradient checkpointing to PT by default. (#9334) * correct tests * correct shape and get_tf_activation * more correction tf * add gradient checkpointing to templates * correct typo
-
- 27 Dec, 2020 1 commit
-
-
Patrick von Platen authored
-
- 25 Dec, 2020 2 commits
-
-
Patrick von Platen authored
* correct gpt2 * fix gpt2 * fix use_cache ordering * correct past tolerance * fix for all cases * style
-
Vasudev Gupta authored
* Created using Colaboratory * mbart-training examples add * link add * Update description Co-authored-by:Suraj Patil <surajp815@gmail.com>
-
- 24 Dec, 2020 8 commits
-
-
Patrick von Platen authored
* fix bart doc * fix docs
-
Bram Vanroy authored
Missing "s" typo
-
Daniele Sartiano authored
* Update modeling_encoder_decoder.py Fixed typo. * typo Co-authored-by:Suraj Patil <surajp815@gmail.com>
-
Ratthachat (Jung) authored
* Create modeling_tf_dpr.py * Add TFDPR * Add back TFPegasus, TFMarian, TFMBart, TFBlenderBot last commit accidentally deleted these 4 lines, so I recover them back * Add TFDPR * Add TFDPR * clean up some comments, add TF input-style doc string * Add TFDPR * Make return_dict=False as default * Fix return_dict bug (in .from_pretrained) * Add get_input_embeddings() * Create test_modeling_tf_dpr.py The current version is already passed all 27 tests! Please see the test run at : https://colab.research.google.com/drive/1czS_m9zy5k-iSJbzA_DP1k1xAAC_sdkf?usp=sharing * fix quality * delete init weights * run fix copies * fix repo consis * del config_class, load_tf_weights They shoud be 'pytorch only' * add config_class back after removing it, test failed ... so totally only removing "use_tf_weights = None" on Lysandre suggestion * newline after .. note:: * import tf, np (Necessary for ModelIntegrationTest) * slow_test from_pretrained with from_pt=True At the moment we don't have TF weights (since we don't have official official TF model) Previously, I did not run slow test, so I missed this bug * Add simple TFDPRModelIntegrationTest Note that this is just a test that TF and Pytorch gives approx. the same output. However, I could not test with the official DPR repo's output yet * upload correct tf model * remove position_ids as missing keys * fix RagSeq generate with context_input_ids fix RagSeq generate with context_input_ids * apply style * delete unused lines * Add test_rag_sequence_generate_batch_from_context_input_ids * Readability improved * stylying * Stylize * typos * add check_model_generate_from_context_input_ids * make style * Apply suggestions from code review * make style2 Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
patrickvonplaten <patrick@huggingface.co>
-
Suraj Patil authored
-
Jungwhan authored
-
Jethro Kuan authored
Fixes #9244 Co-authored-by:Jethro Kuan <jethro.kuan@bytedance.com>
-
Patrick von Platen authored
* adapt templates * adapt config * add test as well * fix output type * fix cache false naming * finish tests * last fix
-