"...gpu/git@developer.sourcefind.cn:gaoqiong/migraphx.git" did not exist on "7668ef6b75a5befa9eb8516257aec8c00e632f56"
- 19 Mar, 2020 23 commits
-
-
Nitish Shirish Keskar authored
torch.cuda.empty_cache() was being called from a TF function (even when torch is unavailable) not sure any replacement is needed if TF OOMs
-
Julien Chaumond authored
-
Kyeongpil Kang authored
I found there are two grammar errors or typo issues in the explanation of the encoding properties. The original sentences: If your was made of multiple \"parts\" such as (question, context), then this would be a vector with for each token the segment it belongs to If your has been truncated into multiple subparts because of a length limit (for BERT for example the sequence length is limited to 512), this will contain all the remaining overflowing parts. I think "input" should be inserted after the phrase "If your".
-
Patrick von Platen authored
* fix conflicts * update bart max length test * correct spelling mistakes * implemented model specific encode function * fix merge conflicts * better naming * save intermediate state -> need to rethink strucuture a bit * leave tf problem as it is for now * current version * add layers.pop * remove ipdb * make style * clean return cut decoding * remove ipdbs * Fix restoring layers in the decoders that doesnt exists. * push good intermediate solution for now * fix conflicts * always good to refuse to merge conflicts when rebasing * fix small bug * improve function calls * remove unused file * add correct scope behavior for t5_generate Co-authored-by:Morgan Funtowicz <funtowiczmo@gmail.com>
-
Julien Chaumond authored
-
husein zolkepli authored
-
Manuel Romero authored
-
Manuel Romero authored
Create card for BERT-Mini finetuned on SQuAD v2
-
Manuel Romero authored
- Only 17MB of Model weights!!
-
Manuel Romero authored
-
Antti Virtanen authored
* Add a model card for FinBERT This is a copy of https://github.com/TurkuNLP/FinBERT/blob/master/README.md. * Added a file for uncased. * Add metadata for cased. * Added metadata for uncased.
-
Lysandre Debut authored
-
Kyeongpil Kang authored
For the tutorial of "How to generate text", the URL link was wrong (it was linked to the tutorial of "How to train a language model"). I fixed the URL.
-
Serkan Karakulak authored
* added return_token_type_ids argument for tokenizers which do not generate return_type_ids by default * fixed styling * Style Co-authored-by:LysandreJik <lysandre.debut@reseau.eseo.fr>
-
Sam Shleifer authored
* config.activation_function
-
mataney authored
* solving bug where for small epochs and large gradient_accumulation_steps we never train * black formatting * no need to change these files
-
Sam Shleifer authored
-
Mohamed El-Geish authored
`T5Tokenizer` instead of `XLNetTokenizer`
-
Matthew Goldey authored
-
Patrick von Platen authored
* fix issue 3289 * fix attention mask if input_ids None behavior
-
Junyi_Li authored
-
Junyi_Li authored
roberta_chinese_base card
-
Junyi_Li authored
albert_chinese_tiny card
-
- 18 Mar, 2020 8 commits
-
-
Kyle Lo authored
* Create README.md * model card * add model card for cased
-
Morgan Funtowicz authored
Remove hardcoded mask_token and use the value provided by the tokenizer.
-
Branden Chan authored
-
Lysandre Debut authored
* XLM-R now passes common tests + Integration tests * Correct mask index * Model input names * Style * Remove text preprocessing * Unneccessary import
-
Patrick von Platen authored
Adding LM Head to Transfo-XL and first step to fixing problem with Adaptive Embeddings in TransfoXL (#3286) * first commit * work in progress * make language generation task pass * update to working version for LM * delete print * remove dead code * make style
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Junyi_Li authored
-
- 17 Mar, 2020 9 commits
-
-
Sam Shleifer authored
* passing * Undo stupid chg * docs * undo rename * delete-cruft * only import if you have torch * Dont rely on dict ordering * Fix dict ordering upstream * docstring link * docstring link * remove trailing comma for 3.5 compat * new name * delegate kwarging * Update kwargs
-
J.P Lee authored
* Update examples/ner/run_ner.py to use AutoModel * Fix missing code and apply `make style` command
-
Manuel Romero authored
-
Julien Chaumond authored
-
Nathan Raw authored
*
✨ Alter base pl transformer to use automodels *🐛 Add batch size env variable to function call *💄 Apply black code style from Makefile *🚚 Move lightning base out of ner directory *✨ Add lightning glue example *💄 self * move _feature_file to base class *✨ Move eval logging to custom callback *💄 Apply black code style *🐛 Add parent to pythonpath, remove copy command *🐛 Add missing max_length kwarg -
Patrick von Platen authored
* change do_samples back * None better default as boolean * adapt do_sample to True in test example * make style
-
Thomas Wolf authored
* memory benchmark rss * have both forward pass and line-by-line mem tracing * cleaned up tracing * refactored and cleaning up API * no f-strings yet... * add GPU mem logging * fix GPU memory monitoring * style and quality * clean up and doc * update with comments * Switching to python 3.6+ * fix quality
-
Jannes authored
* Create README.md * Updated README.md
-
Julien Chaumond authored
-