- 17 Sep, 2020 3 commits
-
-
Stas Bekman authored
-
Stas Bekman authored
* [model cards] ported allenai Deep Encoder, Shallow Decoder models * typo * fix references * add allenai/wmt19-de-en-6-6 model cards * fill-in the missing info for the build script as provided by the searcher.
-
Stas Bekman authored
* ready for PR * cleanup * correct FSMT_PRETRAINED_MODEL_ARCHIVE_LIST * fix * perfectionism * revert change from another PR * odd, already committed this one * non-interactive upload workaround * backup the failed experiment * store langs in config * workaround for localizing model path * doc clean up as in https://github.com/huggingface/transformers/pull/6956 * style * back out debug mode * document: run_eval.py --num_beams 10 * remove unneeded constant * typo * re-use bart's Attention * re-use EncoderLayer, DecoderLayer from bart * refactor * send to cuda and fp16 * cleanup * revert (moved to another PR) * better error message * document run_eval --num_beams * solve the problem of tokenizer finding the right files when model is local * polish, remove hardcoded config * add a note that the file is autogenerated to avoid losing changes * prep for org change, remove unneeded code * switch to model4.pt, update scores * s/python/bash/ * missing init (but doesn't impact the finetuned model) * cleanup * major refactor (reuse-bart) * new model, new expected weights * cleanup * cleanup * full link * fix model type * merge porting notes * style * cleanup * have to create a DecoderConfig object to handle vocab_size properly * doc fix * add note (not a public class) * parametrize * - add bleu scores integration tests * skip test if sacrebleu is not installed * cache heavy models/tokenizers * some tweaks * remove tokens that aren't used * more purging * simplify code * switch to using decoder_start_token_id * add doc * Revert "major refactor (reuse-bart)" This reverts commit 226dad15ca6a9ef4e26178526e878e8fc5c85874. * decouple from bart * remove unused code #1 * remove unused code #2 * remove unused code #3 * update instructions * clean up * move bleu eval to examples * check import only once * move data+gen script into files * reuse via import * take less space * add prepare_seq2seq_batch (auto-tested) * cleanup * recode test to use json instead of yaml * ignore keys not needed * use the new -y in transformers-cli upload -y * [xlm tok] config dict: fix str into int to match definition (#7034) * [s2s] --eval_max_generate_length (#7018) * Fix CI with change of name of nlp (#7054) * nlp -> datasets * More nlp -> datasets * Woopsie * More nlp -> datasets * One last * extending to support allen_nlp wmt models - allow a specific checkpoint file to be passed - more arg settings - scripts for allen_nlp models * sync with changes * s/fsmt-wmt/wmt/ in model names * s/fsmt-wmt/wmt/ in model names (p2) * s/fsmt-wmt/wmt/ in model names (p3) * switch to a better checkpoint * typo * make non-optional args such - adjust tests where possible or skip when there is no other choice * consistency * style * adjust header * cards moved (model rename) * use best custom hparams * update info * remove old cards * cleanup * s/stas/facebook/ * update scores * s/allen_nlp/allenai/ * url maps aren't needed * typo * move all the doc / build /eval generators to their own scripts * cleanup * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * fix indent * duplicated line * style * use the correct add_start_docstrings * oops * resizing can't be done with the core approach, due to 2 dicts * check that the arg is a list * style * style Co-authored-by:
Sam Shleifer <sshleifer@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 16 Sep, 2020 1 commit
-
-
Antoine Louis authored
* Create README.md * Update README.md
-
- 15 Sep, 2020 5 commits
-
-
Patrick von Platen authored
-
Sylvain Gugger authored
-
Pedro Lima authored
From **Language-Agnostic BERT Sentence Embedding** https://ai.googleblog.com/2020/08/language-agnostic-bert-sentence.html
-
tuner007 authored
Model card for PEGASUS finetuned for paraphrasing task
-
Philip May authored
* changed eval table model order * Update install * update mc
-
- 11 Sep, 2020 3 commits
-
-
李明浩 authored
-
李明浩 authored
-
Sagor Sarker authored
* added bangla-bert-base * Apply suggestions from code review Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 10 Sep, 2020 8 commits
-
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
- 09 Sep, 2020 1 commit
-
-
Patrick von Platen authored
-
- 07 Sep, 2020 5 commits
-
-
Mehrdad Farahani authored
ParsBERT v2.0 is a fine-tuned and vocab-reconstructed version of ParsBERT, and it's able to be used in other scopes! It includes these features: - We added some unused-vocab for use in summarization and other scopes. - We fine-tuned the model on vast styles of writing in the Persian language.
-
Abed khooli authored
-
Richard Bownes authored
* Create README.md * Add some custom prompts Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Julien Chaumond authored
-
Julien Chaumond authored
cc @jplu
-
- 06 Sep, 2020 1 commit
-
-
Patrick von Platen authored
-
- 05 Sep, 2020 1 commit
-
-
Steven Liu authored
* create model card for astroGPT * Hotlink to actual image file Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 04 Sep, 2020 1 commit
-
-
Naveenkhasyap authored
* Create Readme.MD for KanBERTo KanBERTo language model readme for Kannada language. * Update model_cards/Naveen-k/KanBERTo/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 03 Sep, 2020 5 commits
-
-
Stefan Engl authored
-
David Mark Nemeskey authored
-
abdullaholuk-loodos authored
Loodos model cards had errors on "Usage" section. It is fixed. Also "electra-base-turkish-uncased" model removed from s3 and re-uploaded as "electra-base-turkish-uncased-discriminator". Its README added. (#6921) Co-authored-by:Abdullah Oluk <abdullaholuk123@gmail.com>
-
Julien Chaumond authored
cc @psorianom @rachelker
-
Antonio V Mendoza authored
Adding the LXMERT pretraining model (MultiModal languageXvision) to HuggingFace's suite of models (#5793) * added template files for LXMERT and competed the configuration_lxmert.py * added modeling, tokization, testing, and finishing touched for lxmert [yet to be tested] * added model card for lxmert * cleaning up lxmert code * Update src/transformers/modeling_lxmert.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/modeling_tf_lxmert.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/modeling_tf_lxmert.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/modeling_lxmert.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * tested torch lxmert, changed documtention, updated outputs, and other small fixes * Update src/transformers/convert_pytorch_checkpoint_to_tf2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/convert_pytorch_checkpoint_to_tf2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/convert_pytorch_checkpoint_to_tf2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * renaming, other small issues, did not change TF code in this commit * added lxmert question answering model in pytorch * added capability to edit number of qa labels for lxmert * made answer optional for lxmert question answering * add option to return hidden_states for lxmert * changed default qa labels for lxmert * changed config archive path * squshing 3 commits: merged UI + testing improvments + more UI and testing * changed some variable names for lxmert * TF LXMERT * Various fixes to LXMERT * Final touches to LXMERT * AutoTokenizer order * Add LXMERT to index.rst and README.md * Merge commit test fixes + Style update * TensorFlow 2.3.0 sequential model changes variable names Remove inherited test * Update src/transformers/modeling_tf_pytorch_utils.py * Update docs/source/model_doc/lxmert.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/model_doc/lxmert.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_lxmert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * added suggestions * Fixes * Final fixes for TF model * Fix docs Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 02 Sep, 2020 1 commit
-
-
David Mark Nemeskey authored
* Create README.md Model card for huBERT. * Update README.md lowercase h * Update model_cards/SZTAKI-HLT/hubert-base-cc/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 01 Sep, 2020 5 commits
-
-
Julien Chaumond authored
-
Rohan Rajpal authored
-
Rohan Rajpal authored
-
Igli Manaj authored
Fix range of possible score, add inference .
-
Tom Grek authored
-