"torchvision/csrc/vscode:/vscode.git/clone" did not exist on "d9e6d60f49e650c2f70b2cf457b2aca081239930"
  1. 07 Aug, 2020 1 commit
  2. 05 Aug, 2020 1 commit
    • Sylvain Gugger's avatar
      Tf model outputs (#6247) · c67d1a02
      Sylvain Gugger authored
      * TF outputs and test on BERT
      
      * Albert to DistilBert
      
      * All remaining TF models except T5
      
      * Documentation
      
      * One file forgotten
      
      * TF outputs and test on BERT
      
      * Albert to DistilBert
      
      * All remaining TF models except T5
      
      * Documentation
      
      * One file forgotten
      
      * Add new models and fix issues
      
      * Quality improvements
      
      * Add T5
      
      * A bit of cleanup
      
      * Fix for slow tests
      
      * Style
      c67d1a02
  3. 03 Aug, 2020 1 commit
  4. 01 Aug, 2020 1 commit
  5. 30 Jul, 2020 1 commit
    • Oren Amsalem's avatar
      Actually the extra_id are from 0-99 and not from 1-100 (#5967) · d24ea708
      Oren Amsalem authored
      a = tokenizer.encode("we got a <extra_id_99>", return_tensors='pt',add_special_tokens=True)
      print(a)
      >tensor([[   62,   530,     3,     9, 32000]])
      a = tokenizer.encode("we got a <extra_id_100>", return_tensors='pt',add_special_tokens=True)
      print(a)
      >tensor([[   62,   530,     3,     9,     3,     2, 25666,   834,    23,    26,
                 834,  2915,  3155]])
      d24ea708
  6. 20 Jul, 2020 1 commit
  7. 13 Jul, 2020 1 commit
  8. 10 Jul, 2020 1 commit
  9. 07 Jul, 2020 2 commits
    • Sam Shleifer's avatar
      Add mbart-large-cc25, support translation finetuning (#5129) · 353b8f1e
      Sam Shleifer authored
      improve unittests for finetuning, especially w.r.t testing frozen parameters
      fix freeze_embeds for T5
      add streamlit setup.cfg
      353b8f1e
    • Quentin Lhoest's avatar
      Add DPR model (#5279) · fbd87921
      Quentin Lhoest authored
      
      
      * beginning of dpr modeling
      
      * wip
      
      * implement forward
      
      * remove biencoder + better init weights
      
      * export dpr model to embed model for nlp lib
      
      * add new api
      
      * remove old code
      
      * make style
      
      * fix dumb typo
      
      * don't load bert weights
      
      * docs
      
      * docs
      
      * style
      
      * move the `k` parameter
      
      * fix init_weights
      
      * add pretrained configs
      
      * minor
      
      * update config names
      
      * style
      
      * better config
      
      * style
      
      * clean code based on PR comments
      
      * change Dpr to DPR
      
      * fix config
      
      * switch encoder config to a dict
      
      * style
      
      * inheritance -> composition
      
      * add messages in assert startements
      
      * add dpr reader tokenizer
      
      * one tokenizer per model
      
      * fix base_model_prefix
      
      * fix imports
      
      * typo
      
      * add convert script
      
      * docs
      
      * change tokenizers conf names
      
      * style
      
      * change tokenizers conf names
      
      * minor
      
      * minor
      
      * fix wrong names
      
      * minor
      
      * remove unused convert functions
      
      * rename convert script
      
      * use return_tensors in tokenizers
      
      * remove n_questions dim
      
      * move generate logic to tokenizer
      
      * style
      
      * add docs
      
      * docs
      
      * quality
      
      * docs
      
      * add tests
      
      * style
      
      * add tokenization tests
      
      * DPR full tests
      
      * Stay true to the attention mask building
      
      * update docs
      
      * missing param in bert input docs
      
      * docs
      
      * style
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      fbd87921
  10. 01 Jul, 2020 2 commits
  11. 22 Jun, 2020 1 commit
  12. 19 Jun, 2020 1 commit
    • Vasily Shamporov's avatar
      Add MobileBert (#4901) · 9a3f9108
      Vasily Shamporov authored
      
      
      * Add MobileBert
      
      * Quality + Conversion script
      
      * style
      
      * Update src/transformers/modeling_mobilebert.py
      
      * Links to S3
      
      * Style
      
      * TFMobileBert
      
      Slight fixes to the pytorch MobileBert
      Style
      
      * MobileBertForMaskedLM (PT + TF)
      
      * MobileBertForNextSentencePrediction (PT + TF)
      
      * MobileFor{MultipleChoice, TokenClassification} (PT + TF)
      
      
      ss
      
      * Tests + Auto
      
      * Doc
      
      * Tests
      
      * Addressing @sgugger's comments
      
      * Adressing @patrickvonplaten's comments
      
      * Style
      
      * Style
      
      * Integration test
      
      * style
      
      * Model card
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      9a3f9108
  13. 18 Jun, 2020 1 commit
  14. 17 Jun, 2020 1 commit
  15. 16 Jun, 2020 2 commits
    • Sylvain Gugger's avatar
      Fix all sphynx warnings (#5068) · 011cc0be
      Sylvain Gugger authored
      011cc0be
    • Yacine Jernite's avatar
      Eli5 examples (#4968) · 49c52025
      Yacine Jernite authored
      
      
      * add eli5 examples
      
      * add dense query script
      
      * query_di
      
      * merging
      
      * merging
      
      * add_utils
      
      * adds nearest neighbor wikipedia
      
      * batch queries
      
      * training_retriever
      
      * new notebooks
      
      * moved retriever traiing script
      
      * finished wiki40b
      
      * max_len_fix
      
      * train_s2s
      
      * retriever_batch_checkpointing
      
      * cleanup
      
      * merge
      
      * dim_fix
      
      * fix_indexer
      
      * fix_wiki40b_snippets
      
      * fix_embed_for_r
      
      * fp32 index
      
      * fix_sparse_q
      
      * joint_training
      
      * remove obsolete datasets
      
      * add_passage_nn_results
      
      * add_passage_nn_results
      
      * add_batch_nn
      
      * add_batch_nn
      
      * add_data_scripts
      
      * notebook
      
      * notebook
      
      * notebook
      
      * fix_multi_gpu
      
      * add_app
      
      * full_caching
      
      * full_caching
      
      * notebook
      
      * sparse_done
      
      * images
      
      * notebook
      
      * add_image_gif
      
      * with_Gif
      
      * add_contr_image
      
      * notebook
      
      * notebook
      
      * notebook
      
      * train_functions
      
      * notebook
      
      * min_retrieval_length
      
      * pandas_option
      
      * notebook
      
      * min_retrieval_length
      
      * notebook
      
      * notebook
      
      * eval_Retriever
      
      * notebook
      
      * images
      
      * notebook
      
      * add_example
      
      * add_example
      
      * notebook
      
      * fireworks
      
      * notebook
      
      * notebook
      
      * joe's notebook comments
      
      * app_update
      
      * notebook
      
      * notebook_link
      
      * captions
      
      * notebook
      
      * assing RetriBert model
      
      * add RetriBert to Auto
      
      * change AutoLMHead to AutoSeq2Seq
      
      * notebook downloads from hf models
      
      * style_black
      
      * style_black
      
      * app_update
      
      * app_update
      
      * fix_app_update
      
      * style
      
      * style
      
      * isort
      
      * Delete WikiELI5training.ipynb
      
      * Delete evaluate_eli5.py
      
      * Delete WikiELI5explore.ipynb
      
      * Delete ExploreWikiELI5Support.html
      
      * Delete explainlikeimfive.py
      
      * Delete wiki_snippets.py
      
      * children before parent
      
      * children before parent
      
      * style_black
      
      * style_black_only
      
      * isort
      
      * isort_new
      
      * Update src/transformers/modeling_retribert.py
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      
      * typo fixes
      
      * app_without_asset
      
      * cleanup
      
      * Delete ELI5animation.gif
      
      * Delete ELI5contrastive.svg
      
      * Delete ELI5wiki_index.svg
      
      * Delete choco_bis.svg
      
      * Delete fireworks.gif
      
      * Delete huggingface_logo.jpg
      
      * Delete huggingface_logo.svg
      
      * Delete Long_Form_Question_Answering_with_ELI5_and_Wikipedia.ipynb
      
      * Delete eli5_app.py
      
      * Delete eli5_utils.py
      
      * readme
      
      * Update README.md
      
      * unused imports
      
      * moved_info
      
      * default_beam
      
      * ftuned model
      
      * disclaimer
      
      * Update src/transformers/modeling_retribert.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * black
      
      * add_doc
      
      * names
      
      * isort_Examples
      
      * isort_Examples
      
      * Add doc to index
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      49c52025
  16. 15 Jun, 2020 1 commit
  17. 12 Jun, 2020 2 commits
  18. 10 Jun, 2020 1 commit
    • Suraj Patil's avatar
      ElectraForQuestionAnswering (#4913) · ef2dcdcc
      Suraj Patil authored
      * ElectraForQuestionAnswering
      
      * udate __init__
      
      * add test for electra qa model
      
      * add ElectraForQuestionAnswering in auto models
      
      * add ElectraForQuestionAnswering in all_model_classes
      
      * fix outputs, input_ids defaults to None
      
      * add ElectraForQuestionAnswering in docs
      
      * remove commented line
      ef2dcdcc
  19. 09 Jun, 2020 1 commit
  20. 08 Jun, 2020 1 commit
  21. 02 Jun, 2020 1 commit
  22. 29 May, 2020 2 commits
  23. 26 May, 2020 1 commit
  24. 25 May, 2020 1 commit
  25. 22 May, 2020 1 commit
  26. 19 May, 2020 1 commit
  27. 18 May, 2020 1 commit
  28. 13 May, 2020 1 commit
  29. 10 May, 2020 1 commit
  30. 07 May, 2020 2 commits
    • Julien Chaumond's avatar
      c99fe038
    • Patrick von Platen's avatar
      Reformer (#3351) · dca34695
      Patrick von Platen authored
      * first copy & past commit from Bert and morgans LSH code
      
      * add easy way to compare to trax original code
      
      * translate most of function
      
      * make trax lsh self attention deterministic with numpy seed + copy paste code
      
      * add same config
      
      * add same config
      
      * make layer init work
      
      * implemented hash_vectors function for lsh attention
      
      * continue reformer translation
      
      * hf LSHSelfAttentionLayer gives same output as trax layer
      
      * refactor code
      
      * refactor code
      
      * refactor code
      
      * refactor
      
      * refactor + add reformer config
      
      * delete bogus file
      
      * split reformer attention layer into two layers
      
      * save intermediate step
      
      * save intermediate step
      
      * make test work
      
      * add complete reformer block layer
      
      * finish reformer layer
      
      * implement causal and self mask
      
      * clean reformer test and refactor code
      
      * fix merge conflicts
      
      * fix merge conflicts
      
      * update init
      
      * fix device for GPU
      
      * fix chunk length init for tests
      
      * include morgans optimization
      
      * improve memory a bit
      
      * improve comment
      
      * factorize num_buckets
      
      * better testing parameters
      
      * make whole model work
      
      * make lm model work
      
      * add t5 copy paste tokenizer
      
      * add chunking feed forward
      
      * clean config
      
      * add improved assert statements
      
      * make tokenizer work
      
      * improve test
      
      * correct typo
      
      * extend config
      
      * add complexer test
      
      * add new axial position embeddings
      
      * add local block attention layer
      
      * clean tests
      
      * refactor
      
      * better testing
      
      * save intermediate progress
      
      * clean test file
      
      * make shorter input length work for model
      
      * allow variable input length
      
      * refactor
      
      * make forward pass for pretrained model work
      
      * add generation possibility
      
      * finish dropout and init
      
      * make style
      
      * refactor
      
      * add first version of RevNet Layers
      
      * make forward pass work and add convert file
      
      * make uploaded model forward pass work
      
      * make uploaded model forward pass work
      
      * refactor code
      
      * add namedtuples and cache buckets
      
      * correct head masks
      
      * refactor
      
      * made reformer more flexible
      
      * make style
      
      * remove set max length
      
      * add attention masks
      
      * fix up tests
      
      * fix lsh attention mask
      
      * make random seed optional for the moment
      
      * improve memory in reformer
      
      * add tests
      
      * make style
      
      * make sure masks work correctly
      
      * detach gradients
      
      * save intermediate
      
      * correct backprob through gather
      
      * make style
      
      * change back num hashes
      
      * rename to labels
      
      * fix rotation shape
      
      * fix detach
      
      * update
      
      * fix trainer
      
      * fix backward dropout
      
      * make reformer more flexible
      
      * fix conflict
      
      * fix
      
      * fix
      
      * add tests for fixed seed in reformer layer
      
      * fix trainer typo
      
      * fix typo in activations
      
      * add fp16 tests
      
      * add fp16 training
      
      * support fp16
      
      * correct gradient bug in reformer
      
      * add fast gelu
      
      * re-add dropout for embedding dropout
      
      * better naming
      
      * better naming
      
      * renaming
      
      * finalize test branch
      
      * finalize tests
      
      * add more tests
      
      * finish tests
      
      * fix
      
      * fix type trainer
      
      * fix fp16 tests
      
      * fix tests
      
      * fix tests
      
      * fix tests
      
      * fix issue with dropout
      
      * fix dropout seeds
      
      * correct random seed on gpu
      
      * finalize random seed for dropout
      
      * finalize random seed for dropout
      
      * remove duplicate line
      
      * correct half precision bug
      
      * make style
      
      * refactor
      
      * refactor
      
      * docstring
      
      * remove sinusoidal position encodings for reformer
      
      * move chunking to modeling_utils
      
      * make style
      
      * clean config
      
      * make style
      
      * fix tests
      
      * fix auto tests
      
      * pretrained models
      
      * fix docstring
      
      * update conversion file
      
      * Update pretrained_models.rst
      
      * fix rst
      
      * fix rst
      
      * update copyright
      
      * fix test path
      
      * fix test path
      
      * fix small issue in test
      
      * include reformer in generation tests
      
      * add docs for axial position encoding
      
      * finish docs
      
      * Update convert_reformer_trax_checkpoint_to_pytorch.py
      
      * remove isort
      
      * include sams comments
      
      * remove wrong comment in utils
      
      * correct typos
      
      * fix typo
      
      * Update reformer.rst
      
      * applied morgans optimization
      
      * make style
      
      * make gpu compatible
      
      * remove bogus file
      
      * big test refactor
      
      * add example for chunking
      
      * fix typo
      
      * add to README
      dca34695
  31. 28 Apr, 2020 2 commits
  32. 27 Apr, 2020 1 commit
  33. 18 Apr, 2020 1 commit
    • Thomas Wolf's avatar
      Cleanup fast tokenizers integration (#3706) · 827d6d6e
      Thomas Wolf authored
      
      
      * First pass on utility classes and python tokenizers
      
      * finishing cleanup pass
      
      * style and quality
      
      * Fix tests
      
      * Updating following @mfuntowicz comment
      
      * style and quality
      
      * Fix Roberta
      
      * fix batch_size/seq_length inBatchEncoding
      
      * add alignement methods + tests
      
      * Fix OpenAI and Transfo-XL tokenizers
      
      * adding trim_offsets=True default for GPT2 et RoBERTa
      
      * style and quality
      
      * fix tests
      
      * add_prefix_space in roberta
      
      * bump up tokenizers to rc7
      
      * style
      
      * unfortunately tensorfow does like these - removing shape/seq_len for now
      
      * Update src/transformers/tokenization_utils.py
      Co-Authored-By: default avatarStefan Schweter <stefan@schweter.it>
      
      * Adding doc and docstrings
      
      * making flake8 happy
      Co-authored-by: default avatarStefan Schweter <stefan@schweter.it>
      827d6d6e