1. 22 Feb, 2023 1 commit
  2. 06 Feb, 2023 1 commit
    • Sylvain Gugger's avatar
      Update quality tooling for formatting (#21480) · 6f79d264
      Sylvain Gugger authored
      * Result of black 23.1
      
      * Update target to Python 3.7
      
      * Switch flake8 to ruff
      
      * Configure isort
      
      * Configure isort
      
      * Apply isort with line limit
      
      * Put the right black version
      
      * adapt black in check copies
      
      * Fix copies
      6f79d264
  3. 05 Dec, 2022 1 commit
    • Kamal Raj Kanakarajan's avatar
      Add BioGPT (#20420) · 13e73668
      Kamal Raj Kanakarajan authored
      * biogpt initial commit
      
      * updated init
      
      * fix faster decoding with use_cache
      
      * 1. fix input_ids and input_embeds with correct device
      2. added _keys_to_ignore_on_load_missing
      3. updated prepare_inputs_for_generation
      
      * add activation_dropout and scale_embedding
      
      * replace fsmt attention with bart attention
      
      * added test
      
      * run make fix-copies
      
      * doc init and fix build
      
      * updated README with proper information
      
      * 1. added tips to docs
      2. updated BioGptTokenizer func
      
      * 1. added tokenizer test
      2. refactor tokenizer
      
      * make fixup
      
      * add biogpt fairseq to hf converter
      
      * updated layer names more
      similar to original checkpoints
      
      * config update doc string and set defaults
      
      * added "#copied" from bart model and
      updated doc strings
      
      * enable model_input_names in tokenizer
      
      * 1.  positionalembedding depending on attention_mask
      2. added attention mask to prepare for generation
      
      * added test to verify past and generation
      
      * BioGptLMHeadModel -> BioGptForCausalLM
      
      * fix typo
      
      * tokenization and test
      Copyright and updated assertion
      
      * updated Copyright and
      one func at time in line
      
      * Copyright updates and
      minor doc fix
      
      * replace assertion with ValueError
      
      * rm extra space
      
      * added code syntax
      
      * revert cmnt position change
      
      * add tokenizer to auto
      
      * updated doc string
      
      * tokenizer doc string update
      
      * biogpt hub model update to microsoft/biogpt
      
      * make fixup
      
      * rm cmnt to fix flake8 5.0.4 vs 6 error
      13e73668
  4. 27 Dec, 2021 1 commit
    • Sylvain Gugger's avatar
      Doc styler v2 (#14950) · 87e6e4fe
      Sylvain Gugger authored
      * New doc styler
      
      * Fix issue with args at the start
      
      * Code sample fixes
      
      * Style code examples in MDX
      
      * Fix more patterns
      
      * Typo
      
      * Typo
      
      * More patterns
      
      * Do without black for now
      
      * Get more info in error
      
      * Docstring style
      
      * Re-enable check
      
      * Quality
      
      * Fix add_end_docstring decorator
      
      * Fix docstring
      87e6e4fe
  5. 21 Dec, 2021 1 commit
    • Sylvain Gugger's avatar
      Mass conversion of documentation from rst to Markdown (#14866) · 27b3031d
      Sylvain Gugger authored
      * Convert docstrings of all configurations and tokenizers
      
      * Processors and fixes
      
      * Last modeling files and fixes to models
      
      * Pipeline modules
      
      * Utils files
      
      * Data submodule
      
      * All the other files
      
      * Style
      
      * Missing examples
      
      * Style again
      
      * Fix copies
      
      * Say bye bye to rst docstrings forever
      27b3031d
  6. 26 Apr, 2021 1 commit
  7. 09 Apr, 2021 1 commit
  8. 31 Mar, 2021 1 commit
  9. 17 Nov, 2020 1 commit
    • Sylvain Gugger's avatar
      Reorganize repo (#8580) · c89bdfbe
      Sylvain Gugger authored
      * Put models in subfolders
      
      * Styling
      
      * Fix imports in tests
      
      * More fixes in test imports
      
      * Sneaky hidden imports
      
      * Fix imports in doc files
      
      * More sneaky imports
      
      * Finish fixing tests
      
      * Fix examples
      
      * Fix path for copies
      
      * More fixes for examples
      
      * Fix dummy files
      
      * More fixes for example
      
      * More model import fixes
      
      * Is this why you're unhappy GitHub?
      
      * Fix imports in conver command
      c89bdfbe
  10. 10 Nov, 2020 1 commit
  11. 29 Oct, 2020 1 commit
  12. 26 Oct, 2020 1 commit
    • Sylvain Gugger's avatar
      Doc styling (#8067) · 08f534d2
      Sylvain Gugger authored
      * Important files
      
      * Styling them all
      
      * Revert "Styling them all"
      
      This reverts commit 7d029395fdae8513b8281cbc2a6c239f8093503e.
      
      * Syling them for realsies
      
      * Fix syntax error
      
      * Fix benchmark_utils
      
      * More fixes
      
      * Fix modeling auto and script
      
      * Remove new line
      
      * Fixes
      
      * More fixes
      
      * Fix more files
      
      * Style
      
      * Add FSMT
      
      * More fixes
      
      * More fixes
      
      * More fixes
      
      * More fixes
      
      * Fixes
      
      * More fixes
      
      * More fixes
      
      * Last fixes
      
      * Make sphinx happy
      08f534d2
  13. 18 Oct, 2020 1 commit
    • Thomas Wolf's avatar
      [Dependencies|tokenizers] Make both SentencePiece and Tokenizers optional dependencies (#7659) · ba8c4d0a
      Thomas Wolf authored
      * splitting fast and slow tokenizers [WIP]
      
      * [WIP] splitting sentencepiece and tokenizers dependencies
      
      * update dummy objects
      
      * add name_or_path to models and tokenizers
      
      * prefix added to file names
      
      * prefix
      
      * styling + quality
      
      * spliting all the tokenizer files - sorting sentencepiece based ones
      
      * update tokenizer version up to 0.9.0
      
      * remove hard dependency on sentencepiece 馃帀
      
      * and removed hard dependency on tokenizers 馃帀
      
      
      
      * update conversion script
      
      * update missing models
      
      * fixing tests
      
      * move test_tokenization_fast to main tokenization tests - fix bugs
      
      * bump up tokenizers
      
      * fix bert_generation
      
      * update ad fix several tokenizers
      
      * keep sentencepiece in deps for now
      
      * fix funnel and deberta tests
      
      * fix fsmt
      
      * fix marian tests
      
      * fix layoutlm
      
      * fix squeezebert and gpt2
      
      * fix T5 tokenization
      
      * fix xlnet tests
      
      * style
      
      * fix mbart
      
      * bump up tokenizers to 0.9.2
      
      * fix model tests
      
      * fix tf models
      
      * fix seq2seq examples
      
      * fix tests without sentencepiece
      
      * fix slow => fast  conversion without sentencepiece
      
      * update auto and bert generation tests
      
      * fix mbart tests
      
      * fix auto and common test without tokenizers
      
      * fix tests without tokenizers
      
      * clean up tests lighten up when tokenizers + sentencepiece are both off
      
      * style quality and tests fixing
      
      * add sentencepiece to doc/examples reqs
      
      * leave sentencepiece on for now
      
      * style quality split hebert and fix pegasus
      
      * WIP Herbert fast
      
      * add sample_text_no_unicode and fix hebert tokenization
      
      * skip FSMT example test for now
      
      * fix style
      
      * fix fsmt in example tests
      
      * update following Lysandre and Sylvain's comments
      
      * Update src/transformers/testing_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/testing_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      ba8c4d0a
  14. 08 Oct, 2020 1 commit
    • Thomas Wolf's avatar
      Adding Fast tokenizers for SentencePiece based tokenizers - Breaking: remove... · 9aeacb58
      Thomas Wolf authored
      
      Adding Fast tokenizers for SentencePiece based tokenizers - Breaking: remove Transfo-XL fast tokenizer (#7141)
      
      * [WIP] SP tokenizers
      
      * fixing tests for T5
      
      * WIP tokenizers
      
      * serialization
      
      * update T5
      
      * WIP T5 tokenization
      
      * slow to fast conversion script
      
      * Refactoring to move tokenzier implementations inside transformers
      
      * Adding gpt - refactoring - quality
      
      * WIP adding several tokenizers to the fast world
      
      * WIP Roberta - moving implementations
      
      * update to dev4 switch file loading to in-memory loading
      
      * Updating and fixing
      
      * advancing on the tokenizers - updating do_lower_case
      
      * style and quality
      
      * moving forward with tokenizers conversion and tests
      
      * MBart, T5
      
      * dumping the fast version of transformer XL
      
      * Adding to autotokenizers + style/quality
      
      * update init and space_between_special_tokens
      
      * style and quality
      
      * bump up tokenizers version
      
      * add protobuf
      
      * fix pickle Bert JP with Mecab
      
      * fix newly added tokenizers
      
      * style and quality
      
      * fix bert japanese
      
      * fix funnel
      
      * limite tokenizer warning to one occurence
      
      * clean up file
      
      * fix new tokenizers
      
      * fast tokenizers deep tests
      
      * WIP adding all the special fast tests on the new fast tokenizers
      
      * quick fix
      
      * adding more fast tokenizers in the fast tests
      
      * all tokenizers in fast version tested
      
      * Adding BertGenerationFast
      
      * bump up setup.py for CI
      
      * remove BertGenerationFast (too early)
      
      * bump up tokenizers version
      
      * Clean old docstrings
      
      * Typo
      
      * Update following Lysandre comments
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      9aeacb58
  15. 23 Sep, 2020 1 commit
  16. 18 Sep, 2020 1 commit
    • Dat Quoc Nguyen's avatar
      Add new pre-trained models BERTweet and PhoBERT (#6129) · af2322c7
      Dat Quoc Nguyen authored
      * Add BERTweet and PhoBERT models
      
      * Update modeling_auto.py
      
      Re-add `bart` to LM_MAPPING
      
      * Update tokenization_auto.py
      
      Re-add `from .configuration_mobilebert import MobileBertConfig`
      not sure why it's replaced by `from transformers.configuration_mobilebert import MobileBertConfig`
      
      * Add BERTweet and PhoBERT to pretrained_models.rst
      
      * Update tokenization_auto.py
      
      Remove BertweetTokenizer and PhobertTokenizer out of tokenization_auto.py (they are currently not supported by AutoTokenizer.
      
      * Update BertweetTokenizer - without nltk
      
      * Update model card for BERTweet
      
      * PhoBERT - with Auto mode - without import fastBPE
      
      * PhoBERT - with Auto mode - without import fastBPE
      
      * BERTweet - with Auto mode - without import fastBPE
      
      * Add PhoBERT and BERTweet to TF modeling auto
      
      * Improve Docstrings for PhobertTokenizer and BertweetTokenizer
      
      * Update PhoBERT and BERTweet model cards
      
      * Fixed a merge conflict in tokenization_auto
      
      * Used black to reformat BERTweet- and PhoBERT-related files
      
      * Used isort to reformat BERTweet- and PhoBERT-related files
      
      * Reformatted BERTweet- and PhoBERT-related files based on flake8
      
      * Updated test files
      
      * Updated test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Updated tf test files
      
      * Update commits from huggingface
      
      * Delete unnecessary files
      
      * Add tokenizers to auto and init files
      
      * Add test files for tokenizers
      
      * Revised model cards
      
      * Update save_vocabulary function in BertweetTokenizer and PhobertTokenizer and test files
      
      * Revised test files
      
      * Update orders of Phobert and Bertweet tokenizers in auto tokenization file
      af2322c7
  17. 04 Sep, 2020 1 commit
  18. 26 Aug, 2020 1 commit
  19. 26 Jun, 2020 1 commit
  20. 11 May, 2020 1 commit
  21. 21 Apr, 2020 1 commit
  22. 18 Apr, 2020 1 commit
    • Thomas Wolf's avatar
      Cleanup fast tokenizers integration (#3706) · 827d6d6e
      Thomas Wolf authored
      
      
      * First pass on utility classes and python tokenizers
      
      * finishing cleanup pass
      
      * style and quality
      
      * Fix tests
      
      * Updating following @mfuntowicz comment
      
      * style and quality
      
      * Fix Roberta
      
      * fix batch_size/seq_length inBatchEncoding
      
      * add alignement methods + tests
      
      * Fix OpenAI and Transfo-XL tokenizers
      
      * adding trim_offsets=True default for GPT2 et RoBERTa
      
      * style and quality
      
      * fix tests
      
      * add_prefix_space in roberta
      
      * bump up tokenizers to rc7
      
      * style
      
      * unfortunately tensorfow does like these - removing shape/seq_len for now
      
      * Update src/transformers/tokenization_utils.py
      Co-Authored-By: default avatarStefan Schweter <stefan@schweter.it>
      
      * Adding doc and docstrings
      
      * making flake8 happy
      Co-authored-by: default avatarStefan Schweter <stefan@schweter.it>
      827d6d6e
  23. 18 Mar, 2020 1 commit
  24. 25 Feb, 2020 1 commit
    • Lysandre Debut's avatar
      Documentation (#2989) · bb7c4685
      Lysandre Debut authored
      * All Tokenizers
      
      BertTokenizer + few fixes
      RobertaTokenizer
      OpenAIGPTTokenizer + Fixes
      GPT2Tokenizer + fixes
      TransfoXLTokenizer
      Correct rst for TransformerXL
      XLMTokenizer + fixes
      XLNet Tokenizer + Style
      DistilBERT + Fix XLNet RST
      CTRLTokenizer
      CamemBERT Tokenizer
      FlaubertTokenizer
      XLMRobertaTokenizer
      cleanup
      
      * cleanup
      bb7c4685
  25. 20 Feb, 2020 1 commit
  26. 27 Jan, 2020 1 commit
  27. 21 Jan, 2020 1 commit
  28. 15 Jan, 2020 1 commit
  29. 06 Jan, 2020 2 commits
  30. 05 Jan, 2020 1 commit
  31. 22 Dec, 2019 4 commits
  32. 21 Dec, 2019 1 commit
    • Aymeric Augustin's avatar
      Reformat source code with black. · fa84ae26
      Aymeric Augustin authored
      This is the result of:
      
          $ black --line-length 119 examples templates transformers utils hubconf.py setup.py
      
      There's a lot of fairly long lines in the project. As a consequence, I'm
      picking the longest widely accepted line length, 119 characters.
      
      This is also Thomas' preference, because it allows for explicit variable
      names, to make the code easier to understand.
      fa84ae26
  33. 20 Dec, 2019 2 commits
  34. 19 Dec, 2019 1 commit
  35. 18 Dec, 2019 1 commit