"vscode:/vscode.git/clone" did not exist on "ef7588b617df3b861b687ab6aefc95fb4e0c5e1e"
- 01 Jun, 2020 1 commit
-
-
Victor SANH authored
-
- 29 May, 2020 2 commits
-
-
Simon Böhm authored
Change the example code to use encode_plus since the token_type_id wasn't being correctly set.
-
Patrick von Platen authored
* add multiple choice for longformer * add models to docs * adapt docstring * add test to longformer * add longformer for mc in init and modeling auto * fix tests
-
- 19 May, 2020 1 commit
-
-
Julien Chaumond authored
* Test case for #3936 * multigpu tests pass on pytorch 1.4.0 * Fixup * multigpu tests pass on pytorch 1.5.0 * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py * rename multigpu to require_multigpu * mode doc
-
- 30 Apr, 2020 1 commit
-
-
Jordan authored
-
- 29 Apr, 2020 1 commit
-
-
Julien Chaumond authored
* [file_utils] use_cdn + documentation * Move to cdn. urls for weights * [urls] Hotfix for bert-base-japanese
-
- 28 Apr, 2020 1 commit
-
-
Patrick von Platen authored
* change encoder decoder style to bart & t5 style * make encoder decoder generation dummy work for bert * make style * clean init config in encoder decoder * add tests for encoder decoder models * refactor and add last tests * refactor and add last tests * fix attn masks for bert encoder decoder * make style * refactor prepare inputs for Bert * refactor * finish encoder decoder * correct typo * add docstring to config * finish * add tests * better naming * make style * fix flake8 * clean docstring * make style * rename
-
- 23 Apr, 2020 1 commit
-
-
Julien Chaumond authored
-
- 21 Apr, 2020 1 commit
-
-
Bharat Raghunathan authored
-
- 17 Apr, 2020 1 commit
-
-
Simon Böhm authored
token_type_id is converted into the segment embedding. For question answering, this needs to highlight whether a token belongs to sequence 0 or 1. encode_plus takes care of correctly setting this parameter automatically.
-
- 16 Apr, 2020 2 commits
-
-
Patrick von Platen authored
-
Sam Shleifer authored
* Delete some copy pasted code
-
- 03 Apr, 2020 1 commit
-
-
Lysandre Debut authored
* Electra wip * helpers * Electra wip * Electra v1 * ELECTRA may be saved/loaded * Generator & Discriminator * Embedding size instead of halving the hidden size * ELECTRA Tokenizer * Revert BERT helpers * ELECTRA Conversion script * Archive maps * PyTorch tests * Start fixing tests * Tests pass * Same configuration for both models * Compatible with base + large * Simplification + weight tying * Archives * Auto + Renaming to standard names * ELECTRA is uncased * Tests * Slight API changes * Update tests * wip * ElectraForTokenClassification * temp * Simpler arch + tests Removed ElectraForPreTraining which will be in a script * Conversion script * Auto model * Update links to S3 * Split ElectraForPreTraining and ElectraForTokenClassification * Actually test PreTraining model * Remove num_labels from configuration * wip * wip * From discriminator and generator to electra * Slight API changes * Better naming * TensorFlow ELECTRA tests * Accurate conversion script * Added to conversion script * Fast ELECTRA tokenizer * Style * Add ELECTRA to README * Modeling Pytorch Doc + Real style * TF Docs * Docs * Correct links * Correct model intialized * random fixes * style * Addressing Patrick's and Sam's comments * Correct links in docs
-
- 01 Apr, 2020 1 commit
-
-
Anirudh Srinivasan authored
-
- 25 Feb, 2020 2 commits
-
-
Lysandre Debut authored
* All Tokenizers BertTokenizer + few fixes RobertaTokenizer OpenAIGPTTokenizer + Fixes GPT2Tokenizer + fixes TransfoXLTokenizer Correct rst for TransformerXL XLMTokenizer + fixes XLNet Tokenizer + Style DistilBERT + Fix XLNet RST CTRLTokenizer CamemBERT Tokenizer FlaubertTokenizer XLMRobertaTokenizer cleanup * cleanup
-
srush authored
* change masking to direct labelings * fix black * switch to ignore index * . * fix black
-
- 21 Feb, 2020 1 commit
-
-
Lysandre Debut authored
-
- 13 Feb, 2020 1 commit
-
-
Sam Shleifer authored
* activations.py contains a mapping from string to activation function * resolves some `gelu` vs `gelu_new` ambiguity
-
- 11 Feb, 2020 1 commit
-
-
Oleksiy Syvokon authored
PyTorch < 1.3 requires multiplication operands to be of the same type. This was violated when using default attention mask (i.e., attention_mask=None in arguments) given BERT in the decoder mode. In particular, this was breaking Model2Model and made tutorial from the quickstart failing.
-
- 07 Feb, 2020 1 commit
-
-
monologg authored
-
- 04 Feb, 2020 1 commit
-
-
Lysandre authored
-
- 03 Feb, 2020 1 commit
-
-
Lysandre authored
Masked indices should have -1 and not -100. Updating documentation + scripts that were forgotten
-
- 28 Jan, 2020 1 commit
-
-
Wietse de Vries authored
-
- 23 Jan, 2020 7 commits
- 15 Jan, 2020 1 commit
-
-
Julien Chaumond authored
-
- 14 Jan, 2020 1 commit
-
-
Lysandre authored
Created a link between the linear layer bias and the model attribute bias. This does not change anything for the user nor for the conversion scripts, but allows the `resize_token_embeddings` method to resize the bias as well as the weights of the decoder. Added a test.
-
- 07 Jan, 2020 2 commits
-
-
Romain Keramitas authored
Signed-off-by:Romain Keramitas <r.keramitas@gmail.com>
-
Genta Indra Winata authored
-
- 06 Jan, 2020 3 commits
-
-
alberduris authored
-
alberduris authored
-
Lysandre authored
-
- 22 Dec, 2019 3 commits
-
-
Aymeric Augustin authored
-
Aymeric Augustin authored
-
Aymeric Augustin authored
This prevents transformers from being importable simply because the CWD is the root of the git repository, while not being importable from other directories. That led to inconsistent behavior, especially in examples. Once you fetch this commit, in your dev environment, you must run: $ pip uninstall transformers $ pip install -e .
-