- 18 Jan, 2022 3 commits
-
-
Li-Huai (Allan) Lin authored
* REALM initial commit * Retriever OK (Update new_gelu). * Encoder prediction score OK * Encoder pretrained model OK * Update retriever comments * Update docs, tests, and imports * Prune unused models * Make embedder as a module `RealmEmbedder` * Add RealmRetrieverOutput * Update tokenization * Pass all tests in test_modeling_realm.py * Prune RealmModel * Update docs * Add training test. * Remove completed TODO * Style & Quality * Prune `RealmModel` * Fixup * Changes: 1. Remove RealmTokenizerFast 2. Update docstrings 3. Add a method to RealmTokenizer to handle candidates tokenization. * Fix up * Style * Add tokenization tests * Update `from_pretrained` tests * Apply suggestions * Style & Quality * Copy BERT model * Fix comment to avoid docstring copying * Make RealmBertModel private * Fix bug * Style * Basic QA * Save * Complete reader logits * Add searcher * Complete searcher & reader * Move block records init to constructor * Fix training bug * Add some outputs to RealmReader * Add finetuned checkpoint variable names parsing * Fix bug * Update REALM config * Add RealmForOpenQA * Update convert_tfrecord logits * Fix bugs * Complete imports * Update docs * Update naming * Add brute-force searcher * Pass realm model tests * Style * Exclude RealmReader from common tests * Fix * Fix * convert docs * up * up * more make style * up * upload * up * Fix * Update src/transformers/__init__.py * adapt testing * change modeling code * fix test * up * up * up * correct more * make retriever work * update * make style * finish main structure * Resolve merge conflict * Make everything work * Style * Fixup * Fixup * Update training test * fix retriever * remove hardcoded path * Fix * Fix modeling test * Update model links * Initial retrieval test * Fix modeling test * Complete retrieval tests * Fix * style * Fix tests * Fix docstring example * Minor fix of retrieval test * Update license headers and docs * Apply suggestions from code review * Style * Apply suggestions from code review * Add an example to RealmEmbedder * Fix Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
* fix doc example - NameError: name 'PATH' is not defined * fix name 'TFRagModel' is not defined * correct TFRagRagSequenceForGeneration * fix name 'tf' is not defined * fix style Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Nicolas Patry authored
* `is_ctc` needs to be updated to `self.type == "ctc". * Adding fast test for this functionality.
-
- 17 Jan, 2022 5 commits
-
-
Yih-Dar authored
* fix doc example - cannot import name 'UniSpeechSatFeatureEncoder' * fix ckpt name Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Stas Bekman authored
add new paper
-
Matt authored
-
MrinalTyagi authored
-
- 16 Jan, 2022 1 commit
-
-
Patrick von Platen authored
-
- 15 Jan, 2022 1 commit
-
-
Stas Bekman authored
* Efficient Software Prebuilds * improve
-
- 14 Jan, 2022 11 commits
-
-
Joao Gante authored
-
Sylvain Gugger authored
* Better dummies * See if this fixes the issue * Fix quality * Style * Add doc for DummyObject
-
Nicolas Patry authored
* Fixing flaky test (hopefully). * tf compliant.
-
Joao Gante authored
* TF Bert inference - support np.ndarray optional arguments * apply np input tests to all TF architectures
-
AK391 authored
* update XLMProphetNet link * update DPR link * change prophetnet link * change link MBART * change link GPT * update gpt2 link * ctrl update link * update Transformer-XL link * Update Reformer link * update xlnet link * bert update link * udpate albert link * roberta update link * update distilbert link * update convbert link * update XLM link * xlm roberta update link * update Flaubert link * update electra link * update funnel transformer and longformer * bart update link * pegasus update link * udpate marianmt link * t5 update link * mt5 update link
-
novice authored
-
SaulLu authored
* add new test * fix in init * more relevant test
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* Make sure all submodules are properly registered * Try to fix tests * Fix tests
-
Joao Gante authored
-
Sylvain Gugger authored
* Check the repo consistency in model templates test * Fix doc template * Fix docstrings * Fix last docstring
-
- 13 Jan, 2022 8 commits
-
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Yanming Wang authored
* Multiple fixes of trainer class with XLA GPU * Make fp16 valid for xla:gpu * Add mark_step in should_log to reduce compilation overhead
-
Carlos Aguayo authored
Fix typo
-
Manuel R. Ciosici authored
* Add AdamW deprecation warning * Add --optim to Trainer * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py * fix style * fix * Regroup adamws together Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Change --adafactor to --optim adafactor * Use Enum for optimizer values * fixup! Change --adafactor to --optim adafactor * fixup! Change --adafactor to --optim adafactor * fixup! Change --adafactor to --optim adafactor * fixup! Use Enum for optimizer values * Improved documentation for --adafactor Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Add mention of no_deprecation_warning Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename OptimizerOptions to OptimizerNames * Use choices for --optim * Move optimizer selection code to a function and add a unit test * Change optimizer names * Rename method Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename method Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Remove TODO comment Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename variable Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename variable Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename function * Rename variable * Parameterize the tests for supported optimizers * Refactor * Attempt to make tests pass on CircleCI * Add a test with apex * rework to add apex to parameterized; add actual train test * fix import when torch is not available * fix optim_test_params when torch is not available * fix optim_test_params when torch is not available * re-org * small re-org * fix test_fused_adam_no_apex * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Remove .value from OptimizerNames * Rename optimizer strings s|--adam_|--adamw_| * Also rename Enum options * small fix * Fix instantiation of OptimizerNames. Remove redundant test * Use ExplicitEnum instead of Enum * Add unit test with string optimizer * Change optimizer default to string value Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas@stason.org>
-
Stas Bekman authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Lysandre Debut authored
* Updated script * Commit everything * Ready for review! * Update .github/workflows/build_documentation.yml Co-authored-by:
Julien Chaumond <julien@huggingface.co> Co-authored-by:
Julien Chaumond <julien@huggingface.co>
-
- 12 Jan, 2022 8 commits
-
-
Edoardo Federici authored
* Update run_summarization.py * Fixed languages and added missing code * fixed obj, docs, removed source_lang and target_lang * make style, run_summarization.py reformatted
-
Jake Tae authored
* refactor: wrap forward pass around no_grad context * Update tests/test_modeling_distilbert.py * fix: rm `no_grad` from non-integration tests * chore: rm whitespace change
-
lewtun authored
* Add ONNX classes to main package * Remove permalinks from ONNX guide * Fix ToC entry * Revert "Add ONNX classes to main package" This reverts commit eb794a5b00d66b0b4eab234987301676d8357630. * Add ONNX classes to main doc * Fix syntax highlighting in doc * Fix text * Add FeaturesManager to doc * Use paths to reference ONNX classes * Add FeaturesManager to init * Add missing ONNX paths
-
Sylvain Gugger authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Leandro von Werra authored
-
Russell Klopfer authored
* use block_size instead of max_seq_length * fixup * remove pad_to_block_size Co-authored-by:Russell Klopfer <russell@kloper.us>
-
Nicolas Patry authored
* Pipeline ASR with LM. * Revamped into `self.decoder`. * Fixing. * 2nd fix. * Update src/transformers/pipelines/__init__.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Fixing. Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 11 Jan, 2022 3 commits
-
-
Sylvain Gugger authored
-
Matt authored
* Update TF test_step to match train_step * Update compile() warning to be clearer about what to pass
-
Vladimir Maryasin authored
All specific tokenizer config properties must be passed to its base class (XLMTokenizer) in order to be saved. This was not the case for do_lowercase config. Thus it was not saved by save_pretrained() method and saving and reloading the tokenizer changed its behaviour. This commit fixes it.
-