- 18 Jan, 2022 14 commits
-
-
matt authored
-
Yih-Dar authored
* fix doc example - MarianForCausalLM example * try to keep copies * fix copies * fix more similar doc examples * fix more * fix style Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
PaulLerner authored
* fix: #14486 do not use BertPooler in DPR * fix tf dpr as well * finish Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
NielsRogge authored
* First draft * More improvements * More improvements * More improvements * Fix embeddings * Add conversion script * Finish conversion script * More improvements * Fix forward pass * Remove print statements * Add weights initialization * Add initialization of decoder weights * Add support for other models in the conversion script * Fix patch_size for huge model * Fix most of the tests * Fix integration test * Fix docs * Fix archive_list * Apply suggestions from code review * Improve documentation * Apply more suggestions * Skip some tests due to non-deterministic behaviour * Fix test_initialization * Remove unneccessary initialization of nn.Embedding * Improve docs * Fix dummies * Remove ViTMAEFeatureExtractor from docs * Add model to README and table of contents * Delete inference file
-
Suraj Patil authored
-
Sylvain Gugger authored
* Ignore empty subfolders when identifying submodules * Update utils/check_inits.py
-
Sylvain Gugger authored
-
Patrick von Platen authored
* [ASR pipeline] correct with lm pipeline * improve error
-
Sylvain Gugger authored
* Style docstrings when making/checking copies * Polish
-
Sylvain Gugger authored
* Fix deprecation warnings for int div Co-authored-by:
mgoldey <matthew.goldey@gmail.com> * Fix import * ensure that tensor output is python scalar * make backward compatible * make code more readable * adapt test functions Co-authored-by:
mgoldey <matthew.goldey@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
-
Li-Huai (Allan) Lin authored
* REALM initial commit * Retriever OK (Update new_gelu). * Encoder prediction score OK * Encoder pretrained model OK * Update retriever comments * Update docs, tests, and imports * Prune unused models * Make embedder as a module `RealmEmbedder` * Add RealmRetrieverOutput * Update tokenization * Pass all tests in test_modeling_realm.py * Prune RealmModel * Update docs * Add training test. * Remove completed TODO * Style & Quality * Prune `RealmModel` * Fixup * Changes: 1. Remove RealmTokenizerFast 2. Update docstrings 3. Add a method to RealmTokenizer to handle candidates tokenization. * Fix up * Style * Add tokenization tests * Update `from_pretrained` tests * Apply suggestions * Style & Quality * Copy BERT model * Fix comment to avoid docstring copying * Make RealmBertModel private * Fix bug * Style * Basic QA * Save * Complete reader logits * Add searcher * Complete searcher & reader * Move block records init to constructor * Fix training bug * Add some outputs to RealmReader * Add finetuned checkpoint variable names parsing * Fix bug * Update REALM config * Add RealmForOpenQA * Update convert_tfrecord logits * Fix bugs * Complete imports * Update docs * Update naming * Add brute-force searcher * Pass realm model tests * Style * Exclude RealmReader from common tests * Fix * Fix * convert docs * up * up * more make style * up * upload * up * Fix * Update src/transformers/__init__.py * adapt testing * change modeling code * fix test * up * up * up * correct more * make retriever work * update * make style * finish main structure * Resolve merge conflict * Make everything work * Style * Fixup * Fixup * Update training test * fix retriever * remove hardcoded path * Fix * Fix modeling test * Update model links * Initial retrieval test * Fix modeling test * Complete retrieval tests * Fix * style * Fix tests * Fix docstring example * Minor fix of retrieval test * Update license headers and docs * Apply suggestions from code review * Style * Apply suggestions from code review * Add an example to RealmEmbedder * Fix Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
* fix doc example - NameError: name 'PATH' is not defined * fix name 'TFRagModel' is not defined * correct TFRagRagSequenceForGeneration * fix name 'tf' is not defined * fix style Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Nicolas Patry authored
* `is_ctc` needs to be updated to `self.type == "ctc". * Adding fast test for this functionality.
-
- 17 Jan, 2022 5 commits
-
-
Yih-Dar authored
* fix doc example - cannot import name 'UniSpeechSatFeatureEncoder' * fix ckpt name Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Stas Bekman authored
add new paper
-
Matt authored
-
MrinalTyagi authored
-
- 16 Jan, 2022 1 commit
-
-
Patrick von Platen authored
-
- 15 Jan, 2022 1 commit
-
-
Stas Bekman authored
* Efficient Software Prebuilds * improve
-
- 14 Jan, 2022 11 commits
-
-
Joao Gante authored
-
Sylvain Gugger authored
* Better dummies * See if this fixes the issue * Fix quality * Style * Add doc for DummyObject
-
Nicolas Patry authored
* Fixing flaky test (hopefully). * tf compliant.
-
Joao Gante authored
* TF Bert inference - support np.ndarray optional arguments * apply np input tests to all TF architectures
-
AK391 authored
* update XLMProphetNet link * update DPR link * change prophetnet link * change link MBART * change link GPT * update gpt2 link * ctrl update link * update Transformer-XL link * Update Reformer link * update xlnet link * bert update link * udpate albert link * roberta update link * update distilbert link * update convbert link * update XLM link * xlm roberta update link * update Flaubert link * update electra link * update funnel transformer and longformer * bart update link * pegasus update link * udpate marianmt link * t5 update link * mt5 update link
-
novice authored
-
SaulLu authored
* add new test * fix in init * more relevant test
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* Make sure all submodules are properly registered * Try to fix tests * Fix tests
-
Joao Gante authored
-
Sylvain Gugger authored
* Check the repo consistency in model templates test * Fix doc template * Fix docstrings * Fix last docstring
-
- 13 Jan, 2022 8 commits
-
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Yanming Wang authored
* Multiple fixes of trainer class with XLA GPU * Make fp16 valid for xla:gpu * Add mark_step in should_log to reduce compilation overhead
-
Carlos Aguayo authored
Fix typo
-
Manuel R. Ciosici authored
* Add AdamW deprecation warning * Add --optim to Trainer * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/optimization.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py * fix style * fix * Regroup adamws together Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Change --adafactor to --optim adafactor * Use Enum for optimizer values * fixup! Change --adafactor to --optim adafactor * fixup! Change --adafactor to --optim adafactor * fixup! Change --adafactor to --optim adafactor * fixup! Use Enum for optimizer values * Improved documentation for --adafactor Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Add mention of no_deprecation_warning Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename OptimizerOptions to OptimizerNames * Use choices for --optim * Move optimizer selection code to a function and add a unit test * Change optimizer names * Rename method Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename method Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Remove TODO comment Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename variable Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename variable Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Rename function * Rename variable * Parameterize the tests for supported optimizers * Refactor * Attempt to make tests pass on CircleCI * Add a test with apex * rework to add apex to parameterized; add actual train test * fix import when torch is not available * fix optim_test_params when torch is not available * fix optim_test_params when torch is not available * re-org * small re-org * fix test_fused_adam_no_apex * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Remove .value from OptimizerNames * Rename optimizer strings s|--adam_|--adamw_| * Also rename Enum options * small fix * Fix instantiation of OptimizerNames. Remove redundant test * Use ExplicitEnum instead of Enum * Add unit test with string optimizer * Change optimizer default to string value Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas@stason.org>
-
Stas Bekman authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Lysandre Debut authored
* Updated script * Commit everything * Ready for review! * Update .github/workflows/build_documentation.yml Co-authored-by:
Julien Chaumond <julien@huggingface.co> Co-authored-by:
Julien Chaumond <julien@huggingface.co>
-