- 03 May, 2022 7 commits
-
-
Pavel Belevich authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Fix RNG reload in resume training from epoch checkpoint * Fix test
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Make Trainer compatible with sharded checkpoints * Add doc
-
Yih-Dar authored
* move test model folders (TODO: fix imports and others) * fix (potentially partially) imports (in model test modules) * fix (potentially partially) imports (in tokenization test modules) * fix (potentially partially) imports (in feature extraction test modules) * fix import utils.test_modeling_tf_core * fix path ../fixtures/ * fix imports about generation.test_generation_flax_utils * fix more imports * fix fixture path * fix get_test_dir * update module_to_test_file * fix get_tests_dir from wrong transformers.utils * update config.yml (CircleCI) * fix style * remove missing imports * update new model script * update check_repo * update SPECIAL_MODULE_TO_TEST_MAP * fix style * add __init__ * update self-scheduled * fix add_new_model scripts * check one way to get location back * python setup.py build install * fix import in test auto * update self-scheduled.yml * update slack notification script * Add comments about artifact names * fix for yolos Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sanchit Gandhi authored
* [FlaxBert] Add ForCausalLM * make style * fix output attentions * Add RobertaForCausalLM * remove comment * fix fx-to-pt model loading * remove comment * add modeling tests * add enc-dec model tests * add big_bird * add electra * make style * make repo-consitency * add to docs * remove roberta test * quality * amend cookiecutter * fix attention_mask bug in flax bert model tester * tighten pt-fx thresholds to 1e-5 * add 'copied from' statements * amend 'copied from' statements * amend 'copied from' statements * quality
-
- 02 May, 2022 17 commits
-
-
Patrick von Platen authored
* [T5 Tokenizer] Model has no fixed position ids - there is no hardcoded max length * [T5 Tokenizer] Model has no fixed position ids - there is no hardcoded max length * correct t5 tokenizer * correct t5 tokenizer * fix test * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * finish Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
* Clean up setup.py * Trigger CI * Upgrade Python used
-
Lysandre Debut authored
* Make sacremoses optional * Pickle
-
Lysandre Debut authored
-
NielsRogge authored
* First draft * Add YolosForObjectDetection * Make forward pass work * Add mid position embeddings * Add interpolation of position encodings * Add expected values * Add YOLOS to tests * Add integration test * Support tiny model as well * Support all models in conversion script * Remove mid_pe_size attribute * Make more tests pass * Add model to README and fix config * Add copied from statements * Rename base_model_prefix to vit * Add missing YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP * Apply suggestions from code review * Apply more suggestions from code review * Convert remaining checkpoints * Improve docstrings * Add YolosFeatureExtractor * Add feature extractor to docs * Add corresponding tests * Fix style * Fix docs * Apply suggestion from code review * Fix bad rebase * Fix some more bad rebase * Fix missing character * Improve docs and variable names Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Zachary Mueller authored
* Update all examples to properly calculate progress bar
-
Zachary Mueller authored
* Propagate and fix imports
-
calpt authored
-
NielsRogge authored
* Clean up tests * Make fixup Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Sylvain Gugger authored
-
yujun authored
* add torch.no_grad when in eval mode * make style quality
-
Martin Pömsl authored
-
Sanchit Gandhi authored
* [FlaxSpeechEncoderDecoder] Fix bug in `decoder_module` * [FlaxEncoderDecoder] Fix bug in `decoder_module`
-
Sylvain Gugger authored
-
Michael Benayoun authored
* Add meta proxy * Uses meta data to trace data dependent control-flow * Remove commented class * Handles torch creating functions * Added type annotation to fix tracing * Tracing works for everything but T5 and GPT-J * Almost all previously supported models pass * All architectures can be traced except T5 * Intermediate commit to have a trace of the comparison operators for HFProxy * Everything works, except loss computation * Everything works * Removed unused import * Overriden methods do not use underlying ops (linear and torch.matmul), and model attributes are copied to the traced version * Fix torch_matmul_override * Change attributes reference to deepcopy * Remove breakpoint and add torch_index_override * Small fix * Fix typo * Replace asserts by explicit exceptions
-
Sanchit Gandhi authored
-
Manan Dey authored
-
- 30 Apr, 2022 2 commits
-
-
Omar U. Espejel authored
* Add translating guide
-
Yih-Dar authored
* Add the check * add missing ckpts * add a list to ignore * call the added check script * better regex pattern Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 29 Apr, 2022 9 commits
-
-
Sylvain Gugger authored
* Result of new doc style with fixes * Add last two files * Bump hf-doc-builder
-
Sylvain Gugger authored
* Replace dict/BatchEncoding instance checks by Mapping * Typo
-
Nicolas Patry authored
This reverts commit 4f3a14e3.
-
Nicolas Patry authored
-
tarzan authored
-
Pavel Belevich authored
-
Joao Gante authored
-
Zachary Mueller authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 28 Apr, 2022 5 commits
-
-
Sylvain Gugger authored
-
Zachary Mueller authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
amyeroberts authored
-
conan1024hao authored
* dd parameter --config_overrides for run_mlm_wwm.py * linter
-