- 28 Mar, 2022 5 commits
-
-
Shang Zhang authored
* update Dockerfile and utils_qa * Update README.md
-
Julien Chaumond authored
-
Kurian Benoy authored
* Update marian.mdx * Update marian.mdx * Update docs/source/model_doc/marian.mdx Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update marian.mdx Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Jia authored
-
Nathan Glenn authored
* fix confusing PIL instructions As stated in the documentation [here](https://pillow.readthedocs.io/en/stable/handbook/image-file-formats.html?highlight=pdf#write-only-formats ), PIL can only write PDF's, not read them. Remove references to reading PDF's via PIL from this page to avoid confusion. * mention PDF in doc examples using PIL Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Be explicit: PDFs must be converted to images * fix formatting Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com>
-
- 27 Mar, 2022 1 commit
-
-
Shamima authored
-
- 25 Mar, 2022 11 commits
-
-
Steven Liu authored
*
✨ create concept guide section *🖍 make fixup *🖍 apply feedback Co-authored-by:Steven <stevhliu@gmail.com>
-
Daniel Stancl authored
* Initial commit * Add TFGPTJModel * Fix a forward pass * Add TFGPTJCausalLM * Add TFGPTJForSequenceClassification * Add TFGPTJForQuestionAnswering * Fix docs * Deal with TF dynamic shapes * Add Loss parents to models * Adjust split and merge heads to handle 4 and 5-dim tensors * Update outputs for @tooslow tests
-
Sanchit Gandhi authored
-
Sanchit Gandhi authored
-
lewtun authored
* Add ONNX support for Blenderbot * Add BlenderbotSmall ONNX configuration * Update serialization table
-
Sylvain Gugger authored
* Sharded checkpoint support * Handle distant sharded checkpoints * Add tests * TODO is done * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Fix docstring * Add example and format * Address review comments * More review comments * End of merge * Revert unintentional change * VsCode what did you do? * Style * Changes * Address final comments * Quality * Moar tests * Move import beneath is_pt_available Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Matt authored
-
Sylvain Gugger authored
-
Atharva Ingle authored
-
Sylvain Gugger authored
* Big file_utils cleanup * This one still needs to be treated separately
-
Michael Benayoun authored
-
- 24 Mar, 2022 13 commits
-
-
NielsRogge authored
Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Yi Heng Lim authored
* Added type hints for PyTorch T5 model * removed a type hint * ran make style * added type hints for ibert pytorch * added type hints for lxmert pytorch * removed kwargs type hint and fixed arguments order
-
Sylvain Gugger authored
-
Robot Jelly authored
* variable naming for Distilbert model * adding unpack inputs at top * make style/quality Co-authored-by:matt <rocketknight1@gmail.com>
-
Sylvain Gugger authored
* Fix doc links in README * Fix name * Fix links in READMEs and doc index * Error if there is something wrong so the CI knows
-
Lysandre Debut authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* update PT/Flax equivalence tests on PT side * overwrite check_outputs in BigBirdModelTest Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Zehua Li authored
* Add missing type hints for ConvBERT flavored models. * Update src/transformers/models/convbert/modeling_convbert.py Co-authored-by:Matt <Rocketknight1@users.noreply.github.com>
-
Dahlbomii authored
* Type hints and TF decorator added * Re-add XLA generation method * Re-add lines that were deleted by conflicting updates * Re-add lines that were deleted by conflicting updates * Re-add lines that were deleted by conflicting updates Co-authored-by:matt <rocketknight1@gmail.com>
-
Yih-Dar authored
* fix * update the expected value in test_fast_integration Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Nathan Cooper authored
* Update readme with how to train offline and fix BPE command * Update examples/research_projects/codeparrot/README.md Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * Update examples/research_projects/codeparrot/README.md Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * Update examples/research_projects/codeparrot/README.md Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * Update examples/research_projects/codeparrot/README.md Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com>
-
Yih-Dar authored
* update doc examples for TFRoberta * fix style * fix style * use TF ckpt * apply suggestion * add the code file to test here * fix style Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 23 Mar, 2022 10 commits
-
-
Patrick von Platen authored
* [Doctests] Make roberta-like meaningfull * correct * final correct * Trigger test * make style * apply suggestion from sylvain
-
Xu Zhao authored
* Make BigBird model compatiable to fp16 dtype. * Use tree_map instead of map * Reformat the code * Fix import order * Convert masks to the correct dtype * Fix format issue * Address comments.
-
Yih-Dar authored
* Update docs/README.md Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Thomas Chaigneau authored
* add GPT-J ONNX config to Transformers * remove token-classification features mapping Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> * add question-answering features mapping Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> * add GPT2 config init to GPT2 config + copie shebang for fix-copies Co-authored-by:
ChainYo <t.chaigneau.tc@gmail.com> Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com>
-
Edward Beeching authored
* Created the Decision Transformer Modle * updating tests, copy to other machine * Added last hidden size to Decision Transformer modelling outputs * Removed copy of original DT file * made a temporary change to gpt2 to have it conform with the Decision Transformer version * Updated tests * Ignoring a file used to test the DT model * added comments to config file * added comments and argument descriptions to decision transformer file * Updated doc * Ran "make style" * Remove old model imports * Removed unused imports, cleaned up init file * Update docs/source/model_doc/decision_transformer.mdx added my username Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Reverted changes made to gpt2 * Removed datasets submodule * Update the modeling outputs to include gpt2 attentions, hidden states and last hidden states * Added support for return of hidden states, attentions and return dict of gpt2 model. * Updated tests to include many of the ModelTesterMixin tests. The following tests are skipped: test_generate_without_input_ids, test_pruning, test_resize_embeddings, test_head_masking, test_attention_outputs, test_hidden_states_output, test_inputs_embeds, test_model_common_attributes * Added missing line to the end of gpt2 file * Added an integration test for the Decision Transformer Test performs and autoregressive evaluation for two time steps * Set done and info to _ to fix failing test * Updated integration test to be deterministic and check expected outputs * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Removed unnecessary config options * Cleaned up commented code and old comments. * Cleaned up commented code. * Changed DecisionTransformer to Decision Transformer * Added Decision Transformer to the main README file * Added copy of GTP2 called DecisionTranformerGPT2Model * isorted imports * isorted imports * Added model to non-English README files * Ran make fix-copies and corrected some cases. * Updated index file to include Decision Transformer * Added gpt2 model as copy inside the Decision Transformer model file * Added the unit test file to the list of TEST_FILES_WITH_NO_COMMON_TESTS * Deleted redundant checkpoint files (I don't know how these got committed) * Removed testing files. (These should have never been committed) * Removed accidentally committed files * Moved the Decision Transformer test to its own directory * Add type hints for Pegasus (#16324) * Funnel type hints (#16323) * add pt funnel type hints * add tf funnel type hints * Add type hints for ProphetNet PyTorch (#16272) * [GLPN] Improve docs (#16331) * Add link to notebook * Add link * Fix bug Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local> * Added type hints for Pytorch Marian calls (#16200) * Added type hinting for forward functions in pytorch marian * typo correction * Removed type hints on functions from BART per Suraj Patil request * fix import pb * fix typo * corrected tuple call * ran black * after fix-copies Some optional tags on primitives were removed, past_key_values in MarianForCausalLM changed from Tuple of Tuple to List * Fixing copies to roformer and pegasus Co-authored-by:
Clementine Fourrier <cfourrie@inria.fr> Co-authored-by:
matt <rocketknight1@gmail.com> * Moved DecisionTransformOutput to modeling_decision_transformer * Moved the example usage to research project and cleaned comments * Made tests ignore the copy of gpt2 in Decision Transformer * Added module output to modelling decision transformer * removed copied gpt2 model from list of transformers models * Updated tests and created __init__ file for new test location * Update README.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/configuration_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Removed unneeded summary type from config file * Fixed copies * Updated pretrained config map to refer to hopper-medium checkpoint * done (#16340) * Added Decision transformer to model docs * Update src/transformers/models/decision_transformer/modeling_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/modeling_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/configuration_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Add type annotations for Rembert/Splinter and copies (#16338) * undo black autoformat * minor fix to rembert forward with default * make fix-copies, make quality * Adding types to template model * Removing List from the template types * Remove `Optional` from a couple of types that don't accept `None` Co-authored-by:
matt <rocketknight1@gmail.com> * [Bug template] Shift responsibilities for long-range (#16344) * Fix code repetition in serialization guide (#16346) * Adopt framework-specific blocks for content (#16342) *
✨ refactor code samples with framework-specific blocks *✨ update training.mdx *🖍 apply feedback * Updates the default branch from master to main (#16326) * Updates the default branch from master to main * Links from `master` to `main` * Typo * Update examples/flax/README.md Co-authored-by:Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Updated model with custom docstring example * Created the Decision Transformer Modle * updating tests, copy to other machine * Added last hidden size to Decision Transformer modelling outputs * Removed copy of original DT file * made a temporary change to gpt2 to have it conform with the Decision Transformer version * Updated tests * Ignoring a file used to test the DT model * added comments to config file * added comments and argument descriptions to decision transformer file * Updated doc * Ran "make style" * Remove old model imports * Removed unused imports, cleaned up init file * Update docs/source/model_doc/decision_transformer.mdx added my username Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Reverted changes made to gpt2 * Removed datasets submodule * Update the modeling outputs to include gpt2 attentions, hidden states and last hidden states * Added support for return of hidden states, attentions and return dict of gpt2 model. * Updated tests to include many of the ModelTesterMixin tests. The following tests are skipped: test_generate_without_input_ids, test_pruning, test_resize_embeddings, test_head_masking, test_attention_outputs, test_hidden_states_output, test_inputs_embeds, test_model_common_attributes * Added missing line to the end of gpt2 file * Added an integration test for the Decision Transformer Test performs and autoregressive evaluation for two time steps * Set done and info to _ to fix failing test * Updated integration test to be deterministic and check expected outputs * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Removed unnecessary config options * Cleaned up commented code and old comments. * Cleaned up commented code. * Changed DecisionTransformer to Decision Transformer * Added Decision Transformer to the main README file * Added copy of GTP2 called DecisionTranformerGPT2Model * isorted imports * isorted imports * Added model to non-English README files * Ran make fix-copies and corrected some cases. * Updated index file to include Decision Transformer * Added gpt2 model as copy inside the Decision Transformer model file * Added the unit test file to the list of TEST_FILES_WITH_NO_COMMON_TESTS * Deleted redundant checkpoint files (I don't know how these got committed) * Removed testing files. (These should have never been committed) * Removed accidentally committed files * Moved the Decision Transformer test to its own directory * Moved DecisionTransformOutput to modeling_decision_transformer * Moved the example usage to research project and cleaned comments * Made tests ignore the copy of gpt2 in Decision Transformer * Added module output to modelling decision transformer * removed copied gpt2 model from list of transformers models * Updated tests and created __init__ file for new test location * Update README.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/configuration_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Removed unneeded summary type from config file * Fixed copies * Updated pretrained config map to refer to hopper-medium checkpoint * Added Decision transformer to model docs * Update src/transformers/models/decision_transformer/modeling_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/modeling_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/decision_transformer/configuration_decision_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Updated model with custom docstring example * Updated copies, config auto, and readme files. Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Dan Tegzes <48134725+Tegzes@users.noreply.github.com> Co-authored-by:
Adam Montgomerie <adam@avanssion.com> Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local> Co-authored-by:
Clémentine Fourrier <22726840+clefourrier@users.noreply.github.com> Co-authored-by:
Clementine Fourrier <cfourrie@inria.fr> Co-authored-by:
matt <rocketknight1@gmail.com> Co-authored-by:
Francesco Saverio Zuppichini <francesco.zuppichini@gmail.com> Co-authored-by:
Jacob Dineen <54680234+jacobdineen@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr>
-
Sylvain Gugger authored
* Make Transformers use cache files when hf.co is down * Fix tests * Was there a random circleCI failure? * Isolate patches * Style * Comment out the failure since it doesn't fail anymore * Better comment
-
OllieBroadhurst authored
* Swap inequalities * Update src/transformers/trainer_callback.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/trainer_callback.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Joao Gante authored
* revert tf gpt2 * add test for unpack_inputs and fix test case * add changes to vision encoder decoder
-
Sylvain Gugger authored
-
João Gustavo A. Amorim authored
-