- 09 Aug, 2022 7 commits
-
-
YouJiacheng authored
* Recover _init_weights value in no_init_weights For potential nested use. In addition, users might modify private no_init_weights as well. * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Remove private variable change check Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Steven Liu authored
-
Nicolas Patry authored
* Adding a new `align_to_words` param to qa pipeline. * Update src/transformers/pipelines/question_answering.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Import protection. Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Younes Belkada authored
* attempt to fix attn mask device * fix bart `_prepare_decoder_attention_mask` - add correct device - run `make fix-copies` to propagate the fix
-
Yih-Dar authored
Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Thomas Chaigneau authored
* update features * MT5OnnxConfig added with updated with tests and docs * fix imports * fix onnc_config_cls for mt5 Co-authored-by: Thomas Chaigneau <thomas.deeptools.ai>
-
Niklas Hansson authored
* feat: add the data2vec conf that are missing https://huggingface.co/docs/transformers/serialization * fix: wrong config
-
- 08 Aug, 2022 17 commits
-
-
Younes Belkada authored
* add correct dtypes when checking for params dtype * forward contrib credits * Update src/transformers/modeling_utils.py Co-authored-by:
Thomas Wang <24695242+thomasw21@users.noreply.github.com> * more comments - added more comments on why we cast only floating point parameters * Update src/transformers/modeling_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
sgugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Thomas Wang <24695242+thomasw21@users.noreply.github.com>
-
AguilaCudicio authored
* Add Spanish translation of summarization.mdx * Apply suggestions from code review Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Ian Castillo authored
* Add file in spanish docs to be translated * Finish translation to Spanish * Improve Spanish wording * Add suggested changes from review
-
Rasmus Arpe Fogh Jensen authored
* Added accelerate gradient accumulation wrapper to run_image_classification_no_trainer.py example script * make fixup changes * PR comments * changed input to Acceletor based on PR comment, ran make fixup * Added comment explaining the sync_gradients statement * Fixed lr scheduler max steps * Changed run_clm_no_trainer.py script to use accelerate gradient accum wrapper * Fixed all scripts except wav2vec2 pretraining to use accelerate gradient accum wrapper * Added accelerate gradient accum wrapper for wav2vec2_pretraining_no_trainer.py script * make fixup and lr_scheduler step inserted back into run_qa_beam_search_no_trainer.py * removed changes to run_wav2vec2_pretraining_no_trainer.py script and fixed using wrong constant in qa_beam_search_no_trainer.py script
-
Mishig Davaadorj authored
-
NielsRogge authored
* Add videomae to doc tests * Add pip install decord Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Steven Liu authored
*
📝 add example of multimodal usage to pipeline tutorial *🖍 apply feedbacks *🖍 apply niels feedback -
Steven Liu authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Ankur Goyal authored
Co-authored-by:Ankur Goyal <ankur@impira.com>
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Fix compatibility with 1.12 * Remove pin from examples requirements * Update torch scatter version * Fix compatibility with 1.12 * Remove pin from examples requirements * Update torch scatter version * fix torch.onnx.symbolic_opset12 import * Reject bad version Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sourab Mangrulkar authored
* updating fsdp documentation * typo fix
-
Sylvain Gugger authored
* Clean up utils.hub * Remove imports * More fixes * Last fix
-
Nicolas Patry authored
* [DX fix] Fixing QA pipeline streaming a dataset. QuestionAnsweringArgumentHandler would iterate over the whole dataset effectively killing all properties of the pipeline. This restores nice properties when using `Dataset` or `Generator` since those are meant to be consumed lazily. * Handling TF better.
-
regisss authored
-
- 06 Aug, 2022 2 commits
-
-
Julien Chaumond authored
* zero chance anyone's using that constant no? * `transformers-cli login` => `huggingface-cli login` * `transformers-cli repo create` => `huggingface-cli repo create` * `make style`
-
Julien Chaumond authored
* Delete valohai.yaml * NLP => ML * typo * website supports https * datasets * 60k + modalities * unrelated link fixing for accelerate * Ok those links were actually broken * Fix link * Make `AutoTokenizer` auto-link * wording tweak * add at least one non-nlp task
-
- 05 Aug, 2022 13 commits
-
-
Julien Chaumond authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* Move cache folder to just huggingface * Thank you VsCode for this needless import * Move to hub * Forgot one
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* Draft new cached_file * Initial draft for config and model * Small fixes * Fix first batch of tests * Look in cache when internet is down * Fix last tests * Bad black, not fixing all quality errors * Make diff less * Implement change for TF and Flax models * Add tokenizer and feature extractor * For compatibility with main * Add utils to move the cache and auto-do it at first use. * Quality * Deal with empty commit shas * Deal with empty etag * Address review comments
-
Sylvain Gugger authored
* Fix pipeline tests * Make sure all pipelines tests run with init changes
-
Sylvain Gugger authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Seunghwan Hong authored
* Refactor `TFSwinLayer` to increase serving compatibility Signed-off-by:
Seunghwan Hong <seunghwan@scatterlab.co.kr> * Fix missed parameters while refactoring Signed-off-by:
Seunghwan Hong <seunghwan@scatterlab.co.kr> * Fix window_reverse to calculate batch size Signed-off-by:
Seunghwan Hong <harrydrippin@gmail.com> Co-Authored-By:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Seunghwan Hong authored
Signed-off-by:Seunghwan Hong <seunghwan@scatterlab.co.kr>
-
Nicolas Patry authored
* Adding a better error message when the model is improperly configured within transformers. * Update src/transformers/pipelines/__init__.py * Black version. * Overriding task aliases so that tokenizer+feature_extractor values are correct. * Fixing task aliases by overriding their names early * X. * Fixing feature-extraction. * black again. * Normalizing `translation` too. * Fixing last few corner cases. translation need to use its non normalized name (translation_XX_to_YY, so that the task_specific_params are correctly overloaded). This can be removed and cleaned up in a later PR. `speech-encode-decoder` actually REQUIRES to pass a `tokenizer` manually so the error needs to be discarded when the `tokenizer` is already there. * doc-builder fix. * Fixing the real issue. * Removing dead code. * Do not import the actual config classes.
-
- 04 Aug, 2022 1 commit
-
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-