- 03 Apr, 2023 17 commits
-
-
Xuehai Pan authored
* [setup] migrate setup script to `pyproject.toml` * [setup] cleanup configurations * remove unused imports
-
Vladimir Blagojevic authored
-
Xuehai Pan authored
* [setup] drop deprecated `distutils` usage * drop deprecated `distutils.util.strtobool` usage * fix import order * reformat docstring by `doc-builder`
-
Ilya authored
-
Younes Belkada authored
* enable PP for T5 * make fixup * fix failing tests
-
Younes Belkada authored
[`Trainer`] Force `is_model_parallel` when model is loaded in multiple GPUs using `accelerate` (#22532) * add `is_model_parallel` arg on Trainer * add warning * adapt from suggestions * revert t5 changes * remove commas * adapt from suggestions
-
zhbh01 authored
-
Thibault Douzon authored
LayoutLMv3TokenizerFast produces empty '臓' token with `offset_mapping = (0, 0)`. Next token is wrongly assumed to also be beginning of word and isn't correctly assigned `pad_token_label`. Modify test with text that produce '臓' token. Remove copy check from LayoutLMv2TokenizerFast for `_batch_encode_plus`. solves issue: #19978
-
Kirill authored
-
larekrow authored
`load_checkpoint()` silently fails because `".qkj_proj." in key` is always `False`, but will eventually cause an error at `model.load_state_dict(state_dict)`.
-
Joao Gante authored
* haha text go brrr (but in gradio)
-
Mohammed Jabir authored
* added biogpt token classifier * fix reviews * Updated modeling_biogpt.py Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com> --------- Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
-
Jungnerd authored
docs: ko: sagemaker.mdx
-
Arthur authored
* draft * update tokenization limma and conversion script * more udpates * initial commit * style * default pad to None * draft tokenization tests * update test * update tokenization tests * nits * update * versioning test * major fix * fix more testst * finish fixing special masks * last nit * more nits * add encode decode tests * add more * fix token type ids * style
-
Eli Simhayev authored
added > 0.5 to `past_observed_mask`
-
amyeroberts authored
* Add out_indices to backbones, deprecate out_features * Update - can specify both out_features and out_indices but not both * Can specify both * Fix copies * Add out_indices to convnextv2 configuration
-
kevinpro authored
-
- 31 Mar, 2023 6 commits
-
-
Sylvain Gugger authored
* Test fetcher v2 * Fix regexes * Remove sanity check * Fake modification to OPT * Fixes some .sep issues * Remove fake OPT change * Fake modif for BERT * Fake modif for init * Exclude SageMaker tests * Fix test and remove fake modif * Fake setup modif * Fake pipeline modif * Remove all fake modifs * Adds options to skip/force tests * [test-all-models] Fake modif for BERT * Try this way * Does the command actually work? * [test-all-models] Try again! * [skip circleci] Remove fake modif * Remove debug statements * Add the list of important models * Quality * Update utils/tests_fetcher.py Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr> * Address review comments * Address review comments * Fix and add test * Apply suggestions from code review Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com> * Address review comments --------- Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Yih-Dar <2521628+ydshieh@users.noreply.github.com>
-
Sabine authored
* update NeptuneCallback docstring * formatting * apply make style --------- Co-authored-by:Aleksander Wojnarowicz <alwojnarowicz@gmail.com>
-
dependabot[bot] authored
Bump redis in /examples/research_projects/decision_transformer Bumps [redis](https://github.com/redis/redis-py) from 4.5.3 to 4.5.4. - [Release notes](https://github.com/redis/redis-py/releases) - [Changelog](https://github.com/redis/redis-py/blob/master/CHANGES) - [Commits](https://github.com/redis/redis-py/compare/v4.5.3...v4.5.4 ) --- updated-dependencies: - dependency-name: redis dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Nicolas Patry authored
* Making sure we can use safetensors to serialize all the time. * Expanding the tests for increased coverage. * Update the test. * Getting current state of affairs. * Tentative fix. * Fixing black version. * Fixing the worst offenders. * Try to modify less files. * Fixing blip_2 (Weird solution right now). * Fixing deta. * Fix blip ? * Missing extra newline. * No deta modification. * Adding some comments. * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Addressing comments. * Addressing comments. * creating warn_once. * Warning_once ! --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
lewtun authored
* Relax checks from to warning * Fix style * Replace warnings with logger * Use warning vs warn
-
- 30 Mar, 2023 10 commits
-
-
Yih-Dar authored
* Enable Nightly + Past CI * put schedule --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Manuel de Prada authored
Docs fix: Multinomial sampling decoding needs "num_beams=1", since by default it is usually not 1. (#22473) Fix: Multinomial sampling needs "num_beams=1", since by default is 5.
-
Joao Gante authored
* Llama now supports max_position_embeddings * Save config; Cosmetic edits
-
Arthur authored
edit default model type and testing path set to hf-internal-testing
-
Roy Hvaara authored
Guard imports that use the tokenizers library
-
amyeroberts authored
Fix ordering of height,width for BLIP
-
Joao Gante authored
* haha tokens go brrrr
-
amyeroberts authored
Skip flaky test for now
-
amyeroberts authored
* Rescale image back if it was scaled during PIL conversion * do_rescale is defined if PIL image passed in
-
amyeroberts authored
* Move common properties to BackboneMixin * Fix failing tests * Update ConvNextV2 backbone
-
- 29 Mar, 2023 7 commits
-
-
Stefan Heng authored
* Update: ignore padding support for TransfoXL training when n_clusters==0 * Update: transformer XL always pad * Update: drop doc
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sabine authored
-
jeffhataws authored
This reverts commit fd81746dbec5f17c8285a0fdc72ca4b4c025cc33.
-
Younes Belkada authored
fix slow test
-