- 22 Feb, 2023 3 commits
-
-
Younes Belkada authored
fix typo
-
Aaron Gokaslan authored
-
Kashif Rasul authored
* added loc and scale outputs from scalers * fix typo * fix tests * fixed formatting * initial StdScaler * move scaling to optional str * calculate std feature for scalers * undid change as it does not help * added StdScaler with weights * added input projection layer and d_model hyperparam * use linear proj * add back layernorm_embedding * add sin-cos pos embeddings * updated scalers * formatting * fix type * fixed test * fix repeated_past_values cal. * fix when keepdim=false * fix default_scale * backward compatibility of scaling config * update integration test expected output * fix style * fix docs * use the actual num_static_real_features in feature_dim cal * clarified docs * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * prediction_length is not optional * fix for reviewer * Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * get rid of un-needed new lines * fix doc * remove unneeded new lines * fix style * static_categorical_features and static_real_features are optional * fix integration test * Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> * fixing docs for multivariate setting * documentation for generate --------- Co-authored-by:
NielsRogge <48327001+NielsRogge@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 21 Feb, 2023 8 commits
-
-
mollerup23 authored
* Adding type hints to call() functions in this file * make fixup * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py * Update src/transformers/models/marian/modeling_tf_marian.py --------- Co-authored-by:
Matt <rocketknight1@gmail.com> Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com>
-
Maria Khalusova authored
* added resources: links to task guides that support these models * minor polishing * conflict resolved * link fix * Update docs/source/en/model_doc/vision-encoder-decoder.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
* fix tvlt ci * fix tvlt ci * fix tvlt ci --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Fix get_class_in_module Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* Add test * Fix GITProcessor * Update --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Arthur authored
-
Ishan Jindal authored
* Update reformer.mdx Fix axial positional encoding calculations * Update docs/source/en/model_doc/reformer.mdx Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Jonatan K艂osko authored
* Add WhisperTokenizerFast * Fixup * Up * Up * Improve tests * Update src/transformers/models/whisper/tokenization_whisper_fast.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Keep stride in whisper pipelien test * Remove unknown token special case * Reduce vocabulary size in tests * Fix vocab size assertion * Sync copied changes from WhisperTokenizer * Skip pipeline tests * Update assertion * Remove Whisper tokenizer dependency on sentencepiece * Format --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
- 20 Feb, 2023 9 commits
-
-
Sylvain Gugger authored
-
Arthur authored
pin pytorch lightning requirement
-
Alara Dirik authored
* Add EfficientNet to transformers
-
Younes Belkada authored
* fix `bnb` decoders bug * make fixup
-
tanreinama authored
* add GPTSAN-Japanese * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN (update for review) * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * fix typo in comment text * add GPTSAN * add GPTSAN * add GPTSAN * add GPTSAN * fix document and comments * fix class name GPTSAN->GPTSan * fix import and test for tokenizer
-
Sylvain Gugger authored
-
Morgan McGuire authored
* add prints * add shape * add reshape * clean up
-
Andy Ehrenberg authored
* add flax whisper implementation * rever change to setup * remove unused imports * revert generation changes * flax whisper docs * docs * import order * import sorting * isort * add dummy objects * doc formatting * formatting * remove trailing whitespaces * fix flax whisper docs * add generation logic to unlock flax whisper * remove scans * give credits to Flax Bart implementation * remove unused imports * add license * remove assert * more credits to Bart * fix style * formatting * support left padding * add flax whisper generation test * remove copied from comments whenever not a full copy * fix docstrings for logits processors * revert change to FlaxForceTokensLogitsProcessor * revert doc changes * improve generation docs * reorganize * formatting * cleanup docs * add tests * handle empty list case * fix forced decoder ids in flax tests * add flax whisper to inits * upate dummy objects * docs for FlaxAutoModelForSpeechSeq2Seq * fix decoder_position_ids computation in pretrained model decode/__call__ fns * add Copied from statements as necessary * compute position_ids only in __call__ and decode methods of pretrained model subclasses * improve readabilityof compute positional embeddings * check dimensionality of input_features instead of hidden_states * copied from statement for init_cache * formatting * fix copies * fix copies * pass attention mask to encoder layers * fix decoder module outputs * set dtype Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * smaller flax model for whisper test * Update src/transformers/generation/flax_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/whisper/modeling_flax_whisper.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update tests/models/whisper/test_modeling_flax_whisper.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * cleanup Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/whisper/modeling_flax_whisper.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * bias cleanup * doc fix * align style for force tokens processor * readability * fix input shape in tests * revert FlaxGenerationMixin docstring * formatting * fix tests * fix imports * consistent encoder hidden states * consistent hidden states * input shapes * typo * partial class trick * partial class for input shape * base_class with correct input shape * partial base classes * match by name * set main_input_name * compare on names * formatting * remove unused import * safer position ids computation * safer position id computation * Update src/transformers/models/whisper/modeling_flax_whisper.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update src/transformers/models/whisper/modeling_flax_whisper.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove identical inherited tests * fix prompt ids in tests * use generation config * use jnp array * better var names * more explicit bias use * import transformers * formatting * test formatting * remove unused imports * remove unused imports * formatting * isort * docs * fix ln orders for encoder hidden states * whisper unique generation stuff * flake * use finfo for attention bias * docs * Update src/transformers/generation/flax_utils.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * docs * add timestamp flax test * jit for timestamps * formatting * clean up timestamps processor * formatting * remove if_true * cleanup --------- Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
AlexWertheim authored
* Reinserted import statement accidentally removed during rebasing. * Added auto_wrap functionality, restructured XLA FSDP logic to more closely match PyTorch FSDP logic. * Fixed flag descriptions; changed several instances of fsdp_ to xla_fsdp_; pass in auto_wrap_policy and auto_wrapper_callable directly to avoid lambda saving. * Moved XLA FSDP logic to be adjacent to Fairscale FSDP logic in trainer. * Formatted changes in accordance with HF style requirements. * Added back in warning which was accidentally removed. * - Merged XLA FSDP training arguments into `fsdp_config` - Added `xla` boolean flag to `fsdp_config` to specify XLA FSDP wrapping - Merged XLA FSDP wrapping logic into FSDP wrapping logic within trainer class * Cleaned up errors, moved argument to fsdp_config - Set `xla` and `xla_fsdp_grad_ckpt` flags by default in fsdp_config - Added missing colons following conditionals - Moved `fsdp_transformer_layer_cls_to_wrap` to `fsdp_config` - Modified `fsdp_transformer_layer_cls_to_wrap` to be list of strings, not just one string - Changed Fairscale FSDP logic to allow for set of layer classes to wrap - Removed unnecessary checks for `xla_fsdp` * Corrected small errors, improved layer class flag - Correctly set default values for `xla` and `xla_fsdp_grad_ckpt` arguments - Made `fsdp_transformer_layer_cls_to_wrap` a list of strings instead of a single string - Added processing to ensure that `fsdp_transformer_layer_cls_to_wrap` works as expected if passed as a single string - Updated PyTorch FSDP logic to accept a list of layers to wrap, as done with XLA FSDP - Replaced instances of `getattr()` with `.get()` for dictionary retrievals with default values, including when setting `fsdp_min_num_params` - Corrected `self.fsdp is not None` to `len(self.fsdp) > 0` - Removed extraneous `xla_fsdp` argument descriptions from outside `fsdp_config` * Changed xla-fsdp-settings to be dictionary - Modified xla-fsdp-settings to be entered directly as dictionary instead of loaded through JSON file - Made small style corrections * Reverted unintentional local_rank TPU check * Do not block XLA FSDP if local rank is -1 * Rebased and applied automatic formatting - Rebased - Applied automatic formatting changes via `make style` * Applied automatic formatting with latest version of black * Replaced expression with * Reran black examples tests src utils ruff examples tests src utils --fix make autogenerate_code make[1]: Entering directory '/usr/local/google/home/awertheim/HF-FSDP-PR/transformers' make[1]: Leaving directory '/usr/local/google/home/awertheim/HF-FSDP-PR/transformers' after additional formatting changes * Additionall automatic formatting changes * Remove unnecessary whitespace characters from src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 17 Feb, 2023 7 commits
-
-
Yih-Dar authored
* fix dynamic module import error --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Younes Belkada authored
* update blip path * Update tests/models/blip/test_modeling_blip.py
-
Younes Belkada authored
[`ImageProcessor`] Refactor default `mean` & `std` to `OPENAI_CLIP_MEAN` & `OPENAI_CLIP_STD` (#21425) * fix default value * add the fix on other models
-
Joao Gante authored
-
Yoshinari Fujinuma authored
Co-authored-by:Yoshinari Fujinuma <fujinuy@amazon.com>
-
Younes Belkada authored
* add `is_longer` * fix docstring * fix config class * fix loss * fix all doctests * fix order * fix last failing tests --------- Co-authored-by:arthur.zucker@gmail.com <arthur.zucker@gmail.com>
-
Younes Belkada authored
* v1 `BitsandbytesConfig` - add v1 - add tests - more user-friendly API - add docs * change to `BitsAndBytesConfig` * replace logic * changes * make fixup * quality * make fixup * fix doc * fix test * update toctree * fix slow test * add tips * add warning * change title * oops * Update docs/source/en/main_classes/quantization.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/utils/bitsandbytes.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * remove unused file * adapt suggestion - add also tests - change logic * update docs * adapt suggestions --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 16 Feb, 2023 11 commits
-
-
Steven Anton authored
* * Properly register parameters in PerceiverMultimodalPreprocessor * Adapt PerceiverTextPreprocessor to work with PerceiverMultimodalPreprocessor * Change a few type hints * Fix formatting; incorrect return type * Return embeddings_wo_pos --------- Co-authored-by:Steven Anton <antonstv@amazon.com>
-
Arthur authored
* add model like clip * update * text model ok * clap text works * some refactor - `CLAPVision` to `CLAPAudio` - refactor kwargs of audio modules * more refactor * more refactor * more refactor * correct fusion * more refactor * new modules * add basic processor * fixup * remove whisper copioed from * audio logits match * add doc * correct filters mel and add maxlength * style * few fixes * forward passes * fixup * fixup * some clean up * remove mels form the dictionnary * pad after the repeat * update padding when dsmaller * fix padding * style * use swin patch merging * use copied from swin * processor with any tokenizer * more copied from * some clean up * more refactor * fix mel when rand_trunc * style * remove unused imports * update processing * remove image processing tests * add testing fiel * fixmodeling issues * replace with `is_longer` * clap in serialization * more refactor * `make fixup` * make fixup * fix feature extractor * update test feature extractor * `make fixup` * clean up config * more clean up * more cleanup * update tests * refactor tests and inits * removeCLAP vision config * remove CLAP from image procssing auto and dummy vision objects * update inits * style * re order classes in modeling clap * Use roberta tokenizer as the other weights are not open sourced * small cleaup * remove tokenization CLAP * processor tokenizr is roberta * update feature extraction doc * remove vclap from model zero shot * update f_min and f_max to frequency_xx * some changes - fix modeling keys - add `is_longer` in the forward pass - make fixup * make fixup * consistent behavior ebtween rand_crop and fusion * add numpy resize and bilinear and documentation * move resizing to image utils * clean feature extraction * import resize from correct file * resize in image transforms * update * style * style * nit * remove unused arguments form the feature extractor * style * few fixes + make fixup * oops * fix more tests * add zero shot audio classification pipeline * update zeroshot classification pipeline * fixup * fix copies * all CI tests pass * make fixup + fix docs * fix docs * fix docs * update tests pip;eline * update zero shot pipeline * update feature extraction clap * update tokenization auto * use nested simplify * update pipeline tests * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * split in two lines * fixes * refactor * clean up * add integration tests * update config docstring * style * update processor * fix processor test * fix feat extractor tests * update docs * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix readmes * fix tips * Update src/transformers/models/auto/configuration_auto.py * update doc and remove todo -> properly explained * fix idx and typo * typoe * cleanup config * cleanup tests, styles and doc * ignore docstyle on image transform * add conversion script * remove the `clap` indx in favor of `CLAP` * update __init * nits * Update src/transformers/pipelines/__init__.py * fix bug * clarifiy config * fix copy * fix init * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * fix model output * fix comment * make fixup * make fixup * rename to `Clap` * replace to `Clap` * replace to `Clap` * repo consistency * again repo-consistency * make fixup * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * add config * changes * update conversion * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove unused function * update based on code reviews * style * more comments * cleanup * clean up * style * apply suggestions * Empty commit * pipeline will be added in a different PR * update calls to audio utils functions * update pipeline init * style * style * styling again * use pad * fix repo-consistency * update utils and add doc for audio utils * clean up resize by using torch. update inits accordingly * style * CLap's tokenizer is RobertA * add audio utils to internal toctreee * update totctree * style * update documentation and normalize naming accross audio utils and feature extraction clap * style * clean up * update doc and typos * fix doctest * update modelin code, got rid of a lot of reshaping * style on added doc audio utils * update modeling clap * style * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * docstringvariables with CLAP * rename key * update modeling CLAP * update audio utils docstring * update processing clap * fix readmes * fix toctree * udpate configuration clap * fix init * make fixup * fix * fix * update naming * update * update checkpoint path * Apply suggestions from code review * Major refactoring * Update src/transformers/models/clap/configuration_clap.py * merge --------- Co-authored-by:
younesbelkada <younesbelkada@gmail.com> Co-authored-by:
Younes Belkada <49240599+younesbelkada@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
-
Sylvain Gugger authored
-
Alissa authored
* Add resources to OPT * Add additional resources for OPT * Remove -{" "} after <PipelineTag pipeline="question-answering" /> * Change bitsnbytes to bitsandbytes * Revert formatting * Revert automatic format changes * Remove - sign after <PipelineTag pipeline="question-answering" /> -
Stas Bekman authored
Update modeling_bloom.py
-
Connor Henderson authored
* refactor: Make direct_import util * edit direct import fn * add docstring * make import function specific to transformers only * edit doc string
-
Jonatas Grosman authored
fix bug in reshaping labels
-
dependabot[bot] authored
Bump werkzeug in /examples/research_projects/decision_transformer Bumps [werkzeug](https://github.com/pallets/werkzeug) from 2.0.3 to 2.2.3. - [Release notes](https://github.com/pallets/werkzeug/releases) - [Changelog](https://github.com/pallets/werkzeug/blob/main/CHANGES.rst) - [Commits](https://github.com/pallets/werkzeug/compare/2.0.3...2.2.3 ) --- updated-dependencies: - dependency-name: werkzeug dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Xiaoyang Chen authored
* Update document of WhisperDecoderLayer * Update modeling_mbart.py * Update doc with utils/check_copies.py --fix_and_overwrite * Update modeling_xlm_prophetnet.py
-
Jannis Vamvas authored
Move X-MOD models to facebook org
-
regisss authored
-
- 15 Feb, 2023 2 commits
-
-
Sylvain Gugger authored
-
Joao Gante authored
-