- 25 Mar, 2021 1 commit
-
-
Amir Tahmasbi authored
* Added embeddings layer * Added layoutlm layers, main model, maskedlm and token classification classes * Added model classes to tf auto models * Added model to PT to TF conversion script * Added model to doc README * Added tests * Removed unused imports * Added layoutlm model, test, and doc for sequence classification, and fix imports in __init__.py * Made tests pass! * Fixed typos in imports and docs * Fixed a typo in embeddings layer * Removed imports * Fixed formatting issues, imports, tests * Added layoutlm layers, main model, maskedlm and token classification classes * Added model classes to tf auto models * Added model to PT to TF conversion script * Removed unused imports * Added layoutlm model, test, and doc for sequence classification, and fix imports in __init__.py * Made tests pass! * Fixed typos in imports and docs * Removed imports * Fixed small formatting issues * Removed duplicates import from main __init__.py * Chnaged deafult arg to true for adding pooling layer to tf layoutlm * Fixed formatting issues * Style * Added copied from to classes copied from bert * Fixed doc strings examples to work with layoutlm inputs * Removed PyTorch reference in doc strings example * Added integration tests * Cleaned up initialization file * Updated model checkpoint identifiers * Fixed imports Co-authored-by:
Amir Tahmasbi <amir@ehsai.ca> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr>
-
- 24 Mar, 2021 4 commits
-
-
Sidd Karamcheti authored
-
Sylvain Gugger authored
* Remove version warning in pretrained BART models * Put it at the base model
-
Lysandre Debut authored
* Removes overflowing bad word IDs * Raise warning
-
imzhengzx authored
the orignal code in line 246 is ``` tokenizer: Optional["PreTrainedTokenizerBase"] = None, ``` it should be ``` tokenizer: Optional[PreTrainedTokenizerBase] = None, ```
-
- 23 Mar, 2021 7 commits
-
-
Philipp Schmid authored
* rewrote is_sagemaker_model_parallel_available * added is_sagemaker_model_parallel_available to SageMakerTrainer * removed unnecessary mp_parameters as TrainingArguments * make style happy * added mp_parameters again to parse mp-specific args.
-
RafaelWO authored
-
Stas Bekman authored
* import refactor * fix the fallback
-
Marta Ma艣lankowska authored
-
Bhadresh Savani authored
-
Stas Bekman authored
-
Sylvain Gugger authored
-
- 22 Mar, 2021 5 commits
-
-
Patrick von Platen authored
* push * finish * finish * make fix copies * change name
-
Ruan Chaves authored
* Modify the _hp_search_setup method on the Trainer class to handle the wandb argument passed by Ray Tune to model config. * Reformat single quotes as double quotes.
-
Boris Dayma authored
* feat: ensure unique artifact id * feat: allow manual init * fix: simplify reinit logic * fix: no dropped value + immediate commits * fix: wandb use in sagemaker * docs: improve documenation and formatting * fix: typos * docs: improve formatting
-
Sidd Karamcheti authored
Add simple one character fix so that on_step_begin and on_step_end are called at the right times (#10839)
-
Sebastian Olsson authored
-
- 19 Mar, 2021 3 commits
-
-
Sylvain Gugger authored
* Initial script * Add script to properly sort imports in init. * Add to the CI * Update utils/custom_init_isort.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Separate scripts that change content from quality * Move class_mapping_update to style_checks Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Philipp Schmid authored
* add uuid.hext to user_agent * add log * changed order of it * renamed as session id * renamed variable * reverted naming of the const
-
Th茅o Matussi猫re authored
* fix backend tokenizer args override: key mismatch * no touching the docs * fix mpnet * add mpnet to test * fix test Co-authored-by:theo <theo@matussie.re>
-
- 18 Mar, 2021 7 commits
-
-
Sylvain Gugger authored
* Fix distributed evaluation * Use logger
-
Vimarsh Chaturvedi authored
* Added check to ensure model name passed to from_pretrained and model are the same * Added test to check from_pretrained throws assert error when passed an incompatiable model name * Modified assert in from_pretrained with f-strings. Modified test to ensure desired assert message is being generated * Added check to ensure config and model has model_type * Fix FlauBERT heads Co-authored-by: vimarsh chaturvedi <vimarsh chaturvedi> Co-authored-by:
Stas Bekman <stas@stason.org> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr>
-
Julien Chaumond authored
* do not gobble certain kinds of requests.ConnectionError * Apply review comments Co-authored-by:Lysandre <lysandre.debut@reseau.eseo.fr>
-
James Thomin authored
This commit fixes a bug in the LengthGroupSampler where if model_input_name is not set, the default value is None instead of "input_ids"
-
Mohamed El-Geish authored
* wav2vec2: support datasets other than LibriSpeech * Formatting run_asr.py to pass code quality test * bundled orthography options and added verbose logs * fixing a typo in timit fine-tuning script * update comment for clarity * resize_lm_head and load custom vocab from file * adding a max_duration_in_seconds filter * do not assign `duration_filter` lambda, use a def * log untransliterated text as well * fix base model for arabic * fix duration filter when target_sr is not set * drop duration_in_seconds when unneeded * script for wav2vec2-large-lv60-timit-asr * fix for "tha" in arabic corpus (huggingface#10581) * adding more options to work with common_voice * PR feedback (huggingface#10581) * small README change
-
Patrick von Platen authored
* Create modeling_flax_eletra with code copied from modeling_flax_bert * Add ElectraForMaskedLM and ElectraForPretraining * Add modeling test for Flax electra and fix naming and arg in Flax Electra model * Add documentation * Fix code style * Create modeling_flax_eletra with code copied from modeling_flax_bert * Add ElectraForMaskedLM and ElectraForPretraining * Add modeling test for Flax electra and fix naming and arg in Flax Electra model * Add documentation * Fix code style * Fix code quality * Adjust tol in assert_almost_equal due to very small difference between model output, ranging 0.0010 - 0.0016 * Remove redundant ElectraPooler * save intermediate * adapt * correct bert flax design * adapt roberta as well * finish roberta flax * finish * apply suggestions * apply suggestions Co-authored-by:Chris Nguyen <anhtu2687@gmail.com>
-
Funtowicz Morgan authored
Signed-off-by:Morgan Funtowicz <funtowiczmo@gmail.com>
-
- 17 Mar, 2021 6 commits
-
-
Mansi Mane authored
* Added debug prints * Added config * Added prints * Added prints * Added extra samples to SequentialDistributedSampler * Added extra samples to SequentialDistributedSampler Updated SequentialDistributedSampler call * Added deubg prints * Removed extra prints * Making predicitons and labels multiple of batchsize * updated number of microbatches * Removed extra prints * Made start_remainder similar to DistributedSamplerWithLoop * Minor spacing update * Added debug prints Added config Added prints Added prints * Added extra samples to SequentialDistributedSampler Updated SequentialDistributedSampler call Added extra samples to SequentialDistributedSampler Added deubg prints Removed extra prints Making predicitons and labels multiple of batchsize updated number of microbatches Removed extra prints Squashing redundant commits * Made start_remainder similar to DistributedSamplerWithLoop Minor spacing update Made start_remainder similar to DistributedSamplerWithLoop * Test and styling * Rename test Co-authored-by:Sylvain Gugger <sylvain.gugger@gmail.com>
-
Sylvain Gugger authored
* Apply black before checking copies * Fix for class methods * Deal with lonely brackets * Remove debug and add forward changes * Separate copies and fix test * Add black as a test dependency
-
Stas Bekman authored
-
Stas Bekman authored
* deepspeed checkpoint loading code plus tests * style * style
-
Stas Bekman authored
-
Sylvain Gugger authored
-
- 16 Mar, 2021 7 commits
-
-
Cheng Li authored
* pass hf optimizer and scheduler to deepspeed if not specified in ds config * pass hf optimizer and scheduler to deepspeed if not specified in ds config * update * make init_deepspeed support config dict * fix docstring formatting * clean up trainer's comments * add new tests * fix type * composit argparse doesn't work * style * add a new test, rename others * document new functionality * complete tests, add docs * style * correct level * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * add new methods to the doc * must tell DS we are using a non-native optimizer * add protection against cpu_offload + HF optimizer combo * fix the cli overrides * sync docs + tests * restore AdamW * better docs * need new version * no longer needed * remove outdate information * refactor duplicated code Co-authored-by:
Stas Bekman <stas@stason.org> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Lysandre Debut authored
* Patches full import failure when sentencepiece is not installed * Dummies :)
-
Lysandre Debut authored
* Patches the full import failure and adds a test * Add comment
-
Lysandre authored
-
Lysandre authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Add DistributedSamplerWithLoop * Fix typo * Test and small fix
-