- 13 Sep, 2022 7 commits
-
-
Wang, Yi authored
only main_process will have HPO, and pass argument to other process Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com>
-
Rahul A R authored
* fixed bug which caused overwrite_cache to always be True (#18967). * reformatting changes
-
Ankur Goyal authored
Co-authored-by:Ankur Goyal <ankur@impira.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Partho authored
* added type hints * fixed typo
-
Joao Gante authored
-
Alara Dirik authored
* fix preprocessing for instance segmentation maps * add support for per-image instance2class_id mapping * edit docstrings for clarity
-
- 12 Sep, 2022 14 commits
-
-
Chris Emezue authored
Fix connected to [this issue](https://github.com/huggingface/transformers/issues/18944)
-
Tobias Nusser authored
Fixed typo itmes --> items
-
Joao Gante authored
-
Partho authored
-
Yih-Dar authored
* fix checkpoint name for wav2vec2 conformer Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Matt authored
* Use int64 throughout TFLongFormer * make style * Do some more fixed casting in TFLongFormer * Fix some wonky "is None" conditionals * Cast all the dtypes, salt the earth * Fix copies to TFLED as well and do some casting there * dtype fix in TFLongformer test * Make fixup * Expand tolerances on the LED tests too (I think this is a TF32 thing) * Expand test tolerances for LED a tiny bit (probably a Tensorfloat thing again)
-
Sylvain Gugger authored
* Align try_to_load_from_cache with huggingface_hub * Fix tests
-
Matt authored
* Update our TF 2.0 input format tip across all models * make style
-
Shijie Wu authored
-
Yih-Dar authored
* create Past CI results as tables for GitHub issue Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Wang, Yi authored
only main_process will have HPO, and pass argument to other process Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com>
-
Shijie Wu authored
-
- 10 Sep, 2022 3 commits
-
-
Sylvain Gugger authored
This reverts commit d8cf3b20.
-
Joao Gante authored
-
Joao Gante authored
-
- 09 Sep, 2022 8 commits
-
-
Bram Vanroy authored
* update black target version * add comment as per https://github.com/huggingface/transformers/pull/18955#issuecomment-1242081649 * revert change Will only update to 3.7 after black 2023 upgrade in January
-
Sylvain Gugger authored
-
Matt authored
* Fix train_step and test_step, correctly enable CLIP fit test * Stop using get_args on older Python versions * Don't use get_origin either * UnionType is actually even newer, don't use that either * Apply the same fix to test_loss_computation * Just realized I was accidentally skipping a bunch of tests! * Fix test_loss_computation for models without separable labels * Fix scalar losses in test_step and train_step * Stop committing your breakpoints * Fix Swin loss shape * Fix Tapas loss shape * Shape fixes for TAPAS, DeIT, HuBERT and ViTMAE * Add loss computation to TFMobileBertForPreTraining * make fixup and move copied from statement * make fixup and move copied from statement * Correct copied from * Add labels and next_sentence_label inputs to TFMobileBERT * Make sure total_loss is always defined * Update tests/test_modeling_tf_common.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Fix copied from * Ensure CTC models get labels in tests * Ensure CTC models get labels in tests * Fix tests for vit_mae * Fix tests for vit_mae * Fix tests for vit_mae * Reduce batch size for wav2vec2 testing because it was causing OOM * Skip some TAPAS tests that are failing * Skip a failing HuBERT test * make style * Fix mobilebertforpretraining test * Skip Wav2Vec2 tests that use huge amounts of mem * Skip keras_fit for Wav2Vec2 as well Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Ekagra Ranjan authored
-
Rafa艂 Jankowski authored
* NeptuneCallback improvements * After review suggestions and deduplication of initial run * Added volatile checkpoints support due to missing post-rebase commit * Update README per review comments - Remove list formatting - Correct Neptune docs link Co-authored-by:Sabine <sabine.nyholm@neptune.ai>
-
Sanchit Gandhi authored
* [JAX] Replace all jax.tree_* calls with jax.tree_util.tree_* * fix double tree_util
-
HuYong authored
* add_ernie * remove Tokenizer in ernie * polish code * format code style * polish code * fix style * update doc * make fix-copies * change model name * change model name * fix dependency * add more copied from * rename ErnieLMHeadModel to ErnieForCausalLM do not expose ErnieLayer update doc * fix * make style * polish code * polish code * fix * fix * fix * fix * fix * final fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Nima Boscarino authored
* Update TRANSLATING.md Update the contact to @GuggerSylvain * Update docs/TRANSLATING.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 08 Sep, 2022 2 commits
-
-
NielsRogge authored
* First draft * Improve conversion script * Make vision encoder work * More improvements * Improve conversion script * Fix quality * Add MultiframeIntegrationTransformer * More improvements * Make MiT output work * Fix quality * Add prompts generator * Add tests * Fix some tests * Fix some more tests * Fix more tests * Improve conversion script * Fix model outputs * Fix more tests * Add XClipProcessor * Use processor in conversion script * Fix integration test * Update README, fix docs * Fix all tests * Add MIT output to XClipOutput * Create better variable names * Rename XClip to XCLIP * Extend conversion script * Add support for large models * Add support for 16 frame models * Add another model' * Fix module issue * Apply suggestions from code review * Add figure to docs * Fix CLIPProcessor issue * Apply suggestions from code review * Delete file * Convert more checkpoints * Convert last checkpoint * Update nielsr to microsoft
-
Devlee247 authored
* fix LayoutXLM wrong link in README * fix LayoutXLM worng link in index.mdx
-
- 07 Sep, 2022 6 commits
-
-
Steven Liu authored
* skip some code examples for doctests * make style * fix code snippet formatting * separate code snippet into two blocks
-
lewtun authored
-
Colin Dean authored
* Starts on a list of external deps required for dev I've found that I need to install MeCab manually on my AS Mac. * Generalizes OS nascent dependency list Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yanming Wang authored
* Fix XLA fp16 and bf16 error checking * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Ankur Goyal authored
* [WIP] Skeleton of VisualQuestionAnweringPipeline extended to support LayoutLM-like models * Fixup * Use the full encoding * Basic refactoring to DocumentQuestionAnsweringPipeline * Cleanup * Improve args, docs, and implement preprocessing * Integrate OCR * Refactor question_answering pipeline * Use refactored QA code in the document qa pipeline * Fix tests * Some small cleanups * Use a string type annotation for Image.Image * Update encoding with image features * Wire through the basic docs * Handle invalid response * Handle empty word_boxes properly * Docstring fix * Integrate Donut model * Fixup * Incorporate comments * Address comments * Initial incorporation of tests * Address Comments * Change assert to ValueError * Comments * Wrap `score` in float to make it JSON serializable * Incorporate AutoModeLForDocumentQuestionAnswering changes * Fixup * Rename postprocess function * Fix auto import * Applying comments * Improve docs * Remove extra assets and add copyright * Address comments Co-authored-by:Ankur Goyal <ankur@impira.com>
-