- 09 Feb, 2021 5 commits
-
-
Julien Plu authored
-
abhishek thakur authored
* Fix example in Wav2Vec2 documentation * fix style
-
Lysandre authored
-
Patrick von Platen authored
* add wav2vec2CTC and deprecate for maskedlm * remove from docs
-
Lysandre authored
-
- 08 Feb, 2021 25 commits
-
-
sandip authored
-
Stas Bekman authored
-
demSd authored
-
Juan Cruz-Benito authored
Removing run_pl_glue.py from text classification docs, include run_xnli.py & run_tf_text_classification.py (#10066) * Removing run_pl_glue.py from seq classification docs * Adding run_tf_text_classification.py * Using :prefix_link: to refer local files * Applying "make style" to the branch * Update docs/source/task_summary.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Removing last underscores Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Lysandre authored
-
Stas Bekman authored
* deepspeed bug fixes and tests * manual wrap?
-
Anthony MOI authored
-
noise-field authored
* Unify logging with f-strings * Get limits from MLflow rather than hardcode * Add a check for parameter length overflow Also constants are marked as internal * Don't stop run in on_train_end This causes bad behaviour when there is a seprarte validation step: validation gets recorded as separate run. * Fix style
-
Olivier authored
* replace -100 token ids with the tokenizer pad_id for compute_metrics * fixed typo for label_ids
-
Lysandre Debut authored
-
demSd authored
* claiming this issue * Integration test for BertGeneration(Encoder and Decoder) * fix code quality
-
Julien Plu authored
* Fix template * Fix template
-
Patrick von Platen authored
-
Julien Plu authored
* Refacto BERT * Restore all the concerned models * Remove print * Update template * Apply Sylvain's and Morgan's comments * Fix cast * Put the cast inside call * Remove cond in ebds * Fix funnel * Restore previous dot product (attention_scores) computation * Add ConvBERT and BART * Make all the S2S models ONNX compliant * Fix test * Fix check copies
-
Julien Plu authored
* Disable temporarily too slow tests * Fix style * Fix template
-
Nicolas Patry authored
* Cleaning up `ConversationalPipeline` to support more than DialoGPT. Currently ConversationalPipeline was heavily biased towards DialoGPT ,which is the default model for this pipeline. This PR proposes changes to put back the modifications specific to DialoGPT into tokenizer-specific behavior wherever possible, by creating `_build_conversation_input_ids` function that takes conversation as input, and returns a list of ints corresponding to the tokens. It feels natural to put here because all models have probably different strategies to build input_ids from the full conversation and it's the tokenizer's job to transform strings into tokens (and vice-versa) If `_build_conversation_input_ids` is missing, previous behavior is used so we don't break anything so far (except for blenderbot where it's a fix). This PR also contains a fix for too long inputs. There used to be dead code for trying to limit the size of incoming input. The introduced fixed is that we limit within `_build_conversation_input_ids` to `tokenizer.model_max_length`. It corresponds to the intent of the removed dead code and is actually better because it corresponds to `model_max_length` which is different from `max_length` (which is a default parameter for `generate`). - Removed `history` logic from the Conversation as it's not relevant anymore because tokenization logic has been moved to tokenizer. And tokenizer cannot save any cache, and conversation cannot know what is relevant or not. Also it's not usable from `blenderbot` because the input_ids are not append only (EOS tokens is always at the end). - Added `iter_texts` method on `Conversation` because all the code was literred with some form of this iteration of past/generated_responses. * Removing torch mention in types. * Adding type checking to `_build_conversation_input_ids`. * Fixing import in strings.
-
Lysandre Debut authored
-
Patrick von Platen authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Lysandre Debut authored
* Correct cast to device * Comment back the slow test
-
sandip authored
-
Stas Bekman authored
-
Stas Bekman authored
-
- 05 Feb, 2021 7 commits
-
-
Stas Bekman authored
* make executable * make executable * same for the template * cleanup
-
Suraj Patil authored
* add prepare_decoder_input_ids_from_labels in s2s models * support lbl smoothing and enc/emb freezing * fix freezing * use pad_token_id from config * remove embed freezing and add warning * prepare decoder_input_ids inside DataCollatorForSeq2Seq
-
Patrick von Platen authored
* Bump minimum Jax requirement to 2.8.0 * update table
-
Patrick von Platen authored
* add big bird * change teacher to mentor * add proposal template * adapt template * delete old template * correct some links * finish template * create big bird from template * add big bird * improve boxes * finish boxes * add pointers for BigBird * finish big bird * up * up * up * up * apply lysandres and sylvains suggestions * delete bogus file * correct markdown * try different style * try different style * finalize
-
Lysandre Debut authored
* Clarify QA pipeline output based on character * Style
-
Lysandre authored
-
Lysandre authored
-
- 04 Feb, 2021 3 commits
-
-
Sylvain Gugger authored
* Update doc for pre-release * Use stable as default * Use the right commit :facepalms:
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-