"vscode:/vscode.git/clone" did not exist on "14b50ed01985e0431308674d8fd143d223032ee2"
  1. 04 Aug, 2020 1 commit
  2. 03 Aug, 2020 1 commit
    • Julien Plu's avatar
      Fix saved model creation (#5468) · 9996f697
      Julien Plu authored
      * Fix TF Serving when output_hidden_states and output_attentions are True
      
      * Add tests for saved model creation + bug fix for multiple choices models
      
      * remove unused import
      
      * Fix the input for several layers
      
      * Fix test
      
      * Fix conflict printing
      
      * Apply style
      
      * Fix XLM and Flaubert for TensorFlow
      
      * Apply style
      
      * Fix TF check version
      
      * Apply style
      
      * Trigger CI
      9996f697
  3. 31 Jul, 2020 3 commits
  4. 30 Jul, 2020 3 commits
    • Stas Bekman's avatar
      typos (#6162) · a2f6d521
      Stas Bekman authored
      * 2 small typos
      
      * more typos
      
      * correct path
      a2f6d521
    • guillaume-be's avatar
      Addition of a DialoguePipeline (#5516) · e642c789
      guillaume-be authored
      
      
      * initial commit for pipeline implementation
      
      Addition of input processing and history concatenation
      
      * Conversation pipeline tested and working for single & multiple conversation inputs
      
      * Added docstrings for dialogue pipeline
      
      * Addition of dialogue pipeline integration tests
      
      * Delete test_t5.py
      
      * Fixed max code length
      
      * Updated styling
      
      * Fixed test broken by formatting tools
      
      * Removed unused import
      
      * Added unit test for DialoguePipeline
      
      * Fixed Tensorflow compatibility
      
      * Fixed multi-framework support using framework flag
      
      * - Fixed docstring
      - Added `min_length_for_response` as an initialization parameter
      - Renamed `*args` to `conversations`, `conversations` being a `Conversation` or a `List[Conversation]`
      - Updated truncation to truncate entire segments of conversations, instead of cutting in the middle of a user/bot input
      
      * - renamed pipeline name from dialogue to conversational
      - removed hardcoded default value of 1000 and use config.max_length instead
      - added `append_response` and `set_history` method to the Conversation class to avoid direct fields mutation
      - fixed bug in history truncation method
      
      * - Updated ConversationalPipeline to accept only active conversations (otherwise a ValueError is raised)
      
      * - Simplified input tensor conversion
      
      * - Updated attention_mask value for Tensorflow compatibility
      
      * - Updated last dialogue reference to conversational & fixed integration tests
      
      * Fixed conflict with master
      
      * Updates following review comments
      
      * Updated formatting
      
      * Added Conversation and ConversationalPipeline to the library __init__, addition of docstrings for Conversation, added both to the docs
      
      * Update src/transformers/pipelines.py
      
      Updated docsting following review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      e642c789
    • Sylvain Gugger's avatar
      Switch from return_tuple to return_dict (#6138) · 91cb9546
      Sylvain Gugger authored
      
      
      * Switch from return_tuple to return_dict
      
      * Fix test
      
      * [WIP] Test TF Flaubert + Add {XLM, Flaubert}{TokenClassification, MultipleC… (#5614)
      
      * Test TF Flaubert + Add {XLM, Flaubert}{TokenClassification, MultipleChoice} models and tests
      
      * AutoModels
      
      
      Tiny tweaks
      
      * Style
      
      * Final changes before merge
      
      * Re-order for simpler review
      
      * Final fixes
      
      * Addressing @sgugger's comments
      
      * Test MultipleChoice
      
      * Rework TF trainer (#6038)
      
      * Fully rework training/prediction loops
      
      * fix method name
      
      * Fix variable name
      
      * Fix property name
      
      * Fix scope
      
      * Fix method name
      
      * Fix tuple index
      
      * Fix tuple index
      
      * Fix indentation
      
      * Fix variable name
      
      * fix eval before log
      
      * Add drop remainder for test dataset
      
      * Fix step number + fix logging datetime
      
      * fix eval loss value
      
      * use global step instead of step + fix logging at step 0
      
      * Fix logging datetime
      
      * Fix global_step usage
      
      * Fix breaking loop + logging datetime
      
      * Fix step in prediction loop
      
      * Fix step breaking
      
      * Fix train/test loops
      
      * Force TF at least 2.2 for the trainer
      
      * Use assert_cardinality to facilitate the dataset size computation
      
      * Log steps per epoch
      
      * Make tfds compliant with TPU
      
      * Make tfds compliant with TPU
      
      * Use TF dataset enumerate instead of the Python one
      
      * revert previous commit
      
      * Fix data_dir
      
      * Apply style
      
      * rebase on master
      
      * Address Sylvain's comments
      
      * Address Sylvain's and Lysandre comments
      
      * Trigger CI
      
      * Remove unused import
      
      * Switch from return_tuple to return_dict
      
      * Fix test
      
      * Add recent model
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      Co-authored-by: default avatarJulien Plu <plu.julien@gmail.com>
      91cb9546
  5. 29 Jul, 2020 2 commits
  6. 28 Jul, 2020 3 commits
  7. 27 Jul, 2020 1 commit
    • Joe Davison's avatar
      Zero shot classification pipeline (#5760) · 3deffc1d
      Joe Davison authored
      * add initial zero-shot pipeline
      
      * change default args
      
      * update default template
      
      * add label string splitting
      
      * add str labels support, remove nli from name
      
      * style
      
      * add input validation and working tf defaults
      
      * tests
      
      * quality check
      
      * add docstring to __call__
      
      * add slow tests
      
      * Change truncation to only_first
      
      also lower precision on tests for readibility
      
      * style
      3deffc1d
  8. 23 Jul, 2020 2 commits
  9. 20 Jul, 2020 2 commits
    • Stas Bekman's avatar
      DataParallel fixes (#5733) · 35cb101e
      Stas Bekman authored
      * DataParallel fixes:
      
      1. switched to a more precise check
      -        if self.args.n_gpu > 1:
      +        if isinstance(model, nn.DataParallel):
      
      2. fix tests - require the same fixup under DataParallel as the training module
      
      * another fix
      35cb101e
    • Pradhy729's avatar
      Trainer support for iterabledataset (#5834) · 290b6e18
      Pradhy729 authored
      * Don't pass sampler for iterable dataset
      
      * Added check for test and eval dataloaders.
      
      * Formatting
      
      * Don't pass sampler for iterable dataset
      
      * Added check for test and eval dataloaders.
      
      * Formatting
      
      * Cleaner if nesting.
      
      * Added test for trainer and iterable dataset
      
      * Formatting for test
      
      * Fixed import when torch is available only.
      
      * Added require torch decorator to helper class
      
      * Moved dataset class inside unittest
      
      * Removed nested if and changed model in test
      
      * Checking torch availability for IterableDataset
      290b6e18
  10. 18 Jul, 2020 3 commits
    • Teven's avatar
      Xlnet outputs (#5883) · 4b506a37
      Teven authored
      Slightly breaking change, changes functionality for `use_cache` in XLNet: if use_cache is True and mem_len is 0 or None (which is the case in the base model config), the model behaves like GPT-2 and returns mems to be used as past in generation. At training time `use_cache` is overriden and always True.
      4b506a37
    • Teven's avatar
      Revert "Xlnet outputs (#5881)" (#5882) · a5580924
      Teven authored
      This reverts commit 13be4872.
      a5580924
    • Teven's avatar
      Xlnet outputs (#5881) · 13be4872
      Teven authored
      Slightly breaking change, changes functionality for `use_cache` in XLNet: if use_cache is True and mem_len is 0 or None (which is the case in the base model config), the model behaves like GPT-2 and returns mems to be used as past in generation. At training time `use_cache` is overriden and always True.
      13be4872
  11. 17 Jul, 2020 3 commits
    • Teven's avatar
      Revert "XLNet `use_cache` refactor (#5770)" (#5854) · 615be03f
      Teven authored
      This reverts commit 0b2da0e5.
      615be03f
    • Teven's avatar
      XLNet `use_cache` refactor (#5770) · 0b2da0e5
      Teven authored
      Slightly breaking change, changes functionality for `use_cache` in XLNet: if use_cache is True and mem_len is 0 or None (which is the case in the base model config), the model behaves like GPT-2 and returns mems to be used as past in generation. At training time `use_cache` is overriden and always True.
      0b2da0e5
    • Patrick von Platen's avatar
      [Reformer] - Cache hidden states and buckets to speed up inference (#5578) · 9d37c56b
      Patrick von Platen authored
      * fix merge rebase
      
      * add intermediate reformer code
      
      * save intermediate caching results
      
      * save intermediate
      
      * save intermediate results
      
      * save intermediate
      
      * upload next step
      
      * fix generate tests
      
      * make tests work
      
      * add named tuple output
      
      * Apply suggestions from code review
      
      * fix use_cache for False case
      
      * fix tensor to gpu
      
      * fix tensor to gpu
      
      * refactor
      
      * refactor and make style
      9d37c56b
  12. 16 Jul, 2020 1 commit
  13. 15 Jul, 2020 3 commits
  14. 14 Jul, 2020 2 commits
    • Sam Shleifer's avatar
    • as-stevens's avatar
      [Reformer classification head] Implement the reformer model classification... · f867000f
      as-stevens authored
      
      [Reformer classification head] Implement the reformer model classification head for text classification (#5198)
      
      * Reformer model head classification implementation for text classification
      
      * Reformat the reformer model classification code
      
      * PR review comments, and test case implementation for reformer for classification head changes
      
      * CI/CD reformer for classification head test import error fix
      
      * CI/CD test case implementation  added ReformerForSequenceClassification to all_model_classes
      
      * Code formatting- fixed
      
      * Normal test cases added for reformer classification head
      
      * Fix test cases implementation for the reformer classification head
      
      * removed token_type_id parameter from the reformer classification head
      
      * fixed the test case for reformer classification head
      
      * merge conflict with master fixed
      
      * merge conflict, changed reformer classification to accept the choice_label parameter added in latest code
      
      * refactored the the reformer classification head test code
      
      * reformer classification head, common transform test cases fixed
      
      * final set of the review comment, rearranging the reformer classes and docstring add to classification forward method
      
      * fixed the compilation error and text case fix for reformer classification head
      
      * Apply suggestions from code review
      
      Remove unnecessary dup
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      f867000f
  15. 13 Jul, 2020 2 commits
  16. 10 Jul, 2020 1 commit
    • Sylvain Gugger's avatar
      Change model outputs types to self-document outputs (#5438) · edfd82f5
      Sylvain Gugger authored
      * [WIP] Proposal for model outputs
      
      * All Bert models
      
      * Make CI green maybe?
      
      * Fix ONNX test
      
      * Isolate ModelOutput from pt and tf
      
      * Formatting
      
      * Add Electra models
      
      * Auto-generate docstrings from outputs
      
      * Add TF outputs
      
      * Add some BERT models
      
      * Revert TF side
      
      * Remove last traces of TF changes
      
      * Fail with a clear error message
      
      * Add Albert and work through Bart
      
      * Add CTRL and DistilBert
      
      * Formatting
      
      * Progress on Bart
      
      * Renames and finish Bart
      
      * Formatting
      
      * Fix last test
      
      * Add DPR
      
      * Finish Electra and add FlauBERT
      
      * Add GPT2
      
      * Add Longformer
      
      * Add MMBT
      
      * Add MobileBert
      
      * Add GPT
      
      * Formatting
      
      * Add Reformer
      
      * Add Roberta
      
      * Add T5
      
      * Add Transformer XL
      
      * Fix test
      
      * Add XLM + fix XLMForTokenClassification
      
      * Style + XLMRoberta
      
      * Add XLNet
      
      * Formatting
      
      * Add doc of return_tuple arg
      edfd82f5
  17. 08 Jul, 2020 2 commits
    • Lorenzo Ampil's avatar
      Fix Inconsistent NER Grouping (Pipeline) (#4987) · 0cc4eae0
      Lorenzo Ampil authored
      
      
      * Add B I handling to grouping
      
      * Add fix to include separate entity as last token
      
      * move last_idx definition outside loop
      
      * Use first entity in entity group as reference for entity type
      
      * Add test cases
      
      * Take out extra class accidentally added
      
      * Return tf ner grouped test to original
      
      * Take out redundant last entity
      
      * Get last_idx safely
      Co-authored-by: default avatarColleterVi <36503688+ColleterVi@users.noreply.github.com>
      
      * Fix first entity comment
      
      * Create separate functions for group_sub_entities and group_entities (splitting call method to testable functions)
      
      * Take out unnecessary last_idx
      
      * Remove additional forward pass test
      
      * Move token classification basic tests to separate class
      
      * Move token classification basic tests back to monocolumninputtestcase
      
      * Move base ner tests to nerpipelinetests
      
      * Take out unused kwargs
      
      * Add back mandatory_keys argument
      
      * Add unitary tests for group_entities in _test_ner_pipeline
      
      * Fix last entity handling
      
      * Fix grouping fucntion used
      
      * Add typing to group_sub_entities and group_entities
      Co-authored-by: default avatarColleterVi <36503688+ColleterVi@users.noreply.github.com>
      0cc4eae0
    • Patrick von Platen's avatar
      [Benchmark] Add benchmarks for TF Training (#5594) · f82a2a5e
      Patrick von Platen authored
      * tf_train
      
      * adapt timing for tpu
      
      * fix timing
      
      * fix timing
      
      * fix timing
      
      * fix timing
      
      * update notebook
      
      * add tests
      f82a2a5e
  18. 07 Jul, 2020 5 commits