1. 18 Nov, 2021 1 commit
  2. 16 Nov, 2021 1 commit
  3. 09 Nov, 2021 1 commit
    • Yih-Dar's avatar
      Add TFViTModel (#13778) · be4a6c64
      Yih-Dar authored
      
      
      * Start the work for TFViTModel
      
      * Convert to TF code - need to check in the follow up commits
      
      * Clean up model code
      
      * Expose TFViTModel
      
      * make style
      
      * make quality
      
      * Add test
      
      * make style & quality
      
      * Fix some imports
      
      * fix wrong usage - *kwargs => ** kwargs
      
      * Fix Conv2D weight loading (PT->TF) issue
      
      * Add tests for images with different sizes + fix model
      
      * Fix some common tests for TFViTModel
      
      * Use inputs instead of input_ids in test_compile_tf_model
      
      * Add a comment about transpose and Conv2D in convert_tf_weight_name_to_pt_weight_name
      
      * Avoid transpose in TFViT call
      
      * Fix Conv2D issue in load_tf2_weights_in_pytorch_model
      
      * Use tf.keras.layers.Conv2D instead of tf.nn.conv2d
      
      * Using simpler heuristic to detect Conv2D layer
      
      * Change convert_tf_weight_name_to_pt_weight_name to return TransposeType
      
      * Check tf_weight_shape is not None before using it
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * fix missing comma
      
      * fix input dtype
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      be4a6c64
  4. 08 Nov, 2021 1 commit
  5. 02 Nov, 2021 1 commit
  6. 01 Nov, 2021 1 commit
    • NielsRogge's avatar
      Add BeitForSemanticSegmentation (#14096) · e20faa6f
      NielsRogge authored
      
      
      * Add first draft
      
      * Make forward pass work
      
      * Improve conversion script
      
      * Add notebook that checks if it works
      
      * Add BeitForSemanticSegmentation to the tests
      
      * More improvements
      
      * Make BeitForSemanticSegmentation consistent with Segformer
      
      * Small bug fix
      
      * Add BeitForSemanticSegmentation to docs
      
      * Make sure model doesn't output hidden states when the user doesn't want to
      
      * Make it possible to convert the large model
      
      * Fix issue
      
      * Fix conversion script for large model
      
      * Add auxiliary_head option to semantic segmentation model
      
      * Apply suggestions from @sgugger's review
      
      * Apply suggestions from code review
      
      * Fix failing test
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      e20faa6f
  7. 29 Oct, 2021 1 commit
  8. 25 Oct, 2021 1 commit
  9. 21 Oct, 2021 1 commit
  10. 11 Oct, 2021 1 commit
  11. 05 Oct, 2021 1 commit
  12. 22 Sep, 2021 1 commit
  13. 20 Sep, 2021 1 commit
  14. 15 Sep, 2021 1 commit
  15. 31 Aug, 2021 2 commits
  16. 24 Aug, 2021 1 commit
  17. 15 Jul, 2021 1 commit
  18. 13 Jul, 2021 1 commit
  19. 01 Jul, 2021 1 commit
  20. 29 Jun, 2021 1 commit
  21. 24 Jun, 2021 1 commit
  22. 23 Jun, 2021 2 commits
  23. 14 Jun, 2021 1 commit
  24. 09 Jun, 2021 1 commit
    • NielsRogge's avatar
      Add DETR (#11653) · d3eacbb8
      NielsRogge authored
      
      
      * Squash all commits of modeling_detr_v7 branch into one
      
      * Improve docs
      
      * Fix tests
      
      * Style
      
      * Improve docs some more and fix most tests
      
      * Fix slow tests of ViT, DeiT and DETR
      
      * Improve replacement of batch norm
      
      * Restructure timm backbone forward
      
      * Make DetrForSegmentation support any timm backbone
      
      * Fix name of output
      
      * Address most comments by @LysandreJik
      
      * Give better names for variables
      
      * Conditional imports + timm in setup.py
      
      * Address additional comments by @sgugger
      
      * Make style, add require_timm and require_vision to tests茅
      
      * Remove train_backbone attribute of DetrConfig, add methods to freeze/unfreeze backbone
      
      * Add png files to fixtures
      
      * Fix type hint
      
      * Add timm to workflows
      
      * Add `BatchNorm2d` to the weight initialization
      
      * Fix retain_grad test
      
      * Replace model checkpoints by Facebook namespace
      
      * Fix name of checkpoint in test
      
      * Add user-friendly message when scipy is not available
      
      * Address most comments by @patrickvonplaten
      
      * Remove return_intermediate_layers attribute of DetrConfig and simplify Joiner
      
      * Better initialization
      
      * Scipy is necessary to get sklearn metrics
      
      * Rename TimmBackbone to DetrTimmConvEncoder and rename DetrJoiner to DetrConvModel
      
      * Make style
      
      * Improve docs and add 2 community notebooks
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      d3eacbb8
  25. 25 May, 2021 1 commit
  26. 20 May, 2021 2 commits
  27. 14 May, 2021 1 commit
  28. 13 May, 2021 1 commit
  29. 11 May, 2021 1 commit
  30. 07 May, 2021 1 commit
    • Vasudev Gupta's avatar
      Add BigBirdPegasus (#10991) · dc3f6758
      Vasudev Gupta authored
      
      
      * init bigbird pegasus
      
      * add debugging nb ; update config
      
      * init conversion
      
      * update conversion script
      
      * complete conversion script
      
      * init forward()
      
      * complete forward()
      
      * add tokenizer
      
      * add some slow tests
      
      * commit current
      
      * fix copies
      
      * add docs
      
      * add conversion script for bigbird-roberta-summarization
      
      * remove TODO
      
      * small fixups
      
      * correct tokenizer
      
      * add bigbird core for now
      
      * fix config
      
      * fix more
      
      * revert pegasus-tokenizer back
      
      * make style
      
      * everything working for pubmed; yayygit status
      
      * complete tests finally
      
      * remove bigbird pegasus tok
      
      * correct tokenizer
      
      * correct tests
      
      * add tokenizer files
      
      * finish make style
      
      * fix test
      
      * update
      
      * make style
      
      * fix tok utils base file
      
      * make fix-copies
      
      * clean a bit
      
      * small update
      
      * fix some suggestions
      
      * add to readme
      
      * fix a bit, clean tests
      
      * fix more tests
      
      * Update src/transformers/__init__.py
      
      * Update src/transformers/__init__.py
      
      * make fix-copies
      
      * complete attn switching, auto-padding left
      
      * make style
      
      * fix auto-padding test
      
      * make style
      
      * fix batched attention tests
      
      * put tolerance at 1e-1 for stand-alone decoder test
      
      * fix docs
      
      * fix tests
      
      * correct slow tokenizer conversion
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * complete remaining suggestions
      
      * fix test
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      dc3f6758
  31. 05 May, 2021 1 commit
  32. 04 May, 2021 1 commit
  33. 29 Apr, 2021 1 commit
  34. 26 Apr, 2021 1 commit
  35. 23 Apr, 2021 2 commits
    • Daniel Stancl's avatar
      Fix cross-attention head mask for Torch encoder-decoder models (#10605) · e3ff165a
      Daniel Stancl authored
      * Fix cross-attention head mask for Torch BART models
      
      * Fix head masking for cross-attention module for the following
      models: BART, Blenderbot, Blenderbot_small, M2M_100, Marian, MBart,
      Pegasus
      
      * Enable test_headmasking for M2M_100 model
      
      * Fix cross_head_mask for FSMT, LED and T5
      
      * This commit fixes `head_mask` for cross-attention modules
      in the following models: FSMT, LED, T5
      
      * It also contains some smaller changes in doc so that
      it is be perfectly clear the shape of `cross_head_mask`
      is the same as of `decoder_head_mask`
      
      * Update template
      
      * Fix template for BartForCausalLM
      
      * Fix cross_head_mask for Speech2Text models
      
      * Fix cross_head_mask in templates
      
      * Fix args order in BartForCausalLM template
      
      * Fix doc in BART templates
      
      * Make more explicit naming
      
      * `cross_head_mask` -> `cross_attn_head_mask`
      
      * `cross_layer_head_mask` -> `cross_attn_layer_head_mask`
      
      * Fix doc
      
      * make style quality
      
      * Fix speech2text docstring
      e3ff165a
    • Sylvain Gugger's avatar
      Trainer push to hub (#11328) · bf2e0cf7
      Sylvain Gugger authored
      
      
      * Initial support for upload to hub
      
      * push -> upload
      
      * Fixes + examples
      
      * Fix torchhub test
      
      * Torchhub test I hate you
      
      * push_model_to_hub -> push_to_hub
      
      * Apply mixin to other pretrained models
      
      * Remove ABC inheritance
      
      * Add tests
      
      * Typo
      
      * Run tests
      
      * Install git-lfs
      
      * Change approach
      
      * Add push_to_hub to all
      
      * Staging test suite
      
      * Typo
      
      * Maybe like this?
      
      * More deps
      
      * Cache
      
      * Adapt name
      
      * Quality
      
      * MOAR tests
      
      * Put it in testing_utils
      
      * Docs + torchhub last hope
      
      * Styling
      
      * Wrong method
      
      * Typos
      
      * Update src/transformers/file_utils.py
      Co-authored-by: default avatarJulien Chaumond <julien@huggingface.co>
      
      * Address review comments
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarJulien Chaumond <julien@huggingface.co>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      bf2e0cf7
  36. 13 Apr, 2021 1 commit