1. 18 May, 2022 1 commit
  2. 24 Feb, 2022 1 commit
  3. 15 Oct, 2021 1 commit
  4. 14 Oct, 2021 1 commit
  5. 11 Jun, 2021 1 commit
    • Shilong Zhang's avatar
      Refactor the baseclass related to transformer (#978) · e05fb560
      Shilong Zhang authored
      
      
      * minor changes
      
      * change to modulist
      
      * change to Sequential
      
      * replace dropout with attn_drop and proj_drop in MultiheadAttention
      
      * add operation_name for attn
      
      * add drop path and move all ffn args to ffncfgs
      
      * fix typo
      
      * fix a bug when use default value of ffn_cfgs
      
      * fix ffns
      
      * add deprecate warning
      
      * fix deprecate warning
      
      * change to pop kwargs
      
      * support register FFN of transformer
      
      * support batch first
      
      * fix batch first wapper
      
      * fix forward wapper
      
      * fix typo
      
      * fix lint
      
      * add unitest for transformer
      
      * fix unitest
      
      * fix equal
      
      * use allclose
      
      * fix comments
      
      * fix comments
      
      * change configdict to dict
      
      * move drop to a file
      
      * add comments for drop path
      
      * add noqa 501
      
      * move bnc wapper to MultiheadAttention
      
      * move bnc wapper to MultiheadAttention
      
      * use dep warning
      
      * resolve comments
      
      * add unitest:
      
      * rename residual to identity
      
      * revert runner
      
      * msda residual to identity
      
      * rename inp_identity to identity
      
      * fix name
      
      * fix transformer
      
      * remove key in msda
      
      * remove assert for key
      Co-authored-by: default avatarHIT-cwh <2892770585@qq.com>
      Co-authored-by: default avatarbkhuang <congee524@gmail.com>
      Co-authored-by: default avatarWenwei Zhang <40779233+ZwwWayne@users.noreply.github.com>
      e05fb560
  6. 25 May, 2021 1 commit
  7. 23 Apr, 2021 1 commit
    • ZhangShilong's avatar
      [Feature]: support Multi-Scale-DeformAttention in deformable-detr (#878) · 54a7ebb4
      ZhangShilong authored
      * add c++ ms_deform_atten
      
      * fix cpp lint
      
      * fix cpp lint
      
      * clang format
      
      * remove cmakefile
      
      * google style
      
      * clang-format precommit
      
      * use clang-format-lint-action
      
      * add transformer base class
      
      * add merge
      
      * add docstr
      
      * add pyargs
      
      * fix according to commments
      
      * resiger module
      
      * change to use basemodule
      
      * add _ between build function
      
      * split the name
      
      * fix according to comments
      
      * fix lint and fix unitest
      
      * fix cpp lint
      
      * fix bug of deformdetr_atten
      
      * fix drop out
      
      * fix residual
      
      * use CUDA_1D_KERNEL_LOOP
      54a7ebb4