"...pipelines/animatediff/pipeline_animatediff_sparsectrl.py" did not exist on "7956c36aaa2b43584f16f7e755e46aa2d7efa721"
  1. 28 Apr, 2021 1 commit
    • Min Xu's avatar
      [feat] save memory by using bucket buffer only in backward (#633) · a5594032
      Min Xu authored
      
      
      * [feat] save memory by using bucket buffer only in backward
      
      - this fixes bug #627
      - added documentation to clarify the buffer's cost and speed/memory
        tradeoff
      - added setup/teardown calls so that the buffer is only allocated
        during the backward pass, saving more memory for forward and stepping
        so that they can be used for things like activations.
      - added a unit test that assert the memory is in range.
      
      Comparing with DDP:
      
        1. buffer size scales with # of FSDP not model size
        2. buffer is only allocated during backward
        3. buffer is used for small tensors only to reduce overhead
        4. overlapping of compute-reduction is very different
      
      * add PR number to changelog
      
      * filled in with memory number on 1.9
      
      * addressed comments
      
      * update comments
      
      * fix for 1.6
      
      * add a todo
      Co-authored-by: default avatarMin Xu <min.xu@acm.org>
      a5594032
  2. 26 Apr, 2021 1 commit
  3. 23 Apr, 2021 1 commit
    • shuyingsunshine21's avatar
      [FSDP] relax checking root condition (#620) · d3b86d65
      shuyingsunshine21 authored
      * relax checking root condition
      
      * formatting
      
      * add unittest
      
      * add unittest to ci test list
      
      * isort for import of unittest
      
      * format black .
      
      * move test to list 1
      
      * add skip no cuda
      
      * black and isort
      d3b86d65
  4. 22 Apr, 2021 2 commits
  5. 19 Apr, 2021 1 commit
    • Min Xu's avatar
      FSDP: fixing training with freezing weights (#614) · 24da3b11
      Min Xu authored
      
      
      * FSDP: fixing training with freezing weights
      
      - an assert is changed to catch this case correctly
      - unit test added (based on Quentin's test code) for this case and
        compare DDP and FSDP
      
      fixes: #610
      
      * added test file to list 1
      
      * Use better and simpler code as suggested by Myle
      
      * testing both methods of freezing as well
      Co-authored-by: default avatarMin Xu <min.xu@acm.org>
      24da3b11
  6. 13 Apr, 2021 3 commits
  7. 08 Apr, 2021 1 commit
  8. 07 Apr, 2021 2 commits
  9. 06 Apr, 2021 1 commit
  10. 04 Apr, 2021 2 commits
  11. 02 Apr, 2021 1 commit
  12. 01 Apr, 2021 1 commit
  13. 31 Mar, 2021 2 commits
    • Min Xu's avatar
      [fix] FSDP: disable single rank process group for auto_wrap_bn and fixed mixed... · a0458b98
      Min Xu authored
      [fix] FSDP: disable single rank process group for auto_wrap_bn and fixed mixed precision regnet test (#556)
      
      * [fix] disable single rank process group for auto_wrap_bn
      
      - beefed up unit test with regnet-like model
      - found that single-rank process group is causing problem
      - disabled it to enable convergence tests on the vissl side
      - use `raise e from None` to get a better assertion output
        in testing.py.
      
      * [test] fix regnet test for ddp+mixed_precision
      
      - need AMP context in FSDP
      - workaround different between ddp & fsdp when bias=True
      - fixed a bug in input data generation that caused different ranks have
        the same data with wrong iteration count.
      - added TODO for need a better loss and grad_scaler and reduced
        iters so there is no nan.
      - added a (disabled) debugging code
      
      * lint
      
      * lint
      
      * add scaler
      
      * lint
      
      * scaler
      
      * add a real loss
      
      * seeding in the ranks
      
      * blance tests
      
      * run AMP DDP==FSDP test only on cuda version 11 and up
      
      * add relu inplace and comment
      
      * make wrap_bn covers more cases in full precision mode
      a0458b98
    • msbaines's avatar
      acb9ef00
  14. 30 Mar, 2021 1 commit
  15. 26 Mar, 2021 1 commit
  16. 25 Mar, 2021 2 commits
  17. 22 Mar, 2021 1 commit
  18. 20 Mar, 2021 1 commit
  19. 19 Mar, 2021 2 commits
  20. 18 Mar, 2021 4 commits
  21. 17 Mar, 2021 1 commit
  22. 12 Mar, 2021 2 commits
  23. 11 Mar, 2021 1 commit
  24. 09 Mar, 2021 2 commits
  25. 08 Mar, 2021 1 commit
    • Min Xu's avatar
      [fix]: handle inputs with containers in mixed precision (#486) · 2e9a14e7
      Min Xu authored
      * [fix]: handle inputs with containers
      
      - this is an issue surfaces by vissl as well
      - fix seems to be super simple
      - also cleaned up two tests with respect to multiple such tests
        running back to back (they don't do that presently)
      
      * cleanup
      
      * fix
      
      * lint
      2e9a14e7
  26. 06 Mar, 2021 1 commit
  27. 05 Mar, 2021 1 commit
    • Min Xu's avatar
      [refactor] enhance wrap and auto_wrap (#467) · a05a79bc
      Min Xu authored
      
      
      * [refactor] enhance wrap and auto_wrap
      
      - Two things were done in this PR
        1. We don't need to import FSDP in wrap.py since the wrapper class
           type is stored in the context now.
        2. We can use a `auto_wrap_policy` function to customize wrapping policy
           for auto_wrap, including size of module, blacklist, exclude list
      - The auto_wrap function got simplified a bit as a minor side effect.
      
      * Update fairscale/nn/wrap/auto_wrap.py
      Co-authored-by: default avatarSean Naren <sean@grid.ai>
      
      * addressed comments
      
      * addressed more comments
      Co-authored-by: default avatarSean Naren <sean@grid.ai>
      a05a79bc