1. 13 May, 2021 2 commits
    • Kai Zhang's avatar
      Auto scale config for multi-node training · e87ed5f0
      Kai Zhang authored
      Summary:
      Pull Request resolved: https://github.com/facebookresearch/d2go/pull/62
      
      Lightning trainer set max step to cfg.SOLVER.MAX_ITER. However, this is the max iteration for all nodes, in multi-node training, we need to scale it down, as well as eval period and other configs.
      This diff calls `auto_scale_world_size` before passing the config to trainer.
      
      Reviewed By: wat3rBro
      
      Differential Revision: D28140877
      
      fbshipit-source-id: 2639ae58773a4ec2a0cc59dfefb2f5d9b1afe1a8
      e87ed5f0
    • Yanghan Wang's avatar
      remove adet's default config from base runner · f3d05021
      Yanghan Wang authored
      Reviewed By: zhanghang1989
      
      Differential Revision: D28346653
      
      fbshipit-source-id: d80a1f824b097c05029edb171739a4928e47e4d8
      f3d05021
  2. 12 May, 2021 1 commit
    • Luis Perez's avatar
      Synchronize PyTorchLightning/pytorch-lightning (revision 7b283e3c@master) to... · 0848c589
      Luis Perez authored
      Synchronize PyTorchLightning/pytorch-lightning (revision 7b283e3c@master) to github/third-party/PyTorchLightning/pytorch-lightning
      
      Summary:
      # Manual
       - remove fixme's in `model_checkpoint.py`, `parameter_monitor.py`, `test_quantization.py`, and `speed_monitor.py` now that `Trainer` is properly annotated.
      - update `test_quantization.py` to `trainer.train_loop.global_step` instead of `trainer.global_step` which is a read-only.
      - update `loop_callback.py` to read from `train_loop` for `batch_idx` (which is no longer available).
      
      # Automatic
      ### New commit log messages
        7b283e3c Bugfix/Multiple dataloaders (#7433)
        d7c44cc6 Docs: sync chlog 1.3.1 (#7478)
        fdf50a5e Mark certain Trainer APIs as protected (#7420)
        ad9118f0 remove trainer hidden state | sanity refactor [1 / n] (#7437)
        4a1134db Log epoch metrics before firing the `on_evaluation_end` hook (#7272)
        b65ae794 Automatically check `DataModule.has_{setup,teardown,prepare_data}` [2/2] (#7238)
        8660d8cf [pre-commit.ci] pre-commit autoupdate (#7475)
        f6fe715e Fix Sphinx argument deprecation (#7464)
      
      Reviewed By: shuyingsunshine21
      
      Differential Revision: D28353491
      
      fbshipit-source-id: 98b87d99e2f09b47b07270858fcbdb5d5299730b
      0848c589
  3. 10 May, 2021 2 commits
  4. 07 May, 2021 2 commits
  5. 06 May, 2021 3 commits
  6. 05 May, 2021 2 commits
    • Yanghan Wang's avatar
      force contiguous when calling `augment_model_with_bundled_inputs` · 04bbc81f
      Yanghan Wang authored
      Summary: `augment_model_with_bundled_inputs` can compress the tensor when values are constant, however it requires contiguous layout and `zero_like` can return non-contiguous ones
      
      Reviewed By: zhanghang1989
      
      Differential Revision: D28224987
      
      fbshipit-source-id: 32b13728ff8fadd53412dbf2d59c4b46e92af04a
      04bbc81f
    • Sam Tsai's avatar
      add enlarge bounging box manipulation · e1961ad4
      Sam Tsai authored
      Summary: Add a bounding manipulation tool to padding bounding box data.
      
      Reviewed By: newstzpz
      
      Differential Revision: D28082071
      
      fbshipit-source-id: f168cae48672c4fa5c4ec98697c57ed7833787ab
      e1961ad4
  7. 04 May, 2021 2 commits
    • Hang Zhang's avatar
      OSS build mask head using fbnet builder · 477ab964
      Hang Zhang authored
      Summary:
      [WIP] Will add pretrained weights and update model url & scores
      
      build mask head using fbnet builder and retrain weights
      
      Reviewed By: wat3rBro
      
      Differential Revision: D27992340
      
      fbshipit-source-id: a216a99954eb3784438d595cd09cbb19e70ec3c3
      477ab964
    • Yanghan Wang's avatar
      move some of `test_meta_arch_rcnn.py` to oss · e84d3414
      Yanghan Wang authored
      Reviewed By: newstzpz
      
      Differential Revision: D27747996
      
      fbshipit-source-id: 6ae3b89c3944098828e246e5a4a89209b8e171a1
      e84d3414
  8. 30 Apr, 2021 1 commit
    • Sam Tsai's avatar
      add keypoints metadata registry · 77ebe09f
      Sam Tsai authored
      Summary:
      1. Add a keypoint metadata registry for registering different keypoint metadata
      2. Add option to inject_coco_dataset for adding keypoint metadata
      
      Reviewed By: newstzpz
      
      Differential Revision: D27730541
      
      fbshipit-source-id: c6ba97f60664fce4dcbb0de80222df7490bc6d5d
      77ebe09f
  9. 29 Apr, 2021 2 commits
  10. 28 Apr, 2021 2 commits
    • Hang Zhang's avatar
      Patch for Quantizing PointRend model · 3e243c1a
      Hang Zhang authored
      Summary: PointRend mask doesn't work for quantization. Add a patch to disable it.
      
      Reviewed By: wat3rBro
      
      Differential Revision: D27800349
      
      fbshipit-source-id: ae0268ee78b000245ebdb2edbfc679a62c85a59a
      3e243c1a
    • Ananth Subramaniam's avatar
      Synchronize PyTorchLightning/pytorch-lightning (revision 7fe8d184@master) to... · a95c7983
      Ananth Subramaniam authored
      Synchronize PyTorchLightning/pytorch-lightning (revision 7fe8d184@master) to github/third-party/PyTorchLightning/pytorch-lightning
      
      Summary:
      ### New commit log messages
        7fe8d184 Do not `shuffle` in `LightningDataModule.from_datasets` for `IterableDataset` (#7053)
        bab72255 [fix] Add barriers before and after setup hook is run (#7202)
        f920ba29 [bugfix] Metric not logged properly in manual optimization (#7228)
        e147127c [feat] Add better support for predict + ddp 2/3 (#7215)
        ca6c87ff Add back `clip_gradients(model)` (#7231)
        3b36d81c Fixed `num_sanity_val_steps` affecting reproducibility of training data shuffling (#7014)
        5cf9afa1 Add fairscale install msg for Sharded Plugins (#7213)
        52a5cee0 Set smarter default for DDP sharded for performance optimization (#6937)
        dd5ec75e Deprecate save_function from model checkpoint callback (#7201)
        ac7d6a35 Fix `NeptuneLogger.log_text(step=None)` (#7194)
        6be0a859 Update teardown for TPU acc (#7211)
        bc3f08b0 [fix] Add barrier to accelerator's teardown (#6814)
        68eac4d9 Enforce Lightning module as source of truth for automatic optimization (#7130)
        44d775fc Update Error message for ProfileConnector (#7204)
        31fcd7d0 Deprecate write_predictions on the LightningModule (#7066)
        591b9cee make bug_report_model minimal (#7191)
        b3fe8366 Move metrics_to_scalars to a dedicated utilities file (#7180)
        f58865aa Properly set `LightningModule.device` after model replacement (#7188)
        8439aead Update FairScale on CI (#7017)
        92af3632 Fix `lr_finder` suggesting too high learning rates (#7076)
        d534e53e add missing predict docs (#7150)
      
      Reviewed By: kazhang
      
      Differential Revision: D28032962
      
      fbshipit-source-id: 18cd01e8ecc13fe25f0890ac0f4b20c3c3e1fed3
      a95c7983
  11. 27 Apr, 2021 1 commit
    • Jacob Szwejbka's avatar
      Remove methods_to_optimize from script · c04ef895
      Jacob Szwejbka authored
      Summary:
      Pull Request resolved: https://github.com/facebookresearch/d2go/pull/54
      
      This arg is being deprecated, and its use case was really only for modules that use functions besides forward for inference.  The new plan is just to optimize every function. Since this script was just created I'm hoping I can edit this without throwing lots of stuff out of wack.
      
      Reviewed By: wat3rBro
      
      Differential Revision: D27954176
      
      fbshipit-source-id: fbe178fcc0404e5d2524d8edb4052e2cd17f43ba
      c04ef895
  12. 23 Apr, 2021 3 commits
  13. 22 Apr, 2021 1 commit
  14. 21 Apr, 2021 4 commits
  15. 20 Apr, 2021 2 commits
  16. 19 Apr, 2021 2 commits
    • Yue (R) Zhao's avatar
      tensorboard add_graph · 625cc2fa
      Yue (R) Zhao authored
      Summary: Add the API to log graph in tensorboard
      
      Reviewed By: wat3rBro
      
      Differential Revision: D27855774
      
      fbshipit-source-id: 415c469c5de0c56fc828d1b95f4be697e0acac84
      625cc2fa
    • Peizhao Zhang's avatar
      Added hooks to report training progress to fblearner and keep alive. · bd6043ee
      Peizhao Zhang authored
      Summary:
      * Added a registry to register functions that could be used to register hooks for training.
        * TRAINER_HOOKS_REGISTRY: List of functions to add hooks for trainer, all functions in the registry will be called to add hooks
        * `func(hooks: List[HookBase]) -> None`
      
      Reviewed By: zhanghang1989
      
      Differential Revision: D27560806
      
      fbshipit-source-id: fcfa02623bfd08508b6083db2d318d08f7e3c0b8
      bd6043ee
  17. 17 Apr, 2021 2 commits
    • Kai Zhang's avatar
      Delegate to model's customization · aeb24a92
      Kai Zhang authored
      Summary: Delegate FX quantization callback's customization to model.
      
      Reviewed By: wat3rBro
      
      Differential Revision: D27669212
      
      fbshipit-source-id: 2715546cf03134896da6f95ecddaf8503ff95d0b
      aeb24a92
    • Kai Zhang's avatar
      E2E QAT Workflow on Lightning · 845d0b2c
      Kai Zhang authored
      Summary:
      As per title and sanity test E2E QAT workflow on Lightning Trainer.
      
      - add `post_training_opts`. This is required to use `all_steps_qat.json` with Lightning. We don't actually support the post_training_opts in this diff though - we leave it part of T83437359.
      - Update .yaml to specify the Quantize-able modules.
      - Update `lightning_train_net.py` to use the QuantizationAwareTraining callback.
      
      Reviewed By: kandluis
      
      Differential Revision: D26304879
      
      fbshipit-source-id: 948bef4817d385d8a0969e4990d7f17ecd6994b7
      845d0b2c
  18. 15 Apr, 2021 3 commits
  19. 14 Apr, 2021 2 commits
  20. 13 Apr, 2021 1 commit
    • Yanghan Wang's avatar
      support rebuild train loader when data has expired · 754a95fa
      Yanghan Wang authored
      Summary:
      - store expiration in meta data when loading dat
      
      - use before_train_hook to rebuild data loader when expiration condition is met.
      
      Reviewed By: zisting
      
      Differential Revision: D27683164
      
      fbshipit-source-id: e3e3c6c15eee7c02c7a1bfed5f4d4d0e67d61a4f
      754a95fa