- 09 Apr, 2021 1 commit
-
-
Ananth Subramaniam authored
Summary: Before: this test would assume only 2 checkpoints were stored: `last.ckpt`, and `FINAL_MODEL_CKPT` Now: this test asserts that at least these 2 checkpoints are stored. In case the config specifies `save_top_k=-1` for instance, we'd save more checkpoints, causing this test to fail Since this test is only loading the last and the final outputs, I'm changing the behavior to assert that these checkpoints must be saved and ignoring other checkpoint files that could be generated. Reviewed By: kazhang Differential Revision: D27671284 fbshipit-source-id: 0419fb46856d048e7b6eba3ff1dc65b7280a9a90
-
- 05 Apr, 2021 1 commit
-
-
Owen Wang authored
Summary: Prediction count evaluator needs to gather it's state before computing metrics, otherwise when parallelized across N GPUs, we only get metrics computed from 1/N of the dataset, increasing our eval signal's variance. Reviewed By: wat3rBro Differential Revision: D27416864 fbshipit-source-id: b2c5334cd5a38bebcd06c6ace1627a6b71645fdd
-
- 31 Mar, 2021 2 commits
-
-
Kai Zhang authored
Reviewed By: newstzpz Differential Revision: D27255960 fbshipit-source-id: 1699ff23d2bc610dffc0215a90a7c1c17e3783c3
-
Sam Tsai authored
Summary: Fixing unit test that was not listed due to rebase error. Reviewed By: newstzpz, wat3rBro Differential Revision: D27456322 fbshipit-source-id: 519c5c086adfb19104ed99234f4f476eb34a79bc
-
- 30 Mar, 2021 1 commit
-
-
Sam Tsai authored
Summary: Separate unit tests into individual folder based on functionality. Reviewed By: wat3rBro Differential Revision: D27132567 fbshipit-source-id: 9a8200be530ca14c7ef42191d59795b05b9800cc
-
- 24 Mar, 2021 2 commits
-
-
Kai Zhang authored
Summary: Evaluate the predictor generated by previous step. This diff modify the lightning_train_net to reuse the evaluation logic by adding a `predictor_path` param. This diff also makes Lightning training backend depends on `cfg.MODEL.DEVICE` so that in evaluate_predictor step, user could set backend by changing model device. This is useful for evaluating int8 quantized model. Reviewed By: newstzpz Differential Revision: D27150609 fbshipit-source-id: fb72da3e81db932c0fa479350150720143e09a3e
-
Kai Zhang authored
Summary: Given that the way to create D2 (https://github.com/facebookresearch/d2go/commit/465cdb842513eb910aa20fcedea1d2edd15dc7b7)go runner and Lightning task are different, get_class was introduced so that in application we could do: ``` if is Lightning: task_cls = get_class(classname) task = task_cls(cfg) else: runner = create_runner(classname) ``` It turns out that we could need to do that in many places: workflow, binaries. This diff revert `get_class` and return class in `create_runner` if the class is a Lightning module. Reviewed By: newstzpz Differential Revision: D26676595 fbshipit-source-id: c3ce2016d09fe073af4c2dd9f98eea4e59ca621b
-
- 20 Mar, 2021 1 commit
-
-
Yanghan Wang authored
Summary: Not d2go.tests is not a library for oss, move utils code to d2go.utils.testing Reviewed By: zhanghang1989 Differential Revision: D26706933 fbshipit-source-id: 85767b66bbb6c67db05e11823beb4840220b2aa3
-
- 18 Mar, 2021 1 commit
-
-
Ananth Subramaniam authored
Summary: `checkpoint_callback` is being phased out. Initially, it was a special way to configure checkpoints, but it makes more sense for those callbacks to be included in the general `callbacks` trainer argument. In 1.2.X, `checkpoint_callback` is expected to be a boolean value only. If `checkpoint_callback=False` **and** an instance of `ModelCheckpoint` is passed in the trainer's `callbacks` arguments, Lightning raises a [misconfiguration error](https://github.com/PyTorchLightning/pytorch-lightning/blob/2f6ce1ae7fff34d16d3707571f6a9a7b0fb0c50a/pytorch_lightning/trainer/connectors/callback_connector.py#L66-L70) Reviewed By: newstzpz Differential Revision: D27139315 fbshipit-source-id: 07ad5ea520583a2e46a9cb2a938f98968265c932
-
- 16 Mar, 2021 1 commit
-
-
Sam Tsai authored
Summary: Extend conversion to support ids beyond cocotext format where ids are strings. Reviewed By: newstzpz Differential Revision: D27018211 fbshipit-source-id: 7282fd4b9a7e9cd19323235ed1a3c3e7b33cb6b4
-
- 11 Mar, 2021 1 commit
-
-
Kai Zhang authored
Summary: Pull Request resolved: https://github.com/facebookresearch/d2go/pull/17 Use PyTorch Lightning checkpoint in the test. Reviewed By: zhanghang1989 Differential Revision: D26962697 fbshipit-source-id: abe635e374c3ada130243f0eaadff34204f04fa1
-
- 05 Mar, 2021 1 commit
-
-
Peizhao Zhang authored
Summary: migrated to iopath for d2go - rest. Reviewed By: zhanghang1989 Differential Revision: D26632160 fbshipit-source-id: 182fcd9ad493a483ab7f342f373a7be2de7d10ca
-
- 04 Mar, 2021 1 commit
-
-
Hang Zhang authored
Summary: Pull Request resolved: https://github.com/facebookresearch/d2go/pull/5 Reviewed By: wat3rBro Differential Revision: D26780956 Pulled By: zhanghang1989 fbshipit-source-id: 26af80bbdf6bcb6af4a8b5d27e655826b34db26a
-
- 03 Mar, 2021 3 commits
-
-
Kai Zhang authored
Summary: As titled. The OSS version only use PyTorch Lightning while internal version leverages some features(e.g. Manifold integration, every_n_step checkpointing). This diff splits train_net.main into smaller functions so that they could be shared across OSS and internal versions. Reviewed By: zhanghang1989 Differential Revision: D26752701 fbshipit-source-id: 7f68e2a81e78193e117517a0ff668ab14b76ea65
-
Kai Zhang authored
Summary: As titled. Make a copy of quantization callback to unblock D2go OSS. Reviewed By: zhanghang1989 Differential Revision: D26735525 fbshipit-source-id: 12b77f04cfa1361e856b26ea218a262da1fadd88
-
facebook-github-bot authored
fbshipit-source-id: f4a8ba78691d8cf46e003ef0bd2e95f170932778
-