- 28 Oct, 2021 1 commit
-
-
Jirka Borovec authored
Co-authored-by:Nicolas Hug <nicolashug@fb.com>
-
- 25 Oct, 2021 1 commit
-
-
Francisco Massa authored
Summary: We would like to track the internal usage of TorchVision components. We're tracking [datasets usage](https://fburl.com/daiquery/dqpmemn3 ) now. This diff expand the tracking to all models. Reviewed By: fmassa Differential Revision: D31441632 fbshipit-source-id: e26072e582ac9f832c2056307ebf0eccf2ed6c9c Co-authored-by:
Kai Zhang <kaizh@fb.com>
-
- 04 Oct, 2021 1 commit
-
-
Philip Meier authored
* add ufmt as code formatter * cleanup * quote ufmt requirement * split imports into more groups * regenerate circleci config * fix CI * clarify local testing utils section * use ufmt pre-commit hook * split relative imports into local category * Revert "split relative imports into local category" This reverts commit f2e224cde2008c56c9347c1f69746d39065cdd51. * pin black and usort dependencies * fix local test utils detection * fix ufmt rev * add reference utils to local category * fix usort config * remove custom categories sorting * Run pre-commit without fixing flake8 * got a double import in merge Co-authored-by:Nicolas Hug <nicolashug@fb.com>
-
- 22 Jun, 2021 1 commit
-
-
Nicolas Hug authored
-
- 30 Mar, 2021 1 commit
-
-
Nicolas Hug authored
* Update URLS of detection models * Empty commit after setting read permission on S3 Co-authored-by:Francisco Massa <fvsmassa@gmail.com>
-
- 16 Nov, 2020 1 commit
-
-
Milos authored
* Fix MNASNet docstrings so they are rendered correctly * Add dot after url link in models docstrings for consistency
-
- 23 Oct, 2020 1 commit
-
-
F-G Fernandez authored
* style: Added annotation typing for resnet * fix: Fixed annotation to pass classes * fix: Fixed annotation typing * fix: Fixed annotation typing * fix: Fixed annotation typing for resnet * refactor: Removed un-necessary import * fix: Fixed constructor typing * style: Added black formatting on _resnet
-
- 16 Mar, 2020 1 commit
-
-
Charles Pao authored
Co-authored-by:Charles Pao <dirtybluer@gmail.com>
-
- 11 Mar, 2020 1 commit
-
-
Lutz Roeder authored
-
- 30 Nov, 2019 1 commit
-
-
driazati authored
* Add tests for results in script vs eager mode This copies some logic from `test_jit.py` to check that a TorchScript'ed model's outputs are the same as outputs from the model in eager mode. To support differences in TorchScript / eager mode outputs, an `unwrapper` function can be provided per-model. * Fix inception, use PYTORCH_TEST_WITH_SLOW * Update * Remove assertNestedTensorObjectsEqual * Add PYTORCH_TEST_WITH_SLOW to CircleCI config * Add MaskRCNN unwrapper * fix prec args * Remove CI changes * update * Update * remove expect changes * Fix tolerance bug * Fix breakages * Fix quantized resnet * Fix merge errors and simplify code * DeepLabV3 has been fixed * Temporarily disable jit compilation
-
- 26 Oct, 2019 1 commit
-
-
raghuramank100 authored
* add quantized models * Modify mobilenet.py documentation and clean up comments Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Move fuse_model method to QuantizableInvertedResidual and clean up args documentation Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Restore relu settings to default in resnet.py Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix missing return in forward Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix missing return in forwards Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Change pretrained -> pretrained_float_models Replace InvertedResidual with block Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Update tests to follow similar structure to test_models.py, allowing for modular testing Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Replace forward method with simple function assignment Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix error in arguments for resnet18 Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * pretrained_float_model argument missing for mobilenet Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * reference script for quantization aware training and post training quantization * reference script for quantization aware training and post training quantization * set pretrained_float_model as False and explicitly provide float model Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Address review comments: 1. Replace forward with _forward 2. Use pretrained models in reference train/eval script 3. Modify test to skip if fbgemm is not supported Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix lint errors. Use _forward for common code between float and quantized models Clean up linting for reference train scripts Test over all quantizable models Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Update default values for args in quantization/train.py Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Update models to conform to new API with quantize argument Remove apex in training script, add post training quant as an option Add support for separate calibration data set. Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix minor errors in train_quantization.py Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Remove duplicate file * Bugfix * Minor improvements on the models * Expose print_freq to evaluate * Minor improvements on train_quantization.py * Ensure that quantized models are created and run on the specified backends Fix errors in test only mode Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Add model urls * Fix errors in quantized model tests. Speedup creation of random quantized model by removing histogram observers Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Move setting qengine prior to convert. Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix lint error Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Add readme.md Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Readme.md Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: * Fix lint
-
- 17 Sep, 2019 1 commit
-
-
eellison authored
-
- 02 Sep, 2019 1 commit
-
-
eellison authored
* make shufflenet scriptable * make resnet18 scriptable * set downsample to identity instead of __constants__ api * use __constants__ for downsample instead of identity * import tensor to fix flake * use torch.Tensor type annotation instead of import
-
- 06 Aug, 2019 1 commit
-
-
Francisco Massa authored
* Add docs for video models * Fix docstrings for resnet and vgg
-
- 19 Jul, 2019 1 commit
-
-
apache2046 authored
Fix the old flatten method which use the size(0) to caculate the batch size, the old method will intruduce Gather opertion in the onnx output, which will faild parsed by tensorRT 5.0 (#1134)
-
- 04 Jul, 2019 1 commit
-
-
ekka authored
* Add paper references to VGG * Add paper references to ResNet and its variants
-
- 26 Jun, 2019 1 commit
-
-
Sergey Zagoruyko authored
* add wide resnet * add docstring for wide resnet * update WRN-50-2 model * add docs * extend WRN docstring * use pytorch storage for WRN * fix rebase * fix typo in docs
-
- 19 May, 2019 1 commit
-
-
Francisco Massa authored
Also move weights from ShuffleNet to PyTorch bucket. Additionally, rename shufflenet to make it consistent with the other models
-
- 17 May, 2019 1 commit
-
-
Sergey Zagoruyko authored
-
- 07 May, 2019 1 commit
-
-
bddppq authored
-
- 30 Apr, 2019 1 commit
-
-
Philip Meier authored
* added progress flag to model getters * flake8 * bug fix * backward commpability
-
- 24 Apr, 2019 1 commit
-
-
Francisco Massa authored
* Add dilation option to ResNet * Add a size check for replace_stride_with_dilation
-
- 15 Apr, 2019 1 commit
-
-
Ross Wightman authored
* Fix ResNeXt model defs with backwards compat for ResNet. * Fix Python 2.x integer div issue
-
- 05 Apr, 2019 1 commit
-
-
Francisco Massa authored
-
- 02 Apr, 2019 1 commit
-
-
Francisco Massa authored
* Add groups support to ResNet * Kill BaseResNet * Make it support multi-machine training
-
- 26 Mar, 2019 1 commit
-
-
Francisco Massa authored
-
- 11 Mar, 2019 1 commit
-
-
ekka authored
In reference to #729 added comments to clarify the naming and action of the layers performing downsampling in resnets.
-
- 11 Dec, 2018 1 commit
-
-
任广辉 authored
-
- 06 Dec, 2018 1 commit
-
-
Yuxin Wu authored
* Zero-init the residual branch in resnet * Add zero_init_residual as an option
-
- 12 Nov, 2018 1 commit
-
-
Amir Arsalan Soltani authored
-
- 06 Nov, 2018 2 commits
-
-
Erik Gaasedelen authored
* use predefined function * cleanup resnet * fix stride * revert inplanes since it isnt constant * lint
-
Adán Mora-Fallas authored
* AdaptiveAvgpool2d integrated The Average Pool Layer on SqueezeNet and ResNet was hardcoded, this was changed, now any the input size is accepted. * error fixed * linting fixed
-
- 01 Oct, 2018 1 commit
-
-
Binghuan Zhang authored
-
- 09 May, 2018 1 commit
-
-
vfdev authored
nit: replace 4 by expansion
-
- 30 Apr, 2018 1 commit
-
-
Ryuichiro Hataya authored
* fix for PyTorch v0.4 (remove .data and add _ for the initializations in nn.init) * fix m.**.**() style to nn.init.**(**) style * remove .idea * fix lines and indents * fix lines and indents * change to use `kaming_normal_` * add `.data` for safety * add nonlinearity='relu' for sure * fix indents
-
- 20 Dec, 2017 1 commit
-
-
Christopher Beckham authored
-
- 22 Oct, 2017 1 commit
-
-
Alykhan Tejani authored
-
- 13 Mar, 2017 1 commit
-
-
Soumith Chintala authored
-
- 27 Feb, 2017 1 commit
-
-
Luke Yeager authored
git ls-files | grep '\.py$' | xargs -n1 -P`nproc` autopep8 -i
-
- 24 Feb, 2017 1 commit
-
-
Luke Yeager authored
-