README.md 8.81 KB
Newer Older
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
1
2
# Submanifold Sparse Convolutional Networks

Ben Graham's avatar
Ben Graham committed
3
This is the Torch/PyTorch library for training Submanifold Sparse Convolutional Networks. If you find this code useful in your research then please cite:
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
4
5
6
7
8

**[Submanifold Sparse Convolutional Networks, https://arxiv.org/abs/1706.01307](https://arxiv.org/abs/1706.01307)** <br />
[Benjamin Graham](https://research.fb.com/people/graham-benjamin/), <br />
[Laurens van der Maaten](https://lvdmaaten.github.io/), <br />

Ben Graham's avatar
Ben Graham committed
9
10
11
12
13
14
15
16
```
@article{SubmanifoldSparseConvNet,
  title={Submanifold Sparse Convolutional Networks},
  author={Graham, Benjamin and van der Maaten, Laurens},
  journal={arXiv preprint arXiv:1706.01307},
  year={2017}
}
```
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
17
18
19

## Spatial sparsity

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
20
This library brings [Spatially-sparse convolutional networks](https://github.com/btgraham/SparseConvNet) to PyTorch. Moreover, it introduces **Submanifold Sparse Convolutions**, that can be used to build computationally efficient sparse VGG/ResNet/DenseNet-style networks.
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
21
22

With regular 3x3 convolutions, the set of active (non-zero) sites grows rapidly:<br />
Ben Graham's avatar
Ben Graham committed
23
![submanifold](img/i.gif) <br />
Ben Graham's avatar
Ben Graham committed
24
With **Submanifold Sparse Convolutions**, the set of active sites is unchanged. Active sites look at their active neighbors (green); non-active sites (red) have no computational overhead: <br />
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
25
![submanifold](img/img.gif) <br />
Ben Graham's avatar
Ben Graham committed
26
Stacking Submanifold Sparse Convolutions to build VGG and ResNet type ConvNets, information can flow along lines or surfaces of active points.<br />
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
27

Ben Graham's avatar
Ben Graham committed
28
Disconnected components don't communicate at first, although they will merge due to the effect of strided operations, either pooling or convolutions. Additionally, adding ConvolutionWithStride2-SubmanifoldConvolution-DeconvolutionWithStride2 paths to the network allows disjoint active sites to communicate; see the 'VGG+' networks in the paper.<br />
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
29
30
![Strided Convolution, convolution, deconvolution](img/img_stridedConv_conv_deconv.gif) <br />
![Strided Convolution, convolution, deconvolution](img/img_stridedConv_conv_deconv.png) <br />
Ben Graham's avatar
Ben Graham committed
31
From left: **(i)** an active point is highlighted; a convolution with stride 2 sees the green active sites **(ii)** and produces output **(iii)**, 'children' of hightlighted active point from (i) are highlighted; a submanifold sparse convolution sees the green active sites **(iv)** and produces output **(v)**; a deconvolution operation sees the green active sites **(vi)**  and produces output **(vii)**.
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52

## Dimensionality and 'submanifolds'

SparseConvNet supports input with different numbers of spatial/temporal dimensions.
Higher dimensional input is more likely to be sparse because of the 'curse of dimensionality'. <br />

  Dimension|Name in 'torch.nn'|Use cases
  :--:|:--:|:--:
  1|TemporalConvolution| Text, audio
  2|SpatialConvolution|Lines in 2D space, e.g. handwriting
  3|VolumetricConvolution|Lines and surfaces in 3D space or (2+1)D space-time
  4| - |Lines, etc,  in (3+1)D space-time

We use the term 'submanifold' to refer to input data that is sparse because it has a lower effective dimension than the space in which it lives, for example a one-dimensional curve in 2+ dimensional space, or a two-dimensional surface in 3+ dimensional space.

In theory, the library supports up to 10 dimensions. In practice, ConvNets with size-3 SVC convolutions in dimension 5+ may be impractical as the number of parameters per convolution is growing exponentially. Possible solutions include factorizing the convolutions (e.g. 3x1x1x..., 1x3x1x..., etc), or switching to a hyper-tetrahedral lattice (see [Sparse 3D convolutional neural networks](http://arxiv.org/abs/1505.02890)).





Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
53
## Hello World
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
54
SparseConvNets can be built either by [defining a function that inherits from torch.nn.Module](examples/Assamese_handwriting/VGGplus.py) or by stacking modules in a [sparseconvnet.Sequential](PyTorch/sparseconvnet/sequential.py):
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
55
56
```
import torch
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
57
import sparseconvnet as scn
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
58
59

# Use the GPU if there is one, otherwise CPU
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
60
use_gpu = torch.cuda.is_available()
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
61
62
63

model = scn.Sequential().add(
    scn.SparseVggNet(2, 1,
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
64
65
66
                     [['C',  8], ['C',  8], ['MP', 3, 2],
                      ['C', 16], ['C', 16], ['MP', 3, 2],
                      ['C', 24], ['C', 24], ['MP', 3, 2]])
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
67
).add(
Ben Graham's avatar
Ben Graham committed
68
    scn.SubmanifoldConvolution(2, 24, 32, 3, False)
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
69
70
71
).add(
    scn.BatchNormReLU(32)
).add(
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
72
73
74
75
    scn.SparseToDense(2,32)
)
if use_gpu:
    model.cuda()
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
76
77

# output will be 10x10
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
78
inputSpatialSize = model.input_spatial_size(torch.LongTensor([10, 10]))
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
79
80
81
82
83
84
85
86
input = scn.InputBatch(2, inputSpatialSize)

msg = [
    " X   X  XXX  X    X    XX     X       X   XX   XXX   X    XXX   ",
    " X   X  X    X    X   X  X    X       X  X  X  X  X  X    X  X  ",
    " XXXXX  XX   X    X   X  X    X   X   X  X  X  XXX   X    X   X ",
    " X   X  X    X    X   X  X     X X X X   X  X  X  X  X    X  X  ",
    " X   X  XXX  XXX  XXX  XX       X   X     XX   X  X  XXX  XXX   "]
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
87

Benjamin Thomas Graham's avatar
tidy  
Benjamin Thomas Graham committed
88
89
#Add a sample using set_location
input.add_sample()
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
90
91
92
93
94
for y, line in enumerate(msg):
    for x, c in enumerate(line):
        if c == 'X':
            location = torch.LongTensor([x, y])
            featureVector = torch.FloatTensor([1])
Benjamin Thomas Graham's avatar
tidy  
Benjamin Thomas Graham committed
95
            input.set_location(location, featureVector, 0)
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
96

Benjamin Thomas Graham's avatar
tidy  
Benjamin Thomas Graham committed
97
98
#Add a sample using set_locations
input.add_sample()
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
99
100
101
102
103
104
105
106
107
locations = []
features = []
for y, line in enumerate(msg):
    for x, c in enumerate(line):
        if c == 'X':
            locations.append([x,y])
            features.append([1])
locations = torch.LongTensor(locations)
features = torch.FloatTensor(features)
Benjamin Thomas Graham's avatar
tidy  
Benjamin Thomas Graham committed
108
input.set_locations(locations, features, 0)
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
109

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
110
111
112
113
114
115
116
117
# Optional: allow metadata preprocessing to be done in batch preparation threads
# to improve GPU utilization.
#
# Parameter:
#    3 if using MP3/2 pooling or C3/2 convolutions for downsizing,
#    2 if using MP2 pooling for downsizing.
input.precomputeMetadata(3)

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
118
119
120
model.train()
if use_gpu:
    input.cuda()
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
121
122
output = model.forward(input)

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
123
# Output is 2x32x10x10: our minibatch has 2 samples, the network has 32 output
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
124
# feature planes, and 10x10 is the spatial size of the output.
125
print(output.size(), output.type())
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
```


## Examples

Examples in the examples folder include
* [Assamese handwriting recognition](https://archive.ics.uci.edu/ml/datasets/Online+Handwritten+Assamese+Characters+Dataset#)
* [Chinese handwriting for recognition](http://www.nlpr.ia.ac.cn/databases/handwriting/Online_database.html)
* [ModelNet-40 Object recognition](http://modelnet.cs.princeton.edu/) using [3dcnn.torch preprocessed data](https://github.com/charlesq34/3dcnn.torch)

Data will be downloaded/preprocessed on the first run, i.e.
```
cd examples/Assamese_handwriting
python VGGplus.py
```

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
142
## Setup
143

Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
144
Tested with Ubuntu 16.04, Python 3 in [Miniconda](https://conda.io/miniconda.html) and PyTorch v0.4 (with merged Tensors/Variables).
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
145
146

```
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
147
148
conda install -c pytorch # OR git clone https://github.com/pytorch/pytorch.git; cd pytorch; python setup.py install; cd ..
conda install -c bioconda google-sparsehash # OR apt-get install libsparsehash-dev
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
149
150
git clone git@github.com:facebookresearch/SparseConvNet.git
cd SparseConvNet/PyTorch/
151
python setup.py install
Benjamin Thomas Graham's avatar
Benjamin Thomas Graham committed
152
153
154
155
156
157
158
```
To run the examples you may also need to install unrar and TorchNet:
```
apt-get install unrar
pip install git+https://github.com/pytorch/tnt.git@master
```

159
160


Ben Graham's avatar
Ben Graham committed
161
162
163
164
165
### Links
1. [ICDAR 2013 Chinese Handwriting Recognition Competition 2013](http://www.nlpr.ia.ac.cn/events/CHRcompetition2013/competition/Home.html) First place in task 3, with test error of 2.61%. Human performance on the test set was 4.81%. [Report](http://www.nlpr.ia.ac.cn/events/CHRcompetition2013/competition/ICDAR%202013%20CHR%20competition.pdf)
2. [Spatially-sparse convolutional neural networks, 2014](http://arxiv.org/abs/1409.6070) SparseConvNets for Chinese handwriting recognition
3. [Fractional max-pooling, 2014](http://arxiv.org/abs/1412.6071) A SparseConvNet with fractional max-pooling achieves an error rate of 3.47% for CIFAR-10.
4. [Sparse 3D convolutional neural networks, BMVC 2015](http://arxiv.org/abs/1505.02890) SparseConvNets for 3D object recognition and (2+1)D video action recognition.
166
5. [Kaggle plankton recognition competition, 2015](https://www.kaggle.com/c/datasciencebowl) Third place. The competition solution is being adapted for research purposes in [EcoTaxa](http://ecotaxa.obs-vlfr.fr/).
Ben Graham's avatar
Ben Graham committed
167
168
169
6. [Kaggle Diabetic Retinopathy Detection, 2015](https://www.kaggle.com/c/diabetic-retinopathy-detection/) First place in the Kaggle Diabetic Retinopathy Detection competition.
7. [Submanifold Sparse Convolutional Networks, 2017](https://arxiv.org/abs/1706.01307) Introduces deep 'submanifold' SparseConvNets.
8. [Workshop on Learning to See from 3D Data, 2017](https://shapenet.cs.stanford.edu/iccv17workshop/) First place in the [semantic segmentation](https://shapenet.cs.stanford.edu/iccv17/) competition. [Report](https://arxiv.org/pdf/1710.06104)
Ben Graham's avatar
Ben Graham committed
170
9. [3D Semantic Segmentation with Submanifold Sparse Convolutional Networks, 2017](https://arxiv.org/abs/1711.10275) Semantic segmentation for the ShapeNet Core55 and NYU-DepthV2 datasets