README.txt 1.12 KB
Newer Older
1
2
3
4
5
.. _tutorials4-index:


Old (new) wines in new bottle
-----------------------------
6
7
8
9
10
11
12
13
14
* **Capsule** `[paper] <https://arxiv.org/abs/1710.09829>`__ `[tutorial]
  <4_old_wines/2_capsule.html>`__ `[code]
  <https://github.com/jermainewang/dgl/tree/master/examples/pytorch/capsule>`__:
  this new computer vision model has two key ideas -- enhancing the feature
  representation in a vector form (instead of a scalar) called *capsule*, and
  replacing max-pooling with dynamic routing. The idea of dynamic routing is to
  integrate a lower level capsule to one (or several) of a higher level one
  with non-parametric message-passing. We show how the later can be nicely
  implemented with DGL APIs.
15

16
17
18
19
20
21
22
* **Transformer** `[paper] <https://arxiv.org/abs/1706.03762>`__ `[tutorial
  (wip)]` `[code (wip)]` and **Universal Transformer** `[paper]
  <https://arxiv.org/abs/1807.03819>`__ `[tutorial (wip)]` `[code (wip)]`:
  these two models replace RNN with several layers of multi-head attention to
  encode and discover structures among tokens of a sentence. These attention
  mechanisms can similarly formulated as graph operations with
  message-passing.