hardware_aware_nas.rst 4.1 KB
Newer Older
1
2
3
Hardware-aware NAS
==================

Yuge Zhang's avatar
Yuge Zhang committed
4
.. This file should be rewritten as a tutorial
5

6
7
End-to-end Multi-trial SPOS Demo
--------------------------------
8

9
To empower affordable DNN on the edge and mobile devices, hardware-aware NAS searches both high accuracy and low latency models. In particular, the search algorithm only considers the models within the target latency constraints during the search process.
10

11
To run this demo, first install nn-Meter by running:
12
13
14

.. code-block:: bash

15
  pip install nn-meter
16
17
18
19
20
21
22

Then run multi-trail SPOS demo:

.. code-block:: bash

  python ${NNI_ROOT}/examples/nas/oneshot/spos/multi_trial.py

23

24
How the demo works
25
^^^^^^^^^^^^^^^^^^
26

27
To support hardware-aware NAS, you first need a ``Strategy`` that supports filtering the models by latency. We provide such a filter named ``LatencyFilter`` in NNI and initialize a ``Random`` strategy with the filter:
28
29
30

.. code-block:: python

31
  simple_strategy = strategy.Random(model_filter=LatencyFilter(threshold=100, predictor=base_predictor))
32
33
34
35

``LatencyFilter`` will predict the models\' latency by using nn-Meter and filter out the models whose latency are larger than the threshold (i.e., ``100`` in this example).
You can also build your own strategies and filters to support more flexible NAS such as sorting the models according to latency.

36
Then, pass this strategy to ``RetiariiExperiment``:
37
38
39

.. code-block:: python

40
  exp = RetiariiExperiment(base_model, trainer, strategy=simple_strategy)
41

42
43
44
45
46
47
48
  exp_config = RetiariiExeConfig('local')
  ...
  exp_config.dummy_input = [1, 3, 32, 32]

  exp.run(exp_config, port)

In ``exp_config``, ``dummy_input`` is required for tracing shape info.
49
50
51
52
53


End-to-end ProxylessNAS with Latency Constraints
------------------------------------------------

54
`ProxylessNAS <https://arxiv.org/abs/1812.00332>`__ is a hardware-aware one-shot NAS algorithm. ProxylessNAS applies the expected latency of the model to build a differentiable metric and design efficient neural network architectures for hardware. The latency loss is added as a regularization term for architecture parameter optimization. In this example, nn-Meter provides a latency estimator to predict expected latency for the mixed operation on other types of mobile and edge hardware. 
55
56
57
58
59
60
61
62
63

To run the one-shot ProxylessNAS demo, first install nn-Meter by running:

.. code-block:: bash

  pip install nn-meter

Then run one-shot ProxylessNAS demo:

Yuge Zhang's avatar
Yuge Zhang committed
64
65
.. code-block:: bash

Yuge Zhang's avatar
Yuge Zhang committed
66
   python ${NNI_ROOT}/examples/nas/oneshot/proxylessnas/main.py --applied_hardware HARDWARE --reference_latency REFERENCE_LATENCY_MS
67
68
69
70

How the demo works
^^^^^^^^^^^^^^^^^^

Yuge Zhang's avatar
Yuge Zhang committed
71
In the implementation of ProxylessNAS ``trainer``, we provide a ``HardwareLatencyEstimator`` which currently builds a lookup table, that stores the measured latency of each candidate building block in the search space. The latency sum of all building blocks in a candidate model will be treated as the model inference latency. The latency prediction is obtained by ``nn-Meter``. ``HardwareLatencyEstimator`` predicts expected latency for the mixed operation based on the path weight of ``ProxylessLayerChoice``. With leveraging ``nn-Meter`` in NNI, users can apply ProxylessNAS to search efficient DNN models on more types of edge devices. 
72
73
74
75
76
77
78
79

Despite of ``applied_hardware`` and ``reference_latency``, There are some other parameters related to hardware-aware ProxylessNAS training in this :githublink:`example <examples/nas/oneshot/proxylessnas/main.py>`:

* ``grad_reg_loss_type``: Regularization type to add hardware related loss. Allowed types include ``"mul#log"`` and ``"add#linear"``. Type of ``mul#log`` is calculate by ``(torch.log(expected_latency) / math.log(reference_latency)) ** beta``. Type of ``"add#linear"`` is calculate by ``reg_lambda * (expected_latency - reference_latency) / reference_latency``. 
* ``grad_reg_loss_lambda``: Regularization params, is set to ``0.1`` by default.
* ``grad_reg_loss_alpha``: Regularization params, is set to ``0.2`` by default.
* ``grad_reg_loss_beta``: Regularization params, is set to ``0.3`` by default.
* ``dummy_input``: The dummy input shape when applied to the target hardware. This parameter is set as (1, 3, 224, 224) by default.