quantization.rst 932 Bytes
Newer Older
J-shang's avatar
J-shang committed
1
2
3
Model Quantization with NNI
===========================

J-shang's avatar
J-shang committed
4
5
6
7
8
9
Quantization refers to compressing models by reducing the number of bits required to represent weights or activations,
which can reduce the computations and the inference time. In the context of deep neural networks, the major numerical
format for model weights is 32-bit float, or FP32. Many research works have demonstrated that weights and activations
can be represented using 8-bit integers without significant loss in accuracy. Even lower bit-widths, such as 4/2/1 bits,
is an active field of research.

10
11
A quantizer is a quantization algorithm implementation in NNI.
You can also :doc:`create your own quantizer <../tutorials/quantization_customize>` using NNI model compression interface.
J-shang's avatar
J-shang committed
12
13
14

.. toctree::
    :hidden:
J-shang's avatar
J-shang committed
15
16
    :maxdepth: 2

17
    Quickstart <../tutorials/quantization_quick_start_mnist>
18
    Quantizer <quantizer>
19
    SpeedUp <../tutorials/quantization_speedup>