onnx_int8.py 236 Bytes
Newer Older
zk's avatar
zk committed
1
2
3
4
5
6
7
8
from onnxruntime.quantization import quantize_dynamic, QuantType

quantize_dynamic(
    model_input="weights/ground.onnx",
    model_output="weights/ground_int8.onnx",
    weight_type=QuantType.QInt8,
)
print("int8 quantization done!")