@@ -12,8 +12,6 @@ Set `"mm_config": {"mm_type": "W-int8-channel-sym-A-int8-channel-sym-dynamic-Vll
...
@@ -12,8 +12,6 @@ Set `"mm_config": {"mm_type": "W-int8-channel-sym-A-int8-channel-sym-dynamic-Vll
-`mm_type`: Specifies the quantized operator
-`mm_type`: Specifies the quantized operator
-`weight_auto_quant: true`: Enables automatic model quantization
-`weight_auto_quant: true`: Enables automatic model quantization
## Quantized Inference
### Offline Quantization
### Offline Quantization
lightx2v also supports direct loading of pre-quantized weights. For offline model quantization, refer to the [documentation](https://github.com/ModelTC/lightx2v/tree/main/tools/convert/readme.md).
lightx2v also supports direct loading of pre-quantized weights. For offline model quantization, refer to the [documentation](https://github.com/ModelTC/lightx2v/tree/main/tools/convert/readme.md).
...
@@ -21,6 +19,9 @@ Configure the [quantization file](https://github.com/ModelTC/lightx2v/tree/main/
...
@@ -21,6 +19,9 @@ Configure the [quantization file](https://github.com/ModelTC/lightx2v/tree/main/
1. Set `dit_quantized_ckpt` to the converted weight path
1. Set `dit_quantized_ckpt` to the converted weight path
2. Set `weight_auto_quant` to `false` in `mm_type`
2. Set `weight_auto_quant` to `false` in `mm_type`