4bit quantization for arbitrary `nn.Parameter` (#1720)
* Add parametrize util for targeting parameters outside of nn.Linear modules * Parametrize 4bit: replace existing prequantized weight * cleanup * Add caching for parametrization * Add tests * Fix tests * Guard for torch < 2.5 * Guard for torch < 2.5 * Another test gaurd for torch >= 2.5
Showing
tests/test_parametrize.py
0 → 100644
Please register or sign in to comment