Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
fengzch-das
nunchaku
Commits
967895a0
Commit
967895a0
authored
Nov 14, 2024
by
Muyang Li
Browse files
add Q-Diffusion
parent
42fe6cc2
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
0 deletions
+1
-0
README.md
README.md
+1
-0
No files found.
README.md
View file @
967895a0
...
@@ -127,6 +127,7 @@ If you find `nunchaku` useful or relevant to your research, please cite our pape
...
@@ -127,6 +127,7 @@ If you find `nunchaku` useful or relevant to your research, please cite our pape
*
[
Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
](
https://arxiv.org/abs/2211.02048
)
, NeurIPS 2022 & T-PAMI 2023
*
[
Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
](
https://arxiv.org/abs/2211.02048
)
, NeurIPS 2022 & T-PAMI 2023
*
[
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
](
https://arxiv.org/abs/2211.10438
)
, ICML 2023
*
[
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
](
https://arxiv.org/abs/2211.10438
)
, ICML 2023
*
[
Q-Diffusion: Quantizing Diffusion Models
](
https://arxiv.org/abs/2302.04304
)
, ICCV 2023
*
[
AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
](
https://arxiv.org/abs/2306.00978
)
, MLSys 2024
*
[
AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
](
https://arxiv.org/abs/2306.00978
)
, MLSys 2024
*
[
DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models
](
https://arxiv.org/abs/2402.19481
)
, CVPR 2024
*
[
DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models
](
https://arxiv.org/abs/2402.19481
)
, CVPR 2024
*
[
QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving
](
https://arxiv.org/abs/2405.04532
)
, ArXiv 2024
*
[
QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving
](
https://arxiv.org/abs/2405.04532
)
, ArXiv 2024
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment