Qwen-Image-Edit =============== Original Qwen-Image-Edit ------------------------ `Qwen-Image-Edit `_ is the image editing version of Qwen-Image. Below is a minimal example for running the 4-bit quantized `Qwen-Image-Edit `_ model with Nunchaku. Nunchaku offers an API compatible with `Diffusers `_, allowing for a familiar user experience. .. literalinclude:: ../../../examples/v1/qwen-image-edit.py :language: python :caption: Running Qwen-Image-Edit (`examples/v1/qwen-image-edit.py `__) :linenos: When using Nunchaku, replace the standard ``QwenImageTransformer2dModel`` with :class:`~nunchaku.models.transformers.transformer_qwenimage.NunchakuQwenImageTransformer2DModel`. The :meth:`~nunchaku.models.transformers.transformer_qwenimage.NunchakuQwenImageTransformer2DModel.from_pretrained` method loads quantized models from either Hugging Face or local file paths. .. note:: - The :func:`~nunchaku.utils.get_precision` function automatically detects whether your GPU supports INT4 or FP4 quantization. Use FP4 models for Blackwell GPUs (RTX 50-series) and INT4 models for other architectures. - Increasing the rank (e.g., to 128) can improve output quality. - To reduce VRAM usage, enable asynchronous CPU offloading with :meth:`~nunchaku.models.transformers.transformer_qwenimage.NunchakuQwenImageTransformer2DModel.set_offload`. For further savings, you may also enable Diffusers' ``pipeline.enable_sequential_cpu_offload()``, but be sure to exclude ``transformer`` from offloading, as Nunchaku's offloading mechanism differs from Diffusers'. With these settings, VRAM usage can be reduced to approximately 3GB. Distilled Qwen-Image-Edit (Qwen-Image-Lightning) ------------------------------------------------ For faster inference, we provide pre-quantized 4-step and 8-step Qwen-Image-Edit models by integrating `Qwen-Image-Lightning LoRAs `_. See the example script below: .. literalinclude:: ../../../examples/v1/qwen-image-edit-lightning.py :language: python :caption: Running Qwen-Image-Edit-Lightning (`examples/v1/qwen-image-edit-lightning.py `__) :linenos: Custom LoRA support is under development.