Model Card (SVDQuant)

Language: English | 中文

Chroma1-HD

Model name

  • Model repo: tonera/Chroma1-HD-SVDQ
  • Base (Diffusers weights path): tonera/Chroma1-HD-SVDQ (repo root)
  • Quantized Transformer weights: tonera/Chroma1-HD-SVDQ/svdq-<precision>_r32-Chroma1-HD.safetensors

Quantization / inference tech

  • Inference engine: Nunchaku (https://github.com/nunchaku-ai/nunchaku)

Nunchaku is a high-performance inference engine for 4-bit (FP4/INT4) low-bit neural networks. Its goal is to significantly reduce VRAM usage and improve inference speed while preserving generation quality as much as possible. It implements and productionizes post-training quantization methods such as SVDQuant, and uses operator/kernel fusion and other optimizations to reduce the extra overhead introduced by low-rank branches.

The Chroma1-HD quantized weights in this repository (e.g. svdq-*_r32-*.safetensors) are meant to be used with Nunchaku for efficient inference on supported GPUs.

You must install Nunchaku before use

  • Official installation docs (recommended source of truth): https://nunchaku.tech/docs/nunchaku/installation/installation.html

(Recommended) Install the official prebuilt wheel

  • Prerequisite: PyTorch >= 2.5 (follow the wheel requirements as the source of truth)
  • Install the nunchaku wheel: pick the wheel matching your environment from GitHub Releases / HuggingFace / ModelScope (note cp311 means Python 3.11):
    • https://github.com/nunchaku-ai/nunchaku/releases
# Example (choose the correct wheel URL for your torch/cuda/python versions)
pip install https://github.com/nunchaku-ai/nunchaku/releases/download/vX.Y.Z/nunchaku-X.Y.Z+torch2.9-cp311-cp311-linux_x86_64.whl
  • Tip (RTX 50 series GPUs): usually CUDA >= 12.8 is recommended, and FP4 models are preferred for better compatibility and performance (follow the official docs).

Usage example (Diffusers + Nunchaku Transformer)

Note: I am pushing for the official Nunchaku PR to be merged: https://github.com/nunchaku-ai/nunchaku/pull/916 Until then, if you want to try it out, you can copy transformer_chroma.py from the repository to nunchaku/models/transformers/transformer_chroma.py.This is a relatively low-performance version; a high-performance version still needs to wait for the PR to be approved.

import torch
from diffusers import ChromaPipeline

from nunchaku import NunchakuChromaTransformer2dModel
from nunchaku.utils import get_precision

MODEL = "Chroma1-HD-SVDQ"
REPO_ID = f"tonera/{MODEL}"

if __name__ == "__main__":
    transformer = NunchakuChromaTransformer2dModel.from_pretrained(
        f"{REPO_ID}/svdq-{get_precision()}_r32-{MODEL}.safetensors"
    )

    pipe = ChromaPipeline.from_pretrained(
        f"{REPO_ID}",
        transformer=transformer,
        torch_dtype=torch.bfloat16,
        use_safetensors=True,
    ).to("cuda")

    prompt = "Make Pikachu hold a sign that says 'Nunchaku is awesome', yarn art style, detailed, vibrant colors"
    image = pipe(prompt=prompt, guidance_scale=2.5, num_inference_steps=40).images[0]
    image.save("Chroma1.png")
Downloads last month
178
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tonera/Chroma1-HD-SVDQ

Unable to build the model tree, the base model loops to the model itself. Learn more.