--- base_model: - BasedBase/WEBGEN-4B-Preview-480B-Double-Distill-FP32 pipeline_tag: text-generation --- # 🧠 WEBGEN-4B-Preview-480B-Double-Distill GGUFs Quantized version of: [BasedBase/WEBGEN-4B-Preview-480B-Double-Distill-FP32](https://huggingface.co/BasedBase/WEBGEN-4B-Preview-480B-Double-Distill-FP32) --- ## 📦 Available GGUFs | Format | Description | |--------|-------------| | **F16** | Full precision (16-bit), better quality, larger size ⚖️ | | **Q8_K_XL** | Quantized (8-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), smaller size, faster inference ⚡ | | **Q4_K_XL** | Quantized (4-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), smaller size, faster inference ⚡ | --- ## 🚀 Usage Example with `llama.cpp`: ```bash ./main -m ./gguf-file-name.gguf -p "Hello world!" ````