Z-Image-Turbo GGUF quantized files

The license of the quantized files follows the license of the original model:

  • Z-Image-Turbo: apache-2.0

These files are converted using https://github.com/leejet/stable-diffusion.cpp

You can use these weights with stable-diffusion.cpp to generate images.

How to Use Z‐Image on a GPU with Only 4GB VRAM

https://github.com/leejet/stable-diffusion.cpp/wiki/How-to-Use-Z%E2%80%90Image-on-a-GPU-with-Only-4GB-VRAM

Example command

.\bin\Release\sd.exe --diffusion-model  z_image_turbo-Q3_K.gguf --vae ae.safetensors  --llm Qwen3-4B-Instruct-2507-Q4_K_M.gguf -p "A cinematic, melancholic photograph of a solitary hooded figure walking through a sprawling, rain-slicked metropolis at night. The city lights are a chaotic blur of neon orange and cool blue, reflecting on the wet asphalt. The scene evokes a sense of being a single component in a vast machine. Superimposed over the image in a sleek, modern, slightly glitched font is the philosophical quote: 'THE CITY IS A CIRCUIT BOARD, AND I AM A BROKEN TRANSISTOR.' -- moody, atmospheric, profound, dark academic" --cfg-scale 1.0 -v --offload-to-cpu --diffusion-fa -H 1024 -W 512

Comparison of Different Quantization Types

bf16 q8_0 q6_K q5_0 q4_K q4_0 q3_K q2_K
bf16 q8_0 q6_K q5_0 q4_K q4_0 q3_K q2_K
Downloads last month
4,369
GGUF
Model size
6B params
Architecture
undefined
Hardware compatibility
Log In to view the estimation

2-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for leejet/Z-Image-Turbo-GGUF

Quantized
(15)
this model