Edit model card

stable-diffusion-3.5-medium-GGUF

Original Model

stabilityai/stable-diffusion-3.5-medium

Run with sd-api-server

  • Version: coming soon

Quantized GGUF Models

Name Quant method Bits Size Use case
clip_g-Q4_0.gguf Q4_0 4 391 MB
clip_g-Q4_1.gguf Q4_1 4 435 MB
clip_g-Q5_0.gguf Q5_0 5 478 MB
clip_g-Q5_1.gguf Q5_1 5 522 MB
clip_g-Q8_0.gguf Q8_0 8 739 MB
clip_g-f16.gguf f16 16 1.39 GB
clip_g.safetensors f16 16 1.39 GB
clip_l-Q4_0.gguf Q4_0 4 69.4 MB
clip_l-Q4_1.gguf Q4_1 4 77.1 MB
clip_l-Q5_0.gguf Q5_0 5 84.8 MB
clip_l-Q5_1.gguf Q5_1 5 92.4 MB
clip_l-Q8_0.gguf Q8_0 8 131 MB
clip_l-f16.gguf f16 16 246 MB
clip_l.safetensors f16 16 246 MB
sd3.5_medium-Q4_0.gguf Q4_0 4 2.08 GB
sd3.5_medium-Q4_1.gguf Q4_1 4 2.22 GB
sd3.5_medium-Q5_0.gguf Q5_0 5 2.36 GB
sd3.5_medium-Q5_1.gguf Q5_1 5 2.50 GB
sd3.5_medium-Q8_0.gguf Q8_0 8 3.19 GB
sd3.5_medium.safetensors f16 16 5.11 GB
t5xxl-Q4_0.gguf Q4_0 4 2.75 GB
t5xxl-Q4_1.gguf Q4_1 4 3.06 GB
t5xxl-Q5_0.gguf Q5_0 5 3.36 GB
t5xxl-Q5_1.gguf Q5_1 5 3.67 GB
t5xxl-Q8_0.gguf Q8_0 8 5.20 GB
t5xxl_fp16.safetensors f16 16 9.79 GB

Quantized with stable-diffusion.cpp master-c3eeb669.

Downloads last month
655
GGUF
Model size
695M params
Architecture
undefined

4-bit

5-bit

8-bit

16-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for second-state/stable-diffusion-3.5-medium-GGUF

Quantized
(4)
this model