Edit model card

Quantized versions of https://huggingface.co/black-forest-labs/FLUX.1-schnell

Tools used for quantization: modded stable-diffusion.cpp, LlamaQuantizer

Work in progress, use at your own risk

How to:

[WIP]

  1. Dowload and build stable-diffusion.cpp
  2. Download one of the models from this repo and
  3. Enter your stable-diffusion.cpp dir
  4. Run the following command:
    ./build/bin/sd --diffusion-model [path to gguf] --vae [path to ae.safetensors] --clip_l [path to clip_l.safetensors] --t5xxl [path to t5xxl_fp16.safetensors] -p "a frog holding a sign saying 'hi' " -o ../frog.png -v --cfg-scale 1.0 --sampling-method euler -v --seed 42 --steps 4
    

Results:

Quant type Size Result (x0.5) Download link
default 23.8 GB flux_frog_default.png flux1-schnell.safetensors.gguf
FP16 23.8 GB flux_frog_F16.png flux1-schnell-F16.gguf
Q8_0 12.6 GB flux_frog_Q8_0.png flux1-schnell-Q8_0.gguf
Q5_0 8.18 GB flux_frog_Q5_0.png flux1-schnell-Q5_0.gguf
Q5_1 8.92 GB flux_frog_Q5_1.png flux1-schnell-Q5_1.gguf
Q4_0 6.69 GB flux_frog_Q4_0.png flux1-schnell-Q4_0.gguf
Q4_1 7.43 GB flux_frog_Q4_1.png flux1-schnell-Q4_1.gguf
Q4_K 6.69 GB flux_frog_Q4_K.png flux1-schnell-Q4_K.gguf
Q2_K 3.9 GB flux_frog_Q2_K.png flux1-schnell-Q2_K.gguf
Downloads last month
1,899
GGUF
Model size
11.9B params
Architecture
undefined

2-bit

4-bit

5-bit

8-bit

16-bit

Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for aifoundry-org/FLUX.1-schnell-Quantized

Quantized
this model

Space using aifoundry-org/FLUX.1-schnell-Quantized 1