FLUX.1

State-of-the-art image generation model from Black Forest Labs on CLORE.AI GPUs.

circle-check

Why FLUX.1?

  • Best quality - Superior to SDXL and Midjourney v5

  • Text rendering - Actually readable text in images

  • Prompt following - Excellent instruction adherence

  • Fast variants - FLUX.1-schnell for quick generation

Model Variants

Model
Speed
Quality
VRAM
License

FLUX.1-schnell

Fast (4 steps)

Great

12GB+

Apache 2.0

FLUX.1-dev

Medium (20 steps)

Excellent

16GB+

Non-commercial

FLUX.1-pro

API only

Best

-

Commercial

Quick Deploy on CLORE.AI

Docker Image:

ghcr.io/huggingface/text-generation-inference:latest

Ports:

For easiest deployment, use ComfyUI with FLUX nodes.

Installation Methods

Method 2: Diffusers

Method 3: Fooocus

Fooocus has built-in FLUX support:

ComfyUI Workflow

FLUX.1-schnell (Fast)

Nodes needed:

  1. Load Diffusion Model → flux1-schnell.safetensors

  2. DualCLIPLoader → clip_l.safetensors + t5xxl_fp16.safetensors

  3. CLIP Text Encode → your prompt

  4. Empty SD3 Latent Image → set dimensions

  5. KSampler → steps: 4, cfg: 1.0

  6. VAE Decode → ae.safetensors

  7. Save Image

FLUX.1-dev (Quality)

Same workflow but:

  • Steps: 20-50

  • CFG: 3.5

  • Use guidance_scale in prompt

Python API

Basic Generation

With Memory Optimization

Batch Generation

FLUX.1-dev (Higher Quality)

Prompt Tips

FLUX excels at:

  • Text in images: "A neon sign that says 'OPEN 24/7'"

  • Complex scenes: "A busy Tokyo street at night with reflections"

  • Specific styles: "Oil painting in the style of Monet"

  • Detailed descriptions: Long, detailed prompts work well

Example Prompts

Memory Optimization

For 12GB VRAM (RTX 3060)

For 8GB VRAM

Use quantized version or ComfyUI with GGUF:

Performance Comparison

Model
Steps
Time (4090)
Quality

FLUX.1-schnell

4

~3 sec

Great

FLUX.1-dev

20

~12 sec

Excellent

FLUX.1-dev

50

~30 sec

Best

SDXL

30

~8 sec

Good

GPU Requirements

Setup
Minimum
Recommended

FLUX.1-schnell

12GB

16GB+

FLUX.1-dev

16GB

24GB+

With CPU offload

8GB

12GB+

Quantized (GGUF)

6GB

8GB+

GPU Presets

RTX 3060 12GB (Budget)

RTX 3090 24GB (Optimal)

RTX 4090 24GB (Performance)

A100 40GB/80GB (Production)

Cost Estimate

GPU
Hourly
Images/Hour

RTX 3060 12GB

~$0.03

~200 (schnell)

RTX 3090 24GB

~$0.06

~600 (schnell)

RTX 4090 24GB

~$0.10

~1000 (schnell)

A100 40GB

~$0.17

~1500 (schnell)

Troubleshooting

Out of Memory

Slow Generation

  • Use FLUX.1-schnell (4 steps)

  • Enable torch.compile: pipe.unet = torch.compile(pipe.unet)

  • Use fp16 instead of bf16 on older GPUs

Poor Quality

  • Use more steps (FLUX-dev: 30-50)

  • Increase guidance_scale (3.0-4.0 for dev)

  • Write more detailed prompts

Next Steps

Last updated

Was this helpful?