AI Agent - Mar 20, 2026

Flux Pro & Dev FAQ: Licensing, LoRA Fine-Tuning, API Rate Limits, and Everything Technical

Flux Pro & Dev FAQ: Licensing, LoRA Fine-Tuning, API Rate Limits, and Everything Technical

Licensing

What license is Flux Dev released under?

Flux Dev is released under the FLUX.1 [dev] Non-Commercial License initially, which was later updated to a more permissive license. Check the current license on the official repository for the latest terms. Key points:

  • Research and personal use: Permitted
  • Commercial use: Check current license terms (license has been updated since initial release)
  • Derivative works: Permitted with attribution
  • Model redistribution: Permitted with license inclusion

What license is Flux Schnell released under?

Flux Schnell uses the Apache 2.0 license — the most permissive option:

  • Commercial use: Fully permitted
  • Modification: Permitted
  • Distribution: Permitted
  • Private use: Permitted
  • Patent use: Permitted

For commercial products, Flux Schnell provides the clearest licensing path.

What about Flux Pro licensing?

Flux Pro is available only via API. Commercial usage rights are included in the API pricing — you pay per generation and can use all generated images commercially. No separate license purchase is required.

Can I sell images generated by Flux?

Flux Pro (API): Yes. Commercial usage is included in API pricing. Flux Schnell: Yes. Apache 2.0 places no restrictions on output usage. Flux Dev: Check current license terms for commercial use provisions.

The copyright status of AI-generated images varies by jurisdiction. In most cases:

  • You have usage rights to images you generate
  • Black Forest Labs does not claim ownership of generated outputs
  • Copyright registration may or may not be available depending on your jurisdiction
  • Consult legal counsel for specific commercial applications

LoRA Fine-Tuning

Can I fine-tune Flux models?

Yes. Both Flux Dev and Flux Schnell can be fine-tuned using LoRA (Low-Rank Adaptation). Flux Pro cannot be fine-tuned (API-only access).

What do I need for LoRA training?

Hardware:

  • Minimum: GPU with 16GB VRAM (RTX 4060 Ti 16GB, RTX 3090)
  • Recommended: GPU with 24GB VRAM (RTX 4090, A5000)
  • For faster training: Multi-GPU setup

Software:

  • Kohya_ss (most popular Flux LoRA training tool)
  • SimpleTuner (alternative training framework)
  • ai-toolkit by ostris (lightweight option)

Training Data:

  • 15-50 high-quality images for the target concept/style
  • Consistent quality and style across training images
  • Accompanying text captions describing each image
  • Higher resolution is better (1024×1024 minimum)

How long does LoRA training take?

GPUTraining StepsApproximate Time
RTX 3090 24GB1,000~2 hours
RTX 4090 24GB1,000~1.5 hours
A100 80GB1,000~45 minutes

Most LoRAs converge between 500-2,000 training steps. Start with 1,000 steps and evaluate.

What training parameters should I use?

Recommended starting parameters:

  • Learning rate: 1e-4 for standard LoRA
  • LoRA rank: 16-32 (higher = more capacity, more VRAM)
  • Batch size: 1-2 (limited by VRAM)
  • Resolution: 1024×1024
  • Optimizer: AdamW8bit
  • Scheduler: Cosine with warmup
  • Training steps: 1,000-2,000

These are starting points. Optimal parameters vary by training data and target concept. Experimentation is expected.

How do I use a LoRA with Flux?

In ComfyUI:

  1. Place the LoRA file in the models/loras/ directory
  2. Add a LoRA Loader node connected to the Flux model loader
  3. Set the LoRA weight (0.5-1.0 is typical; start with 0.7)
  4. Generate as normal — the LoRA influences the output style/content

In Diffusers (Python):

from diffusers import FluxPipeline
import torch

pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-dev")
pipe.load_lora_weights("path/to/your/lora.safetensors")
pipe.to("cuda")

image = pipe("your prompt", num_inference_steps=50).images[0]

Can I share or sell my LoRA?

Yes. LoRA files are derivative works and can be distributed under the terms of the base model’s license. Common distribution platforms:

  • CivitAI (largest community, easy sharing)
  • Hugging Face (developer-oriented)
  • Personal websites

API Rate Limits

What are the rate limits for Flux Pro API?

Rate limits vary by provider:

ProviderFree TierPaid TierEnterprise
BFL Direct1 req/sec10 req/secCustom
Replicate10 req/min (free)Based on planCustom
Fal.ai5 req/min (free)60 req/minCustom
Together AIBased on creditsBased on planCustom

What happens when I hit rate limits?

Most providers return an HTTP 429 (Too Many Requests) with a Retry-After header. Best practices:

  • Implement exponential backoff
  • Use a request queue to manage concurrency
  • Batch non-urgent requests during off-peak times
  • Contact the provider for increased limits if needed

Is there a maximum image resolution via API?

ProviderMaximum ResolutionMaximum Megapixels
BFL Direct2048×2048~4MP
Replicate2048×2048~4MP
Fal.ai2048×2048~4MP

For higher resolutions, generate at maximum supported resolution and upscale using dedicated upscaling models.

Hardware Requirements

What’s the minimum GPU for each Flux model?

ModelMinimum VRAMRecommended VRAMNotes
Flux ProN/A (API only)N/ACloud-based
Flux Dev12GB (with FP8)24GBFP8 quantization required for 12GB
Flux Schnell8GB (with FP8)16GBFastest on 24GB+

Can I run Flux on CPU only?

Technically possible for Flux Schnell (1-step generation) but impractical — generation takes 5-10 minutes per image vs. 0.5 seconds on GPU. Not recommended for any production or regular use.

VRAM optimization techniques

If your GPU has limited VRAM:

  1. FP8 quantization: Reduces model precision, halving VRAM requirements with minimal quality loss
  2. CPU offloading: Moves inactive model components to RAM, reducing peak VRAM usage
  3. Attention slicing: Processes attention in chunks rather than all at once
  4. Tiled VAE: Decodes the image in tiles rather than all at once

ComfyUI supports all these optimizations through node configurations.

Output Specifications

What output formats does Flux support?

The model generates raw pixel data. Output format depends on your pipeline:

  • PNG: Lossless, largest file size (~2-5MB per 1024×1024)
  • JPEG: Lossy compression, smallest file size (~200-500KB)
  • WebP: Good balance of quality and size (~300-800KB)
  • BMP: Uncompressed, very large
  • TIFF: Lossless, suitable for professional workflows
ModelMinimumRecommendedMaximum (diminishing returns)
Flux ProN/A (API handles)N/AN/A
Flux Dev2030-5050
Flux Schnell144

What guidance scale should I use?

Flux uses a different guidance mechanism than Stable Diffusion. Recommended values:

  • Flux Dev: 3.0-4.0 (default 3.5)
  • Flux Schnell: 0 (distilled model, guidance not applicable)

Higher guidance produces more prompt-adherent but potentially less natural images. Lower guidance produces more varied but potentially less accurate results.

Troubleshooting

Generated images look blurry or low-quality

  1. Increase inference steps (try 40-50 for Flux Dev)
  2. Check resolution — ensure you’re generating at 1024×1024 or higher
  3. Verify FP8 quantization isn’t over-compressing (try FP16 if VRAM allows)
  4. Update to the latest model weights

Text in images is garbled or incorrect

  1. Keep text short (1-5 words is most reliable)
  2. Put text in quotes within the prompt: a sign reading "OPEN"
  3. Use Flux Pro for best text rendering
  4. Generate multiple candidates and select the best text rendering
  5. Consider adding text in post-processing for critical applications

Generation is very slow

  1. Verify GPU is being used (check nvidia-smi during generation)
  2. Reduce inference steps
  3. Lower resolution
  4. Close other GPU-consuming applications
  5. Use Flux Schnell for speed-critical applications
  6. Ensure you have the latest CUDA drivers

Out of memory errors

  1. Enable FP8 quantization
  2. Enable CPU offloading
  3. Enable attention slicing
  4. Reduce resolution
  5. Use Flux Schnell (lower VRAM requirements)
  6. Close other applications using GPU memory

References