flux-fp8

flux-fp8

Kijai

Specialized float8 quantized version of FLUX models, offering float8_e4m3fn and float8_e5m2 weight variants for optimized performance while maintaining quality

PropertyValue
AuthorKijai
LicenseMixed (FLUX.1 Non-Commercial / Apache-2.0)
Base ModelsFLUX.1-dev, FLUX.1-schnell, FLUX.1-dev-ControlNet-Union-Pro

What is flux-fp8?

flux-fp8 is a specialized quantized version of various FLUX models, implementing float8 precision to optimize model performance while maintaining quality. It provides both float8_e4m3fn and float8_e5m2 weight variants of popular FLUX models.

Implementation Details

The model implements two specific float8 quantization schemes: float8_e4m3fn and float8_e5m2. These implementations are available for multiple base models including FLUX.1-dev, FLUX.1-schnell, and FLUX.1-dev-ControlNet-Union-Pro.

  • float8_e4m3fn quantization for optimal precision-performance balance
  • float8_e5m2 variant available for specific use cases
  • Maintained compatibility with original model architectures

Core Capabilities

  • Reduced model size while preserving functionality
  • Optimized memory usage through float8 quantization
  • Support for multiple FLUX model variants
  • Compatibility with existing FLUX model pipelines

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its specialized float8 quantization approach, offering both e4m3fn and e5m2 variants that maintain model quality while significantly reducing memory footprint.

Q: What are the recommended use cases?

The model is ideal for scenarios where memory efficiency is crucial but model performance cannot be compromised. It's particularly useful in resource-constrained environments while maintaining FLUX model capabilities.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026