flux-dev-de-distill

flux-dev-de-distill

nyanko7

An experimental AI model that de-distills guidance from FLUX.1-dev, implementing true classifier-free guidance through reversed distillation process.

PropertyValue
LicenseMIT
Base Modelblack-forest-labs/FLUX.1-dev
Research PaperOn Distillation of Guided Diffusion Models
LanguageEnglish

What is flux-dev-de-distill?

flux-dev-de-distill is an experimental AI model that aims to reverse the distillation process of guidance from FLUX.1-dev. Created by nyanko7, this model implements true classifier-free guidance (CFG) by removing the original distilled guidance and reconstructing it from scratch.

Implementation Details

The model follows Algorithm 1 from the paper "On Distillation of Guided Diffusion Models," implementing a student model x(zt) that matches the teacher's output at various time-steps and guidance scales. The training process involved 150K Unsplash images at 1024px resolution, running for 6,000 steps with a global batch size of 32. The training took approximately 12 hours with limited computational resources.

  • Student model initialized with teacher parameters except for w-embedding
  • Guidance scale matching in range w ∈ [1, 4]
  • Uses true CFG instead of distilled CFG
  • Requires custom inference script due to diffusers pipeline incompatibility

Core Capabilities

  • True classifier-free guidance implementation
  • High-resolution image processing (1024px)
  • Flexible guidance scale control
  • Compatible with custom inference implementations

Frequently Asked Questions

Q: What makes this model unique?

This model's uniqueness lies in its approach to de-distilling guidance, effectively reversing the traditional distillation process to achieve true classifier-free guidance. This makes it particularly interesting for researchers and developers working on diffusion models.

Q: What are the recommended use cases?

The model is best suited for research purposes and applications requiring true classifier-free guidance. It's particularly useful for those interested in studying the differences between distilled and true CFG implementations in diffusion models.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026