Paper

Decoupled DMD: CFG Augmentation as the Spear, Distribution Matching as the Shield

by Dongyang Liu ID: arxiv-paper--2511.22677

Diffusion model distillation has emerged as a powerful technique for creating efficient few-step and single-step generators. Among these, Distribution Matching Distillation (DMD) and its variants stand out for their impressive performance, which is widely attributed to their core mechanism of matchi...

High Impact - Citations
2025 Year
ArXiv Venue
Top 45% FNI Rank
Paper Information Summary
Entity Passport
Registry ID arxiv-paper--2511.22677
Provider arxiv

πŸ•ΈοΈ Neural Mesh Hub

Interconnecting Research, Data & Ecosystem

πŸ•ΈοΈ

Intelligence Hive

Multi-source Relation Matrix

Live Index
πŸ“ˆ

Momentum Index

--
🏷️

Contextual Anchors

πŸ“œ

Cite this paper

Academic & Research Attribution

BibTeX
@misc{arxiv_paper__2511.22677,
  author = {Dongyang Liu},
  title = {Decoupled DMD: CFG Augmentation as the Spear, Distribution Matching as the Shield Paper},
  year = {2026},
  howpublished = {\url{https://arxiv.org/abs/2511.22677v1}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
Dongyang Liu. (2026). Decoupled DMD: CFG Augmentation as the Spear, Distribution Matching as the Shield [Paper]. Free2AITools. https://arxiv.org/abs/2511.22677v1

πŸ”¬Technical Deep Dive

Full Specifications [+]

πŸ“ Executive Summary

"Diffusion model distillation has emerged as a powerful technique for creating efficient few-step and single-step generators. Among these, Distribution Matching Distillation (DMD) and its variants stand out for their impressive performance, which is widely attributed to their core mechanism of matching the student's output distribution to that of a pre-trained teacher model. In this work, we challenge this conventional understanding. Through a rigorous decomposition of the DMD training objecti..."

❝ Cite Node

@article{Liu2025Decoupled,
  title={Decoupled DMD: CFG Augmentation as the Spear, Distribution Matching as the Shield},
  author={Dongyang Liu and Peng Gao and David Liu and Ruoyi Du and Zhen Li and Qilong Wu and Xin Jin and Sihan Cao and Shifeng Zhang and Hongsheng Li and Steven Hoi},
  journal={arXiv preprint arXiv:arxiv-paper--2511.22677},
  year={2025}
}

πŸ‘₯ Collaborating Minds

Dongyang Liu Peng Gao David Liu Ruoyi Du Zhen Li Qilong Wu Xin Jin Sihan Cao Shifeng Zhang Hongsheng Li Steven Hoi

Paper Details

Diffusion model distillation has emerged as a powerful technique for creating efficient few-step and single-step generators. Among these, Distribution Matching Distillation (DMD) and its variants stand out for their impressive performance, which is widely attributed to their core mechanism of matching the student's output distribution to that of a pre-trained teacher model. In this work, we challenge this conventional understanding. Through a rigorous decomposition of the DMD training objective, we reveal that in complex tasks like text-to-image generation, where CFG is typically required for desirable few-step performance, the primary driver of few-step distillation is not distribution matching, but a previously overlooked component we identify as CFG Augmentation (CA). We demonstrate that this term acts as the core engine'' of distillation, while the Distribution Matching (DM) term functions as a regularizer'' that ensures training stability and mitigates artifacts. We further validate this decoupling by demonstrating that while the DM term is a highly effective regularizer, it is not unique; simpler non-parametric constraints or GAN-based objectives can serve the same stabilizing function, albeit with different trade-offs. This decoupling of labor motivates a more principled analysis of the properties of both terms, leading to a more systematic and in-depth understanding. This new understanding further enables us to propose principled modifications to the distillation process, such as decoupling the noise schedules for the engine and the regularizer, leading to further performance gains. Notably, our method has been adopted by the Z-Image ( https://github.com/Tongyi-MAI/Z-Image ) project to develop a top-tier 8-step image generation model, empirically validating the generalization and robustness of our findings.

ZEN MODE β€’ Paper Details

Diffusion model distillation has emerged as a powerful technique for creating efficient few-step and single-step generators. Among these, Distribution Matching Distillation (DMD) and its variants stand out for their impressive performance, which is widely attributed to their core mechanism of matching the student's output distribution to that of a pre-trained teacher model. In this work, we challenge this conventional understanding. Through a rigorous decomposition of the DMD training objective, we reveal that in complex tasks like text-to-image generation, where CFG is typically required for desirable few-step performance, the primary driver of few-step distillation is not distribution matching, but a previously overlooked component we identify as CFG Augmentation (CA). We demonstrate that this term acts as the core engine'' of distillation, while the Distribution Matching (DM) term functions as a regularizer'' that ensures training stability and mitigates artifacts. We further validate this decoupling by demonstrating that while the DM term is a highly effective regularizer, it is not unique; simpler non-parametric constraints or GAN-based objectives can serve the same stabilizing function, albeit with different trade-offs. This decoupling of labor motivates a more principled analysis of the properties of both terms, leading to a more systematic and in-depth understanding. This new understanding further enables us to propose principled modifications to the distillation process, such as decoupling the noise schedules for the engine and the regularizer, leading to further performance gains. Notably, our method has been adopted by the Z-Image ( https://github.com/Tongyi-MAI/Z-Image ) project to develop a top-tier 8-step image generation model, empirically validating the generalization and robustness of our findings.