Universal Inverse Distillation for Matching Models with Real-Data Supervision (No GANs)

ICLR 2026 Conference SubmissionAnonymous Authors
Diffusion modelsFlow MatchingAcceleration of diffusion/flow modelsDistillation of diffusion/flow models
Abstract:

While achieving exceptional generative quality, modern diffusion, flow, and other matching models suffer from slow inference, as they require many steps of iterative generation. Recent distillation methods address this by training efficient one-step generators under the guidance of a pre-trained teacher model. However, these methods are often constrained to only one specific framework, e.g., only to diffusion or only to flow models. Furthermore, these methods are naturally data-free, and to benefit from the usage of real data, it is required to use an additional complex adversarial training with an extra discriminator model. In this paper, we present \textbf{RealUID}, a unified distillation framework for all matching models that seamlessly incorporates real data into the distillation procedure without GANs. Our \textbf{RealUID} approach offers a simple theoretical foundation that covers previous distillation methods for Flow Matching and Diffusion models, and is also extended to their modifications, such as Bridge Matching and Stochastic Interpolants.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper proposes RealUID, a unified distillation framework that compresses multi-step matching models (diffusion, flow, bridge matching, stochastic interpolants) into efficient one-step generators while incorporating real data without adversarial training. It resides in the 'Unified Distillation Frameworks' leaf alongside one sibling paper (Distilling Generator Matching). This leaf is part of the broader 'Unified and Theoretical Distillation Frameworks' branch, which contains only three leaves and seven papers total. The positioning suggests a relatively sparse research direction focused on general-purpose, theoretically grounded distillation methods rather than model-specific or domain-specific approaches.

The taxonomy reveals that most distillation research clusters around specialized strategies: Distribution Matching (three leaves, five papers), Score Identity methods (two leaves, three papers), Adversarial/Hybrid approaches (three leaves, five papers), and Consistency/Trajectory methods (three leaves, three papers). RealUID's unified framework contrasts with these narrower paradigms by claiming applicability across diffusion, flow, and bridge matching models. The 'Unified and Theoretical Distillation Frameworks' branch sits apart from domain-specific applications (video, super-resolution, text-to-image) and specialized frameworks (multi-student, variational, autoregressive), indicating the paper targets foundational methodology rather than task-specific optimization.

Among 30 candidates examined, the contribution-level analysis shows mixed novelty signals. The 'RealUID framework' and 'real data incorporation without GANs' contributions each examined 10 candidates with zero refutable overlaps, suggesting these aspects appear relatively novel within the limited search scope. However, the 'Universal Matching loss with real data (RealUM)' contribution found one refutable candidate among 10 examined, indicating at least one prior work in the candidate pool addresses overlapping ideas. The single sibling paper in the same taxonomy leaf likely represents closely related unified distillation work, though the analysis does not specify whether it was among the refutable candidates.

Based on the limited 30-candidate search, the work appears to occupy a less-crowded research direction (unified frameworks) compared to specialized distillation methods. The claim of unifying multiple matching model families without GANs shows some novelty signals, though one contribution has identifiable prior overlap. The analysis covers top-K semantic matches and does not constitute an exhaustive literature review, so additional related work may exist beyond the examined scope.

Taxonomy

Core-task Taxonomy Papers
36
3
Claimed Contributions
30
Contribution Candidate Papers Compared
1
Refutable Paper

Research Landscape Overview

Core task: Distilling matching models into efficient one-step generators. The field addresses the challenge of compressing multi-step diffusion or flow models into fast, single-step samplers while preserving generation quality. The taxonomy reveals several complementary strategies: Distribution Matching Distillation Methods (e.g., Distribution Matching Distillation[4], Improved Distribution Matching[32]) focus on aligning student and teacher output distributions directly; Score Identity and Implicit Matching Distillation approaches (e.g., Score Identity Distillation[2], Score Implicit Matching[19]) leverage score-based objectives to guide the distillation; Adversarial and Hybrid Distillation Methods (e.g., Adversarial Distribution Matching[3], Adversarial Score Identity[20]) incorporate discriminative signals; Consistency and Flow Trajectory Distillation (e.g., Flow Trajectory Distillation[13], Self-Corrected Flow[6]) emphasize trajectory-level consistency; Specialized Distillation Frameworks target domain-specific constraints (e.g., VideoScene[7], Robust Speech Enhancement[34]); Unified and Theoretical Distillation Frameworks aim to provide general principles (e.g., Universal Inverse Distillation[0], Distilling Generator Matching[36]); and Domain-Specific Distillation Applications adapt these ideas to particular modalities. Recent work explores trade-offs between distillation objectives, computational cost, and sample quality. Distribution matching methods like DMD[4] and its variants (Phased DMD[24], Regularized Distribution Matching[25]) offer simplicity but may struggle with mode coverage, while score-based approaches (Score Identity Distillation[2], Few-step Score Identity[14]) provide stronger theoretical grounding at the expense of additional computation. Adversarial techniques (Adversarial Distribution Matching[3], Flow2GAN[23]) can sharpen outputs but risk training instability. Universal Inverse Distillation[0] sits within the Unified and Theoretical Distillation Frameworks branch alongside Distilling Generator Matching[36], emphasizing a principled, general formulation that unifies multiple distillation paradigms. Compared to specialized methods like SwiftBrush[11] or domain-targeted frameworks, Universal Inverse Distillation[0] aims for broader applicability across generative tasks, bridging theoretical insights with practical one-step generation.

Claimed Contributions

RealUID: Universal distillation framework for matching models

The authors introduce RealUID, a unified distillation framework that applies to multiple matching model families (diffusion, flow matching, bridge matching, stochastic interpolants). It unifies prior distillation methods (FGM, SiD, IBMD) under a single theoretical foundation using a linearization technique and connects them to inverse optimization.

10 retrieved papers
Natural incorporation of real data without GANs

The authors propose a method to integrate real data into the distillation procedure by modifying the universal matching loss with weighting coefficients alpha and beta. This approach avoids the architectural modifications and optimization challenges associated with adversarial training and discriminator networks.

10 retrieved papers
Universal Matching loss with real data (RealUM)

The authors define a new loss function that combines terms for both generated and real data distributions, parameterized by coefficients alpha and beta. This loss preserves the property that it yields the same teacher function when applied to real data, enabling real data incorporation while maintaining theoretical consistency.

10 retrieved papers
Can Refute

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

RealUID: Universal distillation framework for matching models

The authors introduce RealUID, a unified distillation framework that applies to multiple matching model families (diffusion, flow matching, bridge matching, stochastic interpolants). It unifies prior distillation methods (FGM, SiD, IBMD) under a single theoretical foundation using a linearization technique and connects them to inverse optimization.

Contribution

Natural incorporation of real data without GANs

The authors propose a method to integrate real data into the distillation procedure by modifying the universal matching loss with weighting coefficients alpha and beta. This approach avoids the architectural modifications and optimization challenges associated with adversarial training and discriminator networks.

Contribution

Universal Matching loss with real data (RealUM)

The authors define a new loss function that combines terms for both generated and real data distributions, parameterized by coefficients alpha and beta. This loss preserves the property that it yields the same teacher function when applied to real data, enabling real data incorporation while maintaining theoretical consistency.

Universal Inverse Distillation for Matching Models with Real-Data Supervision (No GANs) | Novelty Validation