Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install hiyenwong-ai-collection-collection-skills-diffusion-distribution-matchinggit clone https://github.com/hiyenwong/ai_collection.gitcp ai_collection/SKILL.MD ~/.claude/skills/hiyenwong-ai-collection-collection-skills-diffusion-distribution-matching/SKILL.md--- name: diffusion-distribution-matching description: Accelerate diffusion models using continuous-time distribution matching distillation. Analyzes DMD and consistency distillation paradigms for few-step generation. --- # Diffusion Distribution Matching Distillation ## Description Techniques for accelerating diffusion models through distribution matching distillation (DMD) and continuous-time consistency methods. Addresses the limitations of discrete-time DMD (visual artifacts, over-smoothing) by formulating continuous-time distribution matching along the probability flow ODE trajectory. Based on arXiv:2605.06376 "Continuous-Time Distribution Matching for Few-Step Diffusion Distillation". ## Activation Keywords - diffusion distillation - distribution matching distillation - DMD acceleration - consistency distillation - few-step diffusion - continuous-time distribution matching - diffusion model acceleration ## Instructions for Agents ### Step 1: Understand Distillation Paradigms Two main approaches for diffusion acceleration: - **Distribution Matching Distillation (DMD)**: Matches distribution at discrete timesteps - **Consistency Distillation**: Enforces self-consistency along full PF-ODE trajectory ### Step 2: Identify Limitations of Discrete-Time DMD - Sparse supervision at few predefined discrete timesteps - Mode-seeking nature of reverse KL divergence - Visual artifacts and over-smoothed outputs - Often requires complementary techniques (GAN loss, score distillation) ### Step 3: Formulate Continuous-Time Matching - Replace discrete timestep matching with continuous-time formulation - Match distributions along the entire PF-ODE trajectory - This provides denser supervision and avoids mode-seeking issues ### Step 4: Implementation Considerations - Parameterize the student model for few-step generation - Use continuous-time gradient flow for distribution matching - Balance between fidelity and generation speed ### Step 5: Evaluate Trade-offs - Generation steps: fewer steps = faster but potentially lower quality - Distribution fidelity: how well the student matches the teacher - Training complexity: continuous-time vs discrete-time ## Key Concepts - **PF-ODE (Probability Flow ODE)**: The continuous-time formulation of the diffusion process - **Reverse KL Divergence**: Mode-seeking divergence that can cause over-smoothing - **Self-Consistency**: Property that the model produces consistent outputs along the generation trajectory - **Distribution Matching**: Aligning the student's output distribution with the teacher's ## Best Practices 1. Continuous-time formulation provides more robust training than discrete-time 2. Avoid relying solely on reverse KL divergence — consider forward KL or Wasserstein 3. Few-step models should be evaluated on both quality and diversity metrics 4. Consider hybrid approaches combining DMD with adversarial training ## Related Skills - physics-guided-neural-networks: For physics-informed generative models - physics-infused-video-generation: For generative video applications