Zubnet AI學習Wiki › Flow Matching
Models

Flow Matching

Continuous Normalizing Flows, Rectified Flow
一種生成式建模技術,透過在雜訊分佈和資料分佈之間遵循平滑、直接的路徑,學習把雜訊變成資料。不像擴散模型透過固定的加噪過程在多步中加噪除噪,flow matching 學習直線軌跡,能在更少步數裡遍歷,使生成更快。

為什麼重要

Flow matching 在最前沿的影像和影片生成中正快速取代傳統擴散。Flux(Black Forest Labs)、Stable Diffusion 3 和好幾個影片模型都用 flow matching。實際好處:更少生成步數下達到可比甚至更好的品質,這直接意味著更快的推理和更低的成本。

Deep Dive

The intuition: imagine every point in "noise space" connected to a point in "image space" by a straight line. Flow matching trains a neural network to predict the velocity (direction and speed) along these paths at any point. To generate an image, you start from a random noise point and follow the velocity field to arrive at a clean image. The straighter the paths, the fewer steps you need — this is why "rectified flows" (which straighten the paths) are important.

Diffusion vs. Flow Matching

Traditional diffusion models define a fixed forward process (gradually adding Gaussian noise) and learn the reverse process (denoising). The forward process is curved through high-dimensional space, requiring many small steps to reverse accurately (typically 20–50 steps). Flow matching learns more direct paths, often achieving equivalent quality in 4–10 steps. Some formulations (like consistency models) push this to a single step, though with some quality trade-off.

The Unified View

Mathematically, diffusion models and flow matching are both instances of continuous-time generative models — they differ in the probability paths they define between noise and data. This unified perspective is helping researchers design better training objectives and architectures that combine insights from both. The practical implication: the distinction between "diffusion model" and "flow matching model" is becoming more about training methodology than fundamental architecture.

相關概念

← 所有術語
← FLOPs Foundation Model →