<RETURN_TO_BASE

Generalization in Flow Matching Models Driven by Velocity Field Approximation, Not Noise

New research reveals that generalization in flow matching models emerges from the neural network's approximation of velocity fields during early trajectories, challenging the belief that stochasticity drives this process.

Understanding Generalization in Deep Generative Models

Deep generative models like diffusion and flow matching have achieved impressive results in generating realistic and diverse content across images, audio, video, and text. However, understanding how these models generalize—whether they truly learn to produce new data or merely memorize training examples—remains a core challenge. Some studies indicate large diffusion models memorize training samples, while others observe clear generalization when trained on extensive datasets, suggesting a phase shift between memorization and generalization.

Insights from Existing Research on Flow Matching

Prior work has explored closed-form solutions, memorization vs. generalization dynamics, and phases in generative processes. Techniques such as closed-form velocity field regression and smoothed optimal velocity generation have been proposed. Geometric interpretations link transition to generalization with dataset size, and temporal analyses reveal distinct phases dependent on data dimension and sample count. However, validation approaches relying on stochastic backward processes do not apply to flow matching models, leaving gaps in understanding their generalization mechanisms.

New Discoveries: Early Trajectory Failures as the Source of Generalization

Researchers from Université Jean Monnet Saint-Etienne and Université Claude Bernard Lyon demonstrate that generalization in flow matching models arises primarily due to limitations in neural networks approximating the exact velocity field during critical early and late time intervals. This phenomenon corresponds to a transition from stochastic to deterministic behavior in flow matching trajectories. The team proposed a learning algorithm that explicitly regresses against the exact velocity field, resulting in improved generalization on standard image datasets.

Investigating Generalization Sources in Detail

The study challenges assumptions about target stochasticity by employing closed-form optimal velocity field formulations, showing that after small time intervals, the weighted average of conditional flow matching targets equals single expectation values. Experiments on subsampled CIFAR-10 datasets ranging from 10 to 10,000 samples analyze approximation quality between learned and optimal velocity fields. Additionally, hybrid models combining piecewise trajectories governed by optimal velocity fields early and learned velocity fields later, with adjustable threshold parameters, help pinpoint critical periods influencing generalization.

Empirical Flow Matching: Learning with Deterministic Targets

The researchers implemented a learning algorithm regressing against more deterministic targets derived from closed-form formulas. They compared vanilla conditional flow matching, optimal transport flow matching, and empirical flow matching on CIFAR-10 and CelebA datasets. Multiple samples were used to estimate empirical means, and evaluation metrics included Fréchet Inception Distance with Inception-V3 and DINOv2 embeddings for unbiased assessment. The computational complexity was O(M × |B| × d). Increasing sample count M for empirical mean estimation produced less stochastic targets, enhancing stability with modest extra computation when M matched batch size.

Velocity Field Approximation: The Key to Generalization

This research refutes the idea that stochasticity in loss functions is the primary driver of generalization in flow matching models. Instead, the ability—or failure—to closely approximate the exact velocity field, particularly during early trajectory phases, plays the central role. While empirical findings advance understanding of learned models, fully characterizing velocity fields beyond optimal trajectories remains a challenge, inviting future work incorporating architectural inductive biases.

Ethical Considerations

The improved understanding of generative models' generalization has broader implications, including concerns about misuse for deepfakes, privacy breaches, and synthetic content creation. Ethical use of these advanced generative systems must be carefully considered to mitigate potential harms.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский