Equilibrium Matching (EqM) learns a time-invariant gradient field that is compatible with an underlying energy function,
eliminating time/noise conditioning and fixed-horizon integrators.
Conceptually, EqM’s gradient vanishes on the data manifold and increases toward noise,
yielding an equilibrium landscape in which ground-truth samples are stationary points.
Flow Matching learns a varying velocity that only converges to ground truths at the final timestep,
whereas EqM learns a time-invariant gradient landscape that always converges to ground-truth data points.
To train an Equilibrium Matching model, we aim to construct an energy landscape in which the target gradient at ground-truth samples is zero.
To do so, we first define a corruption scheme that provides a transition between data and noise.
Our training obective aims to match a target gradient at these intermediate samples, constructing an implicit energy landscape with
gradient direction pointing from noise to data.
Because of its equilibrium nature, Equilibrium Matching generates samples via optimization on the learned landscape.
In contrast to diffusion/flow models that integrate over a fixed time horizon, EqM decouples sample quality from a prescribed trajectory.
It formulates the sampling process as a gradient descent procedure and supports adaptive step sizes, optimizers, and compute,
offering additional flexibility at inference time.