Project Page

Sinkhorn-Drifting Generative Models

Drifting toward the target is easy. The hard part is ensuring the dynamics cannot stop short of it.

1Vanderbilt University 2University of California, Davis 3Purdue University

Corresponding Authors

All panels use the same particles and targets; only the temperature τ changes. As τ decreases, heuristic normalizations drift into instability and mode collapse, while Sinkhorn preserves coherent transport throughout.

Particle Drift Trajectories (N = 100, steps = 75). Each GIF also displays the current step.

Source Current Target Trajectory

Animated drift trajectories for tau equals 0.01 across all five normalization panels.
τ = 0.01. Extreme low-temperature: heuristic normalization deteriorates sharply, while Sinkhorn remains the most stable.
Animated drift trajectories for tau equals 0.1 across all five normalization panels.
τ = 0.1. Low-temperature breakdown begins: masked baselines turn erratic, while Sinkhorn stays organized.
Animated drift trajectories for tau equals 1 across all five normalization panels.
τ = 1. Intermediate: differences are mild, but Sinkhorn is already the cleanest transport field..
Animated drift trajectories for tau equals 10 across all five normalization panels.
τ = 10. Large-temperature: all methods remain comparatively well behaved.

A drifting generator is driven by two forces: attraction toward real data and repulsion from itself. Under heuristic row-normalized couplings, these forces can cancel before the model has matched the data, leading to low-temperature collapse. We connect this dynamics to a Sinkhorn-divergence gradient flow and show that Sinkhorn scaling provides the right fix: the forces balance only when the generated distribution matches the data distribution.

We establish a theoretical link between the recently proposed “drifting” generative dynamics and gradient flows induced by the Sinkhorn divergence. In a particle discretization, the drift field admits a cross-minus-self decomposition: an attractive term toward the target distribution and a repulsive/self-correction term toward the current model, both expressed via one-sided normalized Gibbs kernels. We show that Sinkhorn divergence yields an analogous cross-minus-self structure, but with each term defined by entropic optimal-transport couplings obtained through two-sided Sinkhorn scaling (i.e., enforcing both marginals). This provides a precise sense in which drifting acts as a surrogate for a Sinkhorn-divergence gradient flow, interpolating between one-sided normalization and full two-sided Sinkhorn scaling. Crucially, this connection resolves an identifiability gap in prior drifting formulations: leveraging the definiteness of the Sinkhorn divergence, we show that zero drift (equilibrium of the dynamics) implies that the model and target measures match. Experiments show that Sinkhorn drifting reduces sensitivity to kernel temperature and improves one-step generative quality, trading off additional training time for a more stable optimization, without altering the inference procedure used by drift methods. These theoretical gains translate to strong low-temperature improvements in practice: on FFHQ-ALAE at the lowest temperature setting we evaluate, Sinkhorn drifting reduces mean FID from 187.7 to 37.1 and mean latent EMD from 453.3 to 144.4, while on MNIST it preserves full class coverage across the temperature sweep.

Keywords: Drifting · Optimal Transport · One-Step Generative Models

Part 2

FFHQ-ALAE

Each row is a class; Baseline is on the left and Sinkhorn is on the right.

FFHQ-ALAE qualitative comparison at tau 0.1.
τ = 0.1. Low-temperature generation is where the gap is clearest: the baseline collapses toward repetitive faces, while Sinkhorn remains diverse and class-faithful.
τ Baseline FID Sinkhorn FID Baseline EMD Sinkhorn EMD
0.1 187.7 37.1 453.3 144.4
1.0 146.3 33.7 370.5 137.6
10.0 45.3 33.9 156.4 138.5

Mean class-conditional FFHQ results across all six classes. Sinkhorn improves both image FID and latent EMD at every temperature.

BibTeX

@misc{he2026sinkhorndriftinggenerativemodels,
  title         = {Sinkhorn-Drifting Generative Models},
  author        = {Ping He and Om Khangaonkar and Hamed Pirsiavash and Yikun Bai and Soheil Kolouri},
  year          = {2026},
  eprint        = {2603.12366},
  archivePrefix = {arXiv},
  primaryClass  = {cs.LG},
  url           = {https://arxiv.org/abs/2603.12366}
}