← Back to Timeline

Poisson Flow Generative Models

Foundational AI

Authors

Yilun Xu, Ziming Liu, Max Tegmark, Tommi Jaakkola

Abstract

We propose a new "Poisson flow" generative model (PFGM) that maps a uniform distribution on a high-dimensional hemisphere into any data distribution. We interpret the data points as electrical charges on the $z=0$ hyperplane in a space augmented with an additional dimension $z$, generating a high-dimensional electric field (the gradient of the solution to Poisson equation). We prove that if these charges flow upward along electric field lines, their initial distribution in the $z=0$ plane transforms into a distribution on the hemisphere of radius $r$ that becomes uniform in the $r \to\infty$ limit. To learn the bijective transformation, we estimate the normalized field in the augmented space. For sampling, we devise a backward ODE that is anchored by the physically meaningful additional dimension: the samples hit the unaugmented data manifold when the $z$ reaches zero. Experimentally, PFGM achieves current state-of-the-art performance among the normalizing flow models on CIFAR-10, with an Inception score of $9.68$ and a FID score of $2.35$. It also performs on par with the state-of-the-art SDE approaches while offering $10\times $ to $20 \times$ acceleration on image generation tasks. Additionally, PFGM appears more tolerant of estimation errors on a weaker network architecture and robust to the step size in the Euler method. The code is available at https://github.com/Newbeeer/poisson_flow .

Concepts

generative models poisson field normalizing flows augmented dimension sampling physics-informed neural networks score-based models stochastic processes diffusion models density estimation inverse problems optimal transport

The Big Picture

Imagine dropping ink into honey. Unlike water, where ink diffuses chaotically, the thick fluid forces it along clean, predictable paths. Now run that process in reverse: start with ink spread uniformly through the honey and watch it concentrate back into a single drop. That’s roughly what MIT researchers Yilun Xu, Ziming Liu, Max Tegmark, and Tommi Jaakkola have built—a generative model grounded in electrostatic field math that synthesizes realistic images faster than nearly any competing approach.

The challenge at the heart of generative AI is simple to state: how do you teach a computer to draw? Current best-in-class methods, diffusion models and score-based models, learn to reverse a noisy scrambling process, gradually recovering coherent images from random static. They produce beautiful results, but they’re slow. Generating a single image can require hundreds of sequential neural network evaluations.

The PFGM team cracked open a textbook on electrostatics and found a better way.

Key Insight: By treating image data as electric charges generating a high-dimensional Poisson field, PFGM creates a smooth, invertible path between noise and data. It achieves the best image quality among normalizing flow models while running 10–20× faster than competing stochastic methods.

How It Works

The trick starts with a geometric insight from classical physics. Take your dataset—say 50,000 images of cats and airplanes from CIFAR-10. Each image lives in a high-dimensional space (3,072 dimensions for CIFAR-10). Now place each image as a positive electric charge on a flat hyperplane embedded in a space with one extra dimension.

Figure 1

Physics takes over from here. Each charge repels every other and gets pushed upward along electric field lines, the invisible paths a test charge would follow, like the curves you see when you scatter iron filings near a magnet. These trajectories are governed by the Poisson equation, a 19th-century partial differential equation that describes electrostatics, gravity, and fluid flow. Here it gets pressed into service to move charges through the extra dimension.

As the charges drift upward, they spread out. The authors prove that when charges cross an imaginary hemisphere of large radius r, their distribution becomes perfectly uniform in the limit of large r. Every complicated, structured data distribution—faces, galaxies, handwritten digits—flattens into the same featureless sphere.

That’s the forward process. Generating new images runs it in reverse:

  1. Sample a random point uniformly from the high-dimensional hemisphere. This is trivially easy.
  2. Integrate backwards along the electric field lines using a backward ODE (ordinary differential equation), moving from large z back toward z=0.
  3. Stop when z hits zero. The point that lands on the hyperplane is your generated image.

Figure 2

The Poisson field can’t be computed analytically at test time, so the team trains a neural network to estimate it. They normalize field vectors to unit length during training, working with direction rather than magnitude. This turns out to be essential for stability. The network learns to predict which direction field lines point at any location, which is enough to trace the backward path from hemisphere to data.

Figure 3

The physical anchoring also buys error tolerance. Score-based approaches tie the “time” variable closely to sample norm, so small estimation errors compound in a correlated way. PFGM’s z-coordinate and sample norm are more loosely coupled, and integration errors don’t cascade as badly.

Why It Matters

On the standard CIFAR-10 benchmark, PFGM achieves an FID score of 2.35 (Fréchet Inception Distance; lower means generated images match real ones more closely) and an Inception score of 9.68 (higher means greater diversity and realism). Among normalizing flow models, that’s the best reported result. PFGM also performs on par with the best SDE-based generators while requiring 10× to 20× fewer neural network evaluations. That’s not a marginal speedup; it’s the difference between seconds and minutes at scale.

The model is unusually forgiving. When the researchers deliberately swapped in a weaker neural architecture—one that causes competing ODE methods to produce garbage—PFGM kept generating decent images. Step size is similarly tolerant: whether you use 10 or 100 integration steps, performance degrades gracefully rather than catastrophically. For practical deployment, where speed-quality tradeoffs get tuned on the fly, that kind of flexibility counts for a lot.

The exchange between physics and AI here runs both ways. A 19th-century equation produces a new machine learning method, but the correspondence raises questions for physics too: why does Poisson flow uniformize distributions so cleanly? What do those field lines tell us about the geometry of data manifolds?

Because the ODEs are invertible, the framework could also support likelihood evaluation and image editing, tasks where pure diffusion models struggle.

Bottom Line: Poisson Flow Generative Models show that an equation written down by Siméon Denis Poisson in 1823 can anchor a 2022 image generator that is faster and more error-tolerant than most of its competitors.

IAIFI Research Highlights

Interdisciplinary Research Achievement
PFGM translates the mathematics of high-dimensional electrostatics (Poisson's equation and *N*-dimensional Coulomb's law) into a practical generative modeling framework. Classical physics intuition, applied carefully, can produce real algorithmic gains in AI.
Impact on Artificial Intelligence
On CIFAR-10, PFGM sets the best FID score (2.35) among normalizing flow models and delivers 10–20× sampling speedups over SDE-based methods, with improved tolerance to network architecture choices and integration step size.
Impact on Fundamental Interactions
The work extends classical electrostatic theory to arbitrary *N* dimensions, proving new geometric results about how charge distributions transform under Poisson field flow. These mathematical physics results hold independently of the machine learning application.
Outlook and References
The approach could extend to higher-dimensional data and other physical flow equations. The paper is [arXiv:2209.11178](https://arxiv.org/abs/2209.11178), presented at NeurIPS 2022.

Original Paper Details

Title
Poisson Flow Generative Models
arXiv ID
[2209.11178](https://arxiv.org/abs/2209.11178)
Authors
Yilun Xu, Ziming Liu, Max Tegmark, Tommi Jaakkola
Abstract
We propose a new "Poisson flow" generative model (PFGM) that maps a uniform distribution on a high-dimensional hemisphere into any data distribution. We interpret the data points as electrical charges on the $z=0$ hyperplane in a space augmented with an additional dimension $z$, generating a high-dimensional electric field (the gradient of the solution to Poisson equation). We prove that if these charges flow upward along electric field lines, their initial distribution in the $z=0$ plane transforms into a distribution on the hemisphere of radius $r$ that becomes uniform in the $r \to\infty$ limit. To learn the bijective transformation, we estimate the normalized field in the augmented space. For sampling, we devise a backward ODE that is anchored by the physically meaningful additional dimension: the samples hit the unaugmented data manifold when the $z$ reaches zero. Experimentally, PFGM achieves current state-of-the-art performance among the normalizing flow models on CIFAR-10, with an Inception score of $9.68$ and a FID score of $2.35$. It also performs on par with the state-of-the-art SDE approaches while offering $10\times $ to $20 \times$ acceleration on image generation tasks. Additionally, PFGM appears more tolerant of estimation errors on a weaker network architecture and robust to the step size in the Euler method. The code is available at https://github.com/Newbeeer/poisson_flow .