← Back to Timeline

Machine learning electronic structure and atomistic properties from the external potential

Foundational AI

Authors

Jigyasa Nigam, Tess Smidt, Geneviève Dusson

Abstract

Electronic structure calculations remain a major bottleneck in atomistic simulations and, not surprisingly, have attracted significant attention in machine learning (ML). Most existing approaches learn a direct map from molecular geometries, typically represented as graphs or encoded local environments, to molecular properties or use ML as a surrogate for electronic structure theory by targeting quantities such as Fock or density matrices expressed in an atomic orbital (AO) basis. Inspired by the Hohenberg-Kohn theorem, in this work, we propose an operator-centered framework in which the external (nuclear) potential, expressed in an AO basis, serves as the model input. From this operator, we construct hierarchical, body-ordered representations of atomic configurations that closely mirror the principles underlying several popular atom-centered descriptors. At the same time, the matrix-valued nature of the external potential provides a natural connection to equivariant message-passing neural networks. In particular, we show that successive products of the external potential provide a scalable route to equivariant message passing and enable an efficient description of long-range effects. We demonstrate that this approach can be used to model molecular properties, such as energies and dipole moments, from the external potential, or learn effective operator-to-operator maps, including mappings to the Fock matrix and the reduced density matrix from which multiple molecular observables can be simultaneously derived.

Concepts

external potential representation equivariant neural networks operator-to-operator learning representation learning symmetry preservation surrogate modeling body-ordered features graph neural networks molecular dynamics multi-task learning inverse problems kernel methods

The Big Picture

Imagine predicting how a molecule will behave without running expensive quantum chemistry calculations. Will it react? What color light does it absorb? How might it bind to a drug target? Most ML approaches treat molecules as graphs or point clouds and learn to map geometry to properties. But quantum mechanics has been hinting at a smarter starting point for decades.

That starting point is the external potential, the electrostatic pull that atomic nuclei exert on electrons. In the 1960s, Hohenberg and Kohn proved that this potential alone contains everything needed to determine a molecule’s lowest-energy state. Every energy level, every reaction rate, every optical property is encoded in how the nuclei push and pull on the electron cloud. A team at MIT and the Université Marie et Louis Pasteur has now built a framework that takes this theorem seriously as an engineering principle, not just a footnote in a textbook.

Jigyasa Nigam, Tess Smidt, and Geneviève Dusson treat the external potential, encoded as a matrix capturing how every nucleus influences the electrons, as the fundamental input to their ML models (arXiv:2602.15345). The payoff is a single framework that predicts molecular properties, learns quantum mechanical operators, and respects the symmetries that physics demands.

Key Insight: Grounding molecular ML in the Hohenberg-Kohn theorem and using the external potential matrix as input gives the framework both the formal backing of density functional theory and the flexibility to learn any quantum mechanical observable.

How It Works

The core move is a change of representation. Instead of encoding a molecule as a graph of atoms and bonds, the team encodes it as a matrix V: the external potential matrix expressed in an atomic orbital (AO) basis. These are standard mathematical functions centered on each atom that describe electron distributions. Each matrix element captures how strongly two basis functions “feel” the nuclear attraction from every atom, producing a dense fingerprint of the entire electrostatic environment.

Figure 1

From this matrix, the team builds increasingly complex features through matrix multiplication. Each successive power of V captures higher-order correlations (two-body, three-body, four-body atomic interactions), mirroring the body-ordered atom-centered descriptors that are standard currency in molecular ML. SOAP (Smooth Overlap of Atomic Positions) and ACE (Atomic Cluster Expansion), the two most widely used descriptor frameworks, are built on exactly these ideas. The matrix structure of V generates all of this naturally, without separate engineering of higher-order terms.

This matrix multiplication route also connects to equivariant message-passing neural networks (MPNNs), the current state of the art for molecular ML. “Equivariant” here means the network’s outputs transform correctly under rotation or reflection: rotate the input molecule, and the predicted properties rotate with it. The team shows that multiplying external potential matrices achieves the same information propagation as MPNN message-passing, with each matrix product carrying atomic interactions one bond-length further. Rotational symmetry comes along for free through the mathematical structure of the AO basis.

This yields two model classes:

  • Op2Prop (Operator-to-Property): predicts scalar or vector quantities like energies and dipole moments directly from V
  • Op2Op (Operator-to-Operator): learns to map V to other quantum mechanical operators, such as the Fock matrix or the reduced density matrix

Op2Op is the more versatile of the two. The Fock matrix encodes how each electron effectively experiences both the nuclei and the averaged influence of all other electrons; it sits at the center of most quantum chemistry calculations. The reduced density matrix encodes how electrons are distributed and correlated across the molecule. Once you have a predicted Fock matrix, you can diagonalize it to get orbital energies and derive multiple molecular observables at once, all from one model.

The team also introduces an effective Op2Op mode, where the model is supervised on physical observables like eigenvalues rather than matrix elements directly. This makes it possible to learn from high-accuracy reference data without needing matched-basis calculations.

Figure 2

Long-range interactions are a persistent headache for molecular ML. Electrostatic effects can stretch across an entire molecule or between distant functional groups, and graph-based approaches handle this poorly because they rely on local neighborhoods. The external potential matrix sidesteps the problem by design: every matrix element involves contributions from all nuclei, so long-range Coulomb interactions are baked in from the start. Successive matrix products propagate this non-local information through the feature hierarchy without any special architectural tricks.

Why It Matters

Electronic structure ML is currently a menagerie of specialized models: some predict energies, some predict electron densities, some learn Hamiltonians. This framework offers a single input representation, the external potential matrix, from which all of these targets can be learned in a consistent theoretical language. Models trained for one task can share representations with models trained for another, and the Hohenberg-Kohn theorem provides formal guarantees about what the models can and cannot learn. That kind of unification has been missing.

Figure 3

The connection to equivariant message-passing networks runs both ways. The external potential framework gives a physically motivated derivation of why message-passing works for molecules. It also suggests new architectural choices grounded in operator algebra rather than graph heuristics. As models scale to larger systems and more complex properties, that theoretical anchor should pay off.

Open questions remain. The AO basis used to discretize the input is a tunable hyperparameter, and choosing it wisely (balancing resolution against computational cost) is an art the paper begins to formalize but doesn’t fully resolve. Scaling to very large molecules, where the full external potential matrix becomes expensive to compute and store, is still unsolved. And the effective Op2Op approach hints at models that could learn from experimental data or high-level theory without requiring matched-basis reference calculations, but that direction is mostly unexplored.

Bottom Line: By treating the external potential matrix as the fundamental input to molecular ML, this work provides a single framework that predicts both molecular properties and quantum mechanical operators, with symmetry and long-range physics built in from the start.

IAIFI Research Highlights

Title
Machine learning electronic structure and atomistic properties from the external potential
arXiv ID
2602.15345
Authors
Jigyasa Nigam, Tess Smidt, Geneviève Dusson
Abstract
Electronic structure calculations remain a major bottleneck in atomistic simulations and, not surprisingly, have attracted significant attention in machine learning (ML). Most existing approaches learn a direct map from molecular geometries, typically represented as graphs or encoded local environments, to molecular properties or use ML as a surrogate for electronic structure theory by targeting quantities such as Fock or density matrices expressed in an atomic orbital (AO) basis. Inspired by the Hohenberg-Kohn theorem, in this work, we propose an operator-centered framework in which the external (nuclear) potential, expressed in an AO basis, serves as the model input. From this operator, we construct hierarchical, body-ordered representations of atomic configurations that closely mirror the principles underlying several popular atom-centered descriptors. At the same time, the matrix-valued nature of the external potential provides a natural connection to equivariant message-passing neural networks. In particular, we show that successive products of the external potential provide a scalable route to equivariant message passing and enable an efficient description of long-range effects. We demonstrate that this approach can be used to model molecular properties, such as energies and dipole moments, from the external potential, or learn effective operator-to-operator maps, including mappings to the Fock matrix and the reduced density matrix from which multiple molecular observables can be simultaneously derived.