Asymptotically Fast Clebsch-Gordan Tensor Products with Vector Spherical Harmonics
Authors
YuQing Xie, Ameya Daigavane, Mit Kotak, Tess Smidt
Abstract
$E(3)$-equivariant neural networks have proven to be effective in a wide range of 3D modeling tasks. A fundamental operation of such networks is the tensor product, which allows interaction between different feature types. Because this operation scales poorly, there has been considerable work towards accelerating this interaction. However, recently \citet{xieprice} have pointed out that most speedups come from a reduction in expressivity rather than true algorithmic improvements on computing Clebsch-Gordan tensor products. A modification of Gaunt tensor product \citep{gaunt} can give a true asymptotic speedup but is incomplete and misses many interactions. In this work, we provide the first complete algorithm which truly provides asymptotic benefits Clebsch-Gordan tensor products. For full CGTP, our algorithm brings runtime complexity from the naive $O(L^6)$ to $O(L^4\log^2 L)$, close to the lower bound of $O(L^4)$. We first show how generalizing fast Fourier based convolution naturally leads to the previously proposed Gaunt tensor product \citep{gaunt}. To remedy antisymmetry issues, we generalize from scalar signals to irrep valued signals, giving us tensor spherical harmonics. We prove a generalized Gaunt formula for the tensor harmonics. Finally, we show that we only need up to vector valued signals to recover the missing interactions of Gaunt tensor product.
Concepts
The Big Picture
Imagine trying to describe the shape of a molecule: not just which atoms are where, but how it would look from every possible angle. To do that faithfully, a neural network needs to “speak the language” of rotations and reflections, treating space the way physics actually works.
E(3)-equivariant neural networks are AI architectures built to respect the symmetries of 3D space (rotations, translations, and reflections). They’re effective for molecular modeling, protein structure prediction, and materials discovery.
But there’s a bottleneck in the math. At the heart of these networks sits the Clebsch-Gordan tensor product (CGTP), a procedure that lets features at different levels of geometric detail interact with each other. It’s slow. In the standard formulation, computational cost scales as O(L⁶), where L is the maximum level of detail the network captures. Double L, and you wait 64 times longer.
For years, researchers tried to speed this up. A study by Xie and Price showed that most proposed shortcuts were quietly discarding information rather than genuinely computing faster.
A team from MIT now has an answer. Their algorithm delivers true speedups without sacrificing expressive power, reducing runtime from O(L⁶) to O(L⁴ log² L), close to the theoretical floor of O(L⁴).
Key Insight: This is the first complete tensor product algorithm that achieves genuine algorithmic speedup, not by throwing away interactions, but by computing all of them more efficiently. The mathematical key: vector spherical harmonics.
How It Works
Start with a signal-processing analogy. Multiplying two functions on a circle in frequency space is just pointwise multiplication; that’s the trick behind Fourier transforms. Can you do something similar on a sphere?
The answer leads to the Gaunt tensor product (GTP), an operation that projects signals onto spherical harmonics (mathematical functions describing patterns on a sphere’s surface, much like Fourier modes describe patterns on a circle), multiplies them pointwise, and transforms back. This is genuinely faster than naive CGTP.

But GTP is incomplete. It misses certain interactions. It can never simulate a cross product, for example.
The reason is subtle. Spherical harmonics come in two flavors based on how they behave under reflection: even (scalar) and odd (pseudoscalar). A pseudoscalar quantity flips sign when you hold it up to a mirror, like the difference between a left-handed and right-handed twist. Standard scalar spherical harmonics can’t encode this handedness, so GTP is blind to those interactions by construction.
The fix: upgrade from scalar-valued signals on the sphere to tensor spherical harmonics, signals that carry not just a number at each point but a small vector pointing in some direction. The authors show that vector-valued signals are sufficient to recover every interaction that scalar harmonics miss. The resulting algorithm is the Vector Signal Tensor Product (VSTP).
VSTP rests on a generalized Gaunt formula, a new identity governing how tensor spherical harmonics multiply. The classical Gaunt formula (which drives GTP) describes how scalar harmonics of degrees l₁ and l₂ decompose when multiplied. The new formula extends this to tensor harmonics, providing the coefficients VSTP needs while proving it captures all CGTP interactions.
The runtime improvement comes from fast spherical harmonic transforms:
- Naive CGTP: O(L⁶), summing over all combinations of angular momentum indices
- CGTP with sparsity: O(L⁵), exploiting known zero coefficients
- Gaunt tensor product: O(L⁴ log² L), using fast SH transforms, but incomplete
- VSTP (this work): O(L⁴ log² L), same asymptotic speed, now complete

Vector-valued signals add just enough structure to cover the interactions scalar harmonics miss, without blowing up the runtime. The authors prove that vector signals are sufficient. You don’t need rank-2 tensors or higher, which keeps the algorithm lean.
Why It Matters
Equivariant networks are scaling up. Models like NequIP, MACE, and Equiformer now handle systems with thousands of atoms: protein folding, catalyst discovery, quantum property prediction. At those scales, the O(L⁶) bottleneck is real.
Capturing finer-grained physical interactions has meant paying steep costs in compute time. L⁶ growth is punishing. VSTP removes that barrier without compromise.
Because VSTP extends GTP to a complete tensor product while keeping the same asymptotic cost, existing pipelines can adopt it without major architectural changes. The results also generalize: the authors frame their work in terms of generalized Fourier transforms, which points toward equivariant networks built on other symmetry groups. In particle physics or crystallography, where different symmetries govern the problem, similar algorithmic strategies may apply.

Bottom Line: By upgrading from scalar to vector spherical harmonics, this paper delivers the first algorithm that computes Clebsch-Gordan tensor products both completely and asymptotically fast, bringing a core bottleneck of 3D equivariant AI from O(L⁶) to O(L⁴ log² L) without losing a single interaction.
IAIFI Research Highlights
This work takes representation theory from quantum mechanics (Clebsch-Gordan coefficients, spherical harmonics, group Fourier transforms) and applies it to a computational bottleneck in modern AI. It's a clean example of how physics-informed thinking can produce algorithmic breakthroughs.
The Vector Signal Tensor Product is the first provably complete and asymptotically fast tensor product for E(3)-equivariant neural networks, letting them scale to higher angular frequencies without the runtime penalty that has held back current models.
The generalized Gaunt formula for tensor spherical harmonics extends a classical result in mathematical physics. It may find independent use in quantum chemistry, atomic physics, and other fields where angular momentum coupling is central.
Future work may extend VSTP to other compact Lie groups and integrate it into hardware-optimized kernels like cuEquivariance and openEquivariance. The preprint is available as [arXiv:2602.21466](https://arxiv.org/abs/2602.21466) from the Smidt group at MIT.
Original Paper Details
Asymptotically Fast Clebsch-Gordan Tensor Products with Vector Spherical Harmonics
[arXiv:2602.21466](https://arxiv.org/abs/2602.21466)
YuQing Xie, Ameya Daigavane, Mit Kotak, Tess Smidt