← Back to Timeline

Multi-Symmetry Ensembles: Improving Diversity and Generalization via Opposing Symmetries

Foundational AI

Authors

Charlotte Loh, Seungwook Han, Shivchander Sudalairaj, Rumen Dangovski, Kai Xu, Florian Wenzel, Marin Soljacic, Akash Srivastava

Abstract

Deep ensembles (DE) have been successful in improving model performance by learning diverse members via the stochasticity of random initialization. While recent works have attempted to promote further diversity in DE via hyperparameters or regularizing loss functions, these methods primarily still rely on a stochastic approach to explore the hypothesis space. In this work, we present Multi-Symmetry Ensembles (MSE), a framework for constructing diverse ensembles by capturing the multiplicity of hypotheses along symmetry axes, which explore the hypothesis space beyond stochastic perturbations of model weights and hyperparameters. We leverage recent advances in contrastive representation learning to create models that separately capture opposing hypotheses of invariant and equivariant functional classes and present a simple ensembling approach to efficiently combine appropriate hypotheses for a given task. We show that MSE effectively captures the multiplicity of conflicting hypotheses that is often required in large, diverse datasets like ImageNet. As a result of their inherent diversity, MSE improves classification performance, uncertainty quantification, and generalization across a series of transfer tasks.

Concepts

ensemble methods contrastive learning equivariant neural networks hypothesis space diversity symmetry preservation representation learning uncertainty quantification transfer learning self-supervised learning group theory data augmentation

The Big Picture

Imagine you’re teaching someone to recognize dogs. You’d probably want them to notice that a dog flipped upside-down is still a dog. But you’d also want them to know that orientation matters: a dog lying on its side looks different from one standing at attention, and that difference carries real information.

Now imagine you’re teaching them to recognize flowers. For flowers, orientation is almost irrelevant. A rose rotated 90 degrees is still unmistakably a rose. These two lessons are in fundamental tension, and that tension sits at the heart of a longstanding challenge in machine learning.

When researchers build ensembles of neural networks (collections of models that vote together on predictions), they typically rely on randomness to create diversity. Each model starts from a different random configuration and, through the chaos of training, lands somewhere different in the solution space. But this approach has a ceiling. No matter how many randomly initialized models you combine, they’ll cluster around the same basic assumption about how the world works. They’re like a team of people who all attended the same school: diverse in small ways, shaped by the same core beliefs.

Researchers at MIT and the MIT-IBM Watson AI Lab took a different approach. Their framework, Multi-Symmetry Ensembles (MSE), doesn’t shuffle the deck randomly. It deliberately trains models that hold opposing views about the geometric structure of the world, then puts them to a vote.

Key Insight: MSE builds diverse ensembles by pairing models that treat the same transformations, like rotations, in fundamentally opposite ways: one ignoring the transformation (invariant), one tracking it precisely (equivariant). This structured opposition creates diversity that random initialization simply cannot.

How It Works

The central concept is the distinction between invariance and equivariance, two ways a model can respond to geometric transformations like rotation, reflection, or flipping.

Figure 1

An invariant model, shown a rotated image, produces the same output regardless of rotation. It has learned to ignore the change entirely. An equivariant model tracks the rotation, its output changing in sync with how the input changed, preserving orientation information. Neither approach is universally better.

For flowers, invariance wins. For dogs with characteristic postures, equivariance wins. In a massive dataset like ImageNet, containing both flowers and dogs plus a thousand other categories, both assumptions are simultaneously correct for different slices of the data. Traditional ensembles can’t navigate this. Even a large ensemble of equivariant models remains fundamentally equivariant; you can’t average your way to invariance.

MSE builds the ensemble from both camps:

  1. Contrastive pretraining with opposing symmetries. The team uses recent advances in self-supervised learning, where models learn from unlabeled data by comparing examples against each other. They apply Equivariant Self-Supervised Learning (E-SSL) to pretrain separate models that explicitly encode invariant versus equivariant representations. One model is trained to be agnostic to rotations; another is trained to be sensitive to them.

  2. Task-agnostic diversity. This diversity is baked into the representations (the internal features a model learns) before any task-specific training. For transfer learning, where a model trained on one dataset is later applied to a new task, this is a big deal. Geometric assumptions shape everything downstream, so getting them right at the representation level pays dividends across many tasks.

  3. Greedy ensembling. Once diverse members are trained, the researchers use a greedy selection strategy to combine them. Rather than blindly averaging all models, they pick combinations that maximize validation performance, efficiently identifying which symmetry assumptions help for a given task.

Figure 2

The team tested MSE on ImageNet classification, uncertainty calibration (how well a model knows when it doesn’t know), and a battery of transfer tasks. The key result: combining a high-accuracy equivariant model with a weaker invariant model can outperform an ensemble of multiple high-accuracy equivariant models. The weaker model isn’t dead weight. It carries information the stronger model fundamentally cannot represent. Diversity, not raw accuracy, drives ensemble performance.

Figure 3

The framework extends beyond rotational symmetry too. MSE generalizes to other symmetry groups, including horizontal flipping and color jitter. Combining models with opposing hypotheses across multiple symmetry axes pushes performance further. The geometry of the hypothesis space is rich, and MSE systematically explores it.

Why It Matters

MSE reframes the diversity problem for ensemble learning. Instead of asking “how do we make models more randomly different,” it asks “what are the structurally distinct ways of understanding this data, and can we explicitly represent each?” The connection to symmetry group theory is direct, and it suggests far richer ensemble designs are possible.

There’s a deeper point here about physics and AI: geometric symmetry isn’t just a mathematical nicety. It’s a hypothesis about how the world works, and real-world datasets are full of objects that disagree about which symmetries apply. A model forced to commit to one geometric worldview will always be blind to the others.

Holding multiple conflicting geometric hypotheses at once is closer to how physicists actually think. Different physical regimes obey different symmetries. The art is knowing which applies when.

Plenty of open questions remain. Can MSE extend to more exotic symmetry groups relevant to physics, like gauge symmetries or Lorentz invariance? Could the greedy ensembling become fully adaptive at inference time, routing different inputs to the most appropriate ensemble member? And as models grow larger and pretraining more expensive, can MSE’s diversity gains be achieved at lower computational cost?

Bottom Line: Multi-Symmetry Ensembles show that the best path to diverse, reliable neural networks isn’t more randomness but structured geometric disagreement. By pairing models with opposing symmetry hypotheses, MSE achieves classification, calibration, and transfer performance that purely stochastic ensembles cannot match.

IAIFI Research Highlights

Interdisciplinary Research Achievement
This work imports concepts from physics (symmetry groups, invariance, and equivariance) directly into the engineering of machine learning systems. The mathematical language of fundamental physics turns out to be useful for designing AI systems too.
Impact on Artificial Intelligence
MSE offers a new way to think about ensemble diversity. Structured symmetry-based disagreement between models outperforms stochastic diversity alone, with measurable gains in classification accuracy, uncertainty quantification, and cross-dataset transfer.
Impact on Fundamental Interactions
Large visual datasets contain inherently conflicting symmetry hypotheses. Physicists know that the correct symmetry group depends on context, and here that observation is formalized and built into the model design itself.
Outlook and References
Future work may extend MSE to symmetry groups central to physics (gauge invariance, Lorentz symmetry) and to adaptive inference-time ensemble routing. The paper is available at [arXiv:2303.02484](https://arxiv.org/abs/2303.02484), and the code at [github.com/clott3/multi-sym-ensem](https://github.com/clott3/multi-sym-ensem).

Original Paper Details

Title
Multi-Symmetry Ensembles: Improving Diversity and Generalization via Opposing Symmetries
arXiv ID
2303.02484
Authors
Charlotte Loh, Seungwook Han, Shivchander Sudalairaj, Rumen Dangovski, Kai Xu, Florian Wenzel, Marin Soljacic, Akash Srivastava
Abstract
Deep ensembles (DE) have been successful in improving model performance by learning diverse members via the stochasticity of random initialization. While recent works have attempted to promote further diversity in DE via hyperparameters or regularizing loss functions, these methods primarily still rely on a stochastic approach to explore the hypothesis space. In this work, we present Multi-Symmetry Ensembles (MSE), a framework for constructing diverse ensembles by capturing the multiplicity of hypotheses along symmetry axes, which explore the hypothesis space beyond stochastic perturbations of model weights and hyperparameters. We leverage recent advances in contrastive representation learning to create models that separately capture opposing hypotheses of invariant and equivariant functional classes and present a simple ensembling approach to efficiently combine appropriate hypotheses for a given task. We show that MSE effectively captures the multiplicity of conflicting hypotheses that is often required in large, diverse datasets like ImageNet. As a result of their inherent diversity, MSE improves classification performance, uncertainty quantification, and generalization across a series of transfer tasks.