← Back to Timeline

Equivariant Contrastive Learning

Foundational AI

Authors

Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash Srivastava, Brian Cheung, Pulkit Agrawal, Marin Soljačić

Abstract

In state-of-the-art self-supervised learning (SSL) pre-training produces semantically good representations by encouraging them to be invariant under meaningful transformations prescribed from human knowledge. In fact, the property of invariance is a trivial instance of a broader class called equivariance, which can be intuitively understood as the property that representations transform according to the way the inputs transform. Here, we show that rather than using only invariance, pre-training that encourages non-trivial equivariance to some transformations, while maintaining invariance to other transformations, can be used to improve the semantic quality of representations. Specifically, we extend popular SSL methods to a more general framework which we name Equivariant Self-Supervised Learning (E-SSL). In E-SSL, a simple additional pre-training objective encourages equivariance by predicting the transformations applied to the input. We demonstrate E-SSL's effectiveness empirically on several popular computer vision benchmarks, e.g. improving SimCLR to 72.5% linear probe accuracy on ImageNet. Furthermore, we demonstrate usefulness of E-SSL for applications beyond computer vision; in particular, we show its utility on regression problems in photonics science. Our code, datasets and pre-trained models are available at https://github.com/rdangovs/essl to aid further research in E-SSL.

Concepts

self-supervised learning contrastive learning representation learning equivariant neural networks equivariant pretext task data augmentation symmetry preservation group theory transfer learning multi-task learning regression scientific workflows

The Big Picture

Think about how you recognize a dog. Whether it’s upright, flipped on its side, or rotated in a photo, you still know it’s a dog. That’s invariance: your brain ignores certain transformations. But imagine instead you’re trying to figure out which way a dog is oriented. Now you need to actually track that rotation, not ignore it. That’s a fundamentally different kind of knowledge, and a richer one.

Modern AI systems learn representations (compact internal summaries of what they observe) through self-supervised learning, where a model trains itself on unlabeled data by solving cleverly designed puzzles. The dominant approach enforces invariance: treat a photo and its flipped, color-shifted version as identical. This produces powerful features. But a team at MIT asked: what if some transformations are too informative to throw away?

Their answer is Equivariant Self-Supervised Learning (E-SSL), a framework where models learn to be invariant to some transformations while staying sensitive to others in a structured, predictable way. The representations it produces beat existing methods on standard benchmarks and work well in physics applications where symmetry matters.

Key Insight: Invariance is just a special case of a broader property called equivariance. By training models to track certain transformations rather than ignore them, E-SSL learns more informative representations, improving ImageNet accuracy and proving useful for regression tasks in photonics science.

How It Works

The core idea rests on equivariance: when you transform an input, the representation transforms in a predictable way. Invariance is the trivial case where the representation doesn’t change at all. Non-trivial equivariance means the representation moves in a structured, knowable way.

The researchers started with a diagnostic experiment. They took SimCLR, a popular invariant SSL method, and tested what happened when they made it either invariant or sensitive to a second transformation on top of standard random cropping.

Figure 1

The results split cleanly. For horizontal flips and grayscale, invariance helped. But for four-fold rotations, vertical flips, 2×2 jigsaws, four-fold Gaussian blurs, and color inversions, invariance actively hurt, while sensitivity to those same transformations improved performance. The pattern: if invariance to a transformation is harmful, equivariance to it tends to help.

E-SSL captures this with a two-part training objective:

  1. Standard invariant SSL loss (e.g., SimCLR’s contrastive loss, which rewards grouping different augmented views of the same image together): train the encoder, the network that compresses an image into a representation, to produce similar outputs for different views of the same image.
  2. Equivariance prediction loss: a lightweight prediction head identifies which transformation was applied, forcing the encoder to retain transformation information rather than discard it.

No architectural overhaul required. The prediction head is a simple classifier over discrete transformation classes, such as which of four rotation angles was applied. Minimal overhead, real impact on what the encoder learns.

Figure 2

The team applied E-SSL to four popular SSL methods (SimCLR, BYOL, Barlow Twins, and VICReg) and found consistent gains. On ImageNet, evaluated with a linear probe (a frozen representation tested with a single trainable layer on top), E-SSL pushed SimCLR to 72.5% accuracy. In a field where gains come slowly, that matters.

Why It Matters

Human knowledge about transformations shouldn’t just specify what to ignore. It should actively shape the structure of learned representations. The symmetries and near-symmetries of a problem are often its most informative features, and E-SSL puts that intuition to work as a training signal.

The photonics application makes this concrete. The team applied E-SSL to regression tasks in photonics, predicting physical properties of materials from simulation data. Domain knowledge tells you which transformations preserve or predictably alter the output. Encoding those as equivariance targets rather than invariances improved regression performance.

Many scientific problems involve data with known symmetries, whether rotational, translational, or gauge symmetries (the abstract symmetries underlying the fundamental forces of nature). Current SSL methods either ignore or discard that structure. E-SSL converts it into a training signal. The approach should transfer to any domain where the geometry of the data is partially understood but labels are expensive.

Bottom Line: The best self-supervised representations come not from maximally ignoring transformations, but from being selective: staying blind to some while staying sharp on others. A simple idea with real empirical payoff and a natural fit for physics.

IAIFI Research Highlights

Interdisciplinary Research Achievement
E-SSL applies equivariance, a concept central to fundamental physics, as a practical SSL training objective. The bridge from representation learning theory to physics is demonstrated on photonics regression tasks.
Impact on Artificial Intelligence
By generalizing invariant SSL to an equivariant framework, E-SSL improves linear probe accuracy on ImageNet and consistently boosts SimCLR, BYOL, Barlow Twins, and VICReg with minimal architectural changes.
Impact on Fundamental Interactions
Domain-specific transformation knowledge in photonics can be encoded as equivariance targets, improving prediction accuracy for physical property regression from simulation data.
Outlook and References
Future directions include extending E-SSL to continuous transformation groups and other scientific domains with known symmetry structure; the paper appeared at ICLR 2022 with code and pretrained models available at [github.com/rdangovs/essl](https://github.com/rdangovs/essl). See also: [arXiv:2111.00899](https://arxiv.org/abs/2111.00899).

Original Paper Details

Title
Equivariant Contrastive Learning
arXiv ID
2111.00899
Authors
Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash Srivastava, Brian Cheung, Pulkit Agrawal, Marin Soljačić
Abstract
In state-of-the-art self-supervised learning (SSL) pre-training produces semantically good representations by encouraging them to be invariant under meaningful transformations prescribed from human knowledge. In fact, the property of invariance is a trivial instance of a broader class called equivariance, which can be intuitively understood as the property that representations transform according to the way the inputs transform. Here, we show that rather than using only invariance, pre-training that encourages non-trivial equivariance to some transformations, while maintaining invariance to other transformations, can be used to improve the semantic quality of representations. Specifically, we extend popular SSL methods to a more general framework which we name Equivariant Self-Supervised Learning (E-SSL). In E-SSL, a simple additional pre-training objective encourages equivariance by predicting the transformations applied to the input. We demonstrate E-SSL's effectiveness empirically on several popular computer vision benchmarks, e.g. improving SimCLR to 72.5% linear probe accuracy on ImageNet. Furthermore, we demonstrate usefulness of E-SSL for applications beyond computer vision; in particular, we show its utility on regression problems in photonics science. Our code, datasets and pre-trained models are available at https://github.com/rdangovs/essl to aid further research in E-SSL.