← Back to Timeline

AI-Driven Robotics for Optics

Experimental Physics

Authors

Shiekh Zia Uddin, Sachin Vaidya, Shrish Choudhary, Zhuo Chen, Raafat K. Salib, Luke Huang, Dirk R. Englund, Marin Soljačić

Abstract

Optics is foundational to research in many areas of science and engineering, including nanophotonics, quantum information, materials science, biomedical imaging, and metrology. However, the design, assembly, and alignment of optical experiments remain predominantly manual, limiting throughput and reproducibility. Automating such experiments is challenging due to the strict, non-negotiable precision requirements and the diversity of optical configurations found in typical laboratories. Here, we introduce a platform that integrates generative artificial intelligence, computer vision, and robotics to automate free-space optical experiments. The platform translates user-defined goals into valid optical configurations, assembles them using a robotic arm, and performs micrometer-scale fine alignment using a robot-deployable tool. It then executes a range of automated measurements, including beam characterization, polarization mapping, and spectroscopy, with consistency surpassing that of human operators. This work demonstrates the first flexible, AI-driven automation platform for optics, offering a path towards remote operation, cloud labs, and high-throughput discovery in the optical sciences.

Concepts

scientific workflows robotic optical assembly llm optical agent fine-tuning micrometer alignment tool experimental design computer vision automated discovery data augmentation tensor networks inverse problems active learning

The Big Picture

A researcher hunches over a laser table at midnight, tweaking a tiny mirror by fractions of a millimeter, watching interference fringes shimmer on the wall. One wrong nudge, a vibration from a passing truck or a slip of the wrist, and hours of careful alignment come undone. Components must sit within microns of their intended positions. This is how optical experiments still get built: by hand, one human at a time.

Optics shows up across modern science. Gravitational wave detectors, quantum computers, cancer-screening microscopes, astronomical spectrographs: light is the go-to probe. But the labs themselves haven’t changed much in fifty years. Researchers still slide lenses and mirrors into position on vibration-isolated tables, iterating by hand until the alignment is right.

A team from MIT’s IAIFI and Research Laboratory of Electronics has built the first robotic platform that can design, assemble, align, and run free-space optical experiments from end to end, automatically. You describe what you want in plain English; the system handles the rest.

Key Insight: By combining fine-tuned large language models with computer vision and a robotic arm, the researchers automated every stage of a free-space optical experiment, from goal specification to physical assembly to measurement, achieving consistency that surpasses human operators.

How It Works

The platform has three stages, each covering one phase of what a human experimentalist would normally do.

Figure 1

Stage 1: Design via fine-tuned LLMs. A researcher types what they want: “Give me a Mach-Zehnder interferometer” or “a Michelson interferometer with the beam entering at 30 degrees and the beamsplitter near coordinate (20 cm, 3 cm).” An Optics Agent takes that request and generates a list of optical components with their x-y positions and orientations.

The Optics Agent runs on LLaMA3.1-8B-Instruct, fine-tuned using QuanTA (Quantum-informed Tensor Adaptation). QuanTA borrows mathematical structures from quantum physics to update the model efficiently without retraining all its billions of parameters. Training data came from a synthetically augmented dataset covering four standard configurations: Michelson interferometer, Mach-Zehnder interferometer, Hong-Ou-Mandel interferometer, and a 4f imaging system.

Each configuration was generated through geometric transformations and automatically captioned to mimic how real users describe experiments. A validation step checks whether the output is physically feasible before passing it downstream.

Stage 2: Robotic assembly. A Coding Agent translates the validated layout into step-by-step robot instructions. The robotic arm, equipped with a computer vision system:

  • Identifies individual optical components on the table
  • Picks each one and places it at its target position with sub-millimeter precision
  • Applies a positional error correction loop using camera feedback
  • Deploys a purpose-built fine-alignment tool for the micrometer-scale adjustments that human operators typically spend the most time on

Stage 3: Automated measurements. Once assembled and aligned, the platform runs a full suite of measurements without further human input. The team demonstrated beam characterization, polarization mapping, transmission spectroscopy, and real-space optimization. Across repeated trials, the automated platform showed less variation than trained human operators performing the same tasks.

Figure 2

Free-space experiments are unforgiving. Laser light travels through open air rather than fiber-optic cables, and the tolerances are brutal: an angular misalignment of a fraction of a degree can walk a beam entirely off a detector. The fine-alignment tool solves this by letting the robot handle coarse pick-and-place, then handing off to a specialized instrument that works at the precision the physics demands.

Why It Matters

This platform could sharply accelerate optical research. A lab that currently needs days to set up a new interferometric experiment could cut that to hours or minutes, reproducibly, with results that don’t depend on which postdoc happened to be available. For quantum information science, where alignment is everything and reproducibility determines whether results can be trusted, that’s a big deal.

The longer arc points toward cloud labs: remote-access facilities where researchers anywhere in the world specify an optical experiment, have it assembled and run by a robot, and receive data back without setting foot in a physical lab. Chemistry and genomics already have versions of this. The work here makes it conceivable for optics too.

Then there’s high-throughput discovery. Systematically scanning large spaces of optical configurations to find new phenomena or optimal designs is the kind of thing no human team could do by hand.

The fine-tuned LLM here isn’t just retrieving information. It generates layouts that satisfy spatial constraints, physical feasibility requirements, and equipment limitations. Whether these capabilities generalize beyond the four demonstrated setup types, and how the system handles novel or ambiguous requests, remain open questions.

Bottom Line: The first flexible, AI-driven robotic platform for free-space optics handles design, assembly, alignment, and measurement with consistency that outperforms human operators. Cloud labs and high-throughput optical discovery are no longer hypothetical.

IAIFI Research Highlights

Interdisciplinary Research Achievement
This project combines LLM fine-tuning, computer vision, and generative design with the precision demands of experimental optics, sitting squarely at IAIFI's intersection of AI and physics.
Impact on Artificial Intelligence
The paper lays out a practical pipeline for deploying fine-tuned LLMs as physical-world design agents, where outputs must satisfy hard geometric and physical constraints rather than just sound plausible.
Impact on Fundamental Interactions
Automating optical experiments removes a major bottleneck in quantum information, nanophotonics, and precision measurement, fields where optical setups are essential for probing fundamental physics.
Outlook and References
Future work could extend the platform to more complex configurations and toward fully remote cloud-lab operation; the paper is available at [arXiv:2505.17985](https://arxiv.org/abs/2505.17985) and was authored by researchers from MIT Physics, EECS, and the NSF IAIFI.

Original Paper Details

Title
AI-Driven Robotics for Optics
arXiv ID
2505.17985
Authors
Shiekh Zia Uddin, Sachin Vaidya, Shrish Choudhary, Zhuo Chen, Raafat K. Salib, Luke Huang, Dirk R. Englund, Marin Soljačić
Abstract
Optics is foundational to research in many areas of science and engineering, including nanophotonics, quantum information, materials science, biomedical imaging, and metrology. However, the design, assembly, and alignment of optical experiments remain predominantly manual, limiting throughput and reproducibility. Automating such experiments is challenging due to the strict, non-negotiable precision requirements and the diversity of optical configurations found in typical laboratories. Here, we introduce a platform that integrates generative artificial intelligence, computer vision, and robotics to automate free-space optical experiments. The platform translates user-defined goals into valid optical configurations, assembles them using a robotic arm, and performs micrometer-scale fine alignment using a robot-deployable tool. It then executes a range of automated measurements, including beam characterization, polarization mapping, and spectroscopy, with consistency surpassing that of human operators. This work demonstrates the first flexible, AI-driven automation platform for optics, offering a path towards remote operation, cloud labs, and high-throughput discovery in the optical sciences.