Coherent Light Computes: How Optical AI Chips Could Change Everything

Let's imagine this scene: A beam of light carries not just information, but does the computing.
That sounds like science fiction - yet researchers now show it's possible.
Today we're exploring how optical computing is making a leap with the paper “Direct tensor processing with coherent light”.
We'll walk through what the breakthrough is, how it works, and why it matters for AI hardware.
What the challenge is
Modern AI depends on tensor operations - large-scale matrix multiplications, convolutions, attention-layers. These run on GPUs today and demand lots of power, memory bandwidth, and time.
Optical computing has long promised high bandwidth, massive parallelism, and low energy use. But so far it has had limits: many optical approaches handle vector-matrix multiplication, scalar operations, or require multiple passes of light.
What was missing was an optical system that can handle tensor operations - the kind of large matrix-matrix multiplications and multi-dimensional computations used in modern neural networks - in one shot.
What the researchers propose
The team introduces a method called Parallel Optical Matrix-Matrix Multiplication (POMMM).
In this method, a single propagation of coherent light (one pass through an optical system) encodes and computes matrix-matrix multiplications in parallel.
They encode data into the amplitude and phase of the light wave, and exploit the spatial position, phase gradients and frequency (wavelength) domain to map inputs to outputs.
The system supports real-valued and complex-valued domains, and the authors show consistency with GPU-based matrix multiplications.
How it works (overview)
The two matrices (or tensor levels) are encoded into light fields. One via amplitude/phase modulations for matrix A, the other via encoded positions or wavelengths for matrix B.
The coherent light then propagates through lenses, Fourier transforms, phase-elements and detectors. As it propagates, the light naturally performs the equivalent of dot-product operations and summations because of optical Fourier transform properties.
Because the operations happen in parallel in the optical domain, many dot-products run simultaneously. Then the resulting field is captured (by a camera or detector) and mapped back to the digital domain.
They also show multi-wavelength extensions: by using different wavelengths they encode additional dimensions (higher-order tensors) and expand the system's scale. (Bohrium)
Why it matters
This is a substantial step for AI hardware. Here's why:
It offers a way to scale tensor operations much faster than current electronic hardware by harnessing light's natural parallelism. It reduces the number of sequential operations hardware needs to execute. Instead of many clock cycles, you get a “single-shot” propagation.
- Energy efficiency: optical propagation can reduce switching losses, memory bottlenecks and overheads typical in GPUs.
- Compatibility: The method shows it can implement operations used in convolutional neural networks (CNNs) and vision transformers (ViTs) - meaning the optical approach is not just a toy, but relevant to current AI architectures.
The limitations & what to watch
Of course, this is early. The hardware prototype is still lab-scale. Integration into full systems will take time. Some important caveats:
- Accuracy and error-rates: the optical system must match digital precision, manage noise, calibration, and real-world conditions. The researchers show promising match-ups but scaling adds challenge.
- Integration with existing silicon hardware: Photonic chips, detectors, modulators, memory all must tie in with digital systems.
- The ecosystem: For many enterprises, moving to optical computing means revamping parts of their compute stack - not trivial.
Final thoughts
This work shows that coherent light can carry out complex tensor operations in one propagation. That's a meaningful leap towards optical AI hardware.
For you - if you're working on AI architecture, hardware strategy or compute infrastructure - this signals a coming shift. It suggests that in the next few years, the compute bottleneck for large-scale AI might look very different.
