Photonic processor could enable ultrafast AI computations with extreme energy efficiency

The deep neural network models that power today’s most demanding machine-learning applications have grown so large and complex that they are pushing the limits of traditional electronic computing hardware.

Photonic hardware, which can perform machine-learning computations with light, offers a faster and more energy-efficient alternative. However, there are some types of neural network computations that a photonic device can’t perform, requiring the use of off-chip electronics or other techniques that hamper speed and efficiency.

Building on a decade of research, scientists from MIT and elsewhere have developed a new photonic chip that overcomes these roadblocks. They demonstrated a fully integrated photonic processor that can perform all the key computations of a deep neural network optically on the chip.

The optical device was able to complete the key computations for a machine-learning classification task in less than half a nanosecond while achieving more than 92 percent accuracy — performance that is on par with traditional hardware.

The chip, composed of interconnected modules that form an optical neural network, is fabricated using commercial foundry processes, which could enable the scaling of the technology and its integration into electronics.

In the long run, the photonic processor could lead to faster and more energy-efficient deep learning for computationally demanding applications like lidar, scientific research in astronomy and particle physics, or high-speed telecommunications.

“There are a lot of cases where how well the model performs isn’t the only thing that matters, but also how fast you can get an answer. Now that we have an end-to-end system that can run a neural network in optics, at a nanosecond time scale, we can start thinking at a higher level about applications and algorithms,” says Saumil Bandyopadhyay ’17, MEng ’18, PhD ’23, a visiting scientist in the Quantum Photonics and AI Group within the Research Laboratory of Electronics (RLE) and a postdoc at NTT Research, Inc., who is the lead author of a paper on the new chip.

Bandyopadhyay is joined on the paper by Alexander Sludds ’18, MEng ’19, PhD ’23; Nicholas Harris PhD ’17; Darius Bunandar PhD ’19; Stefan Krastanov, a former RLE research scientist who is now an assistant professor at the University of Massachusetts at Amherst; Ryan Hamerly, a visiting scientist at RLE and senior scientist at NTT Research; Matthew Streshinsky, a former silicon photonics lead at Nokia who is now co-founder and CEO of Enosemi; Michael Hochberg, president of Periplous, LLC; and Dirk Englund, a professor in the Department of Electrical Engineering and Computer Science, principal investigator of the Quantum Photonics and Artificial Intelligence Group and of RLE, and senior author of the paper. The research appears today in Nature Photonics.

Machine learning with light

Deep neural networks are composed of many interconnected layers of nodes, or neurons, that operate on input data to produce an output. One key operation in a deep neural network involves the use of linear algebra to perform matrix multiplication, which transforms data as it is passed from layer to layer.

But in addition to these linear operations, deep neural networks perform nonlinear operations that help the model learn more intricate patterns. Nonlinear operations, like activation functions, give deep neural networks the power to solve complex problems.

In 2017, Englund’s group, along with researchers in the lab of Marin Soljačić, the Cecil and Ida Green Professor of Physics, demonstrated an optical neural network on a single photonic chip that could perform matrix multiplication with light.

But at the time, the device couldn’t perform nonlinear operations on the chip. Optical data had to be converted into electrical signals and sent to a digital processor to perform nonlinear operations.

“Nonlinearity in optics is quite challenging because photons don’t interact with each other very easily. That makes it very power consuming to trigger optical nonlinearities, so it becomes challenging to build a system that can do it in a scalable way,” Bandyopadhyay explains.

They overcame that challenge by designing devices called nonlinear optical function units (NOFUs), which combine electronics and optics to implement nonlinear operations on the chip.

The researchers built an optical deep neural network on a photonic chip using three layers of devices that perform linear and nonlinear operations.

A fully-integrated network

At the outset, their system encodes the parameters of a deep neural network into light. Then, an array of programmable beamsplitters, which was demonstrated in the 2017 paper, performs matrix multiplication on those inputs.

The data then pass to programmable NOFUs, which implement nonlinear functions by siphoning off a small amount of light to photodiodes that convert optical signals to electric current. This process, which eliminates the need for an external amplifier, consumes very little energy.

“We stay in the optical domain the whole time, until the end when we want to read out the answer. This enables us to achieve ultra-low latency,” Bandyopadhyay says.

Achieving such low latency enabled them to efficiently train a deep neural network on the chip, a process known as in situ training that typically consumes a huge amount of energy in digital hardware.

“This is especially useful for systems where you are doing in-domain processing of optical signals, like navigation or telecommunications, but also in systems that you want to learn in real time,” he says.

The photonic system achieved more than 96 percent accuracy during training tests and more than 92 percent accuracy during inference, which is comparable to traditional hardware. In addition, the chip performs key computations in less than half a nanosecond.     

“This work demonstrates that computing — at its essence, the mapping of inputs to outputs — can be compiled onto new architectures of linear and nonlinear physics that enable a fundamentally different scaling law of computation versus effort needed,” says Englund.

The entire circuit was fabricated using the same infrastructure and foundry processes that produce CMOS computer chips. This could enable the chip to be manufactured at scale, using tried-and-true techniques that introduce very little error into the fabrication process.

Scaling up their device and integrating it with real-world electronics like cameras or telecommunications systems will be a major focus of future work, Bandyopadhyay says. In addition, the researchers want to explore algorithms that can leverage the advantages of optics to train systems faster and with better energy efficiency.

This research was funded, in part, by the U.S. National Science Foundation, the U.S. Air Force Office of Scientific Research, and NTT Research.

Salesforce: UK set to lead agentic AI revolution

Salesforce has unveiled the findings of its UK AI Readiness Index, signalling the nation is in a position to spearhead the next wave of AI innovation, also known as agentic AI. The report places the UK ahead of its G7 counterparts in terms of AI adoption…

A data designer driven to collaborate with communities

It is fairly common in public discourse for someone to announce, “I brought data to this discussion,” thus casting their own conclusions as empirical and rational. It is less common to ask: Where did the data come from? How was it collected? Why is there data about some things but not others?

MIT Associate Professor Catherine D’Ignazio SM ’14 does ask those kinds of questions. A scholar with a far-reaching portfolio of work, she has a strong interest in applying data to social issues — often to help the disempowered gain access to numbers, and to help provide a fuller picture of civic problems we are trying to address.

“If we want an educated citizenry to participate in our democracy with data and data-driven arguments, we should think about how we design our data infrastructures to support that,” says D’Ignazio.

Take, for example, the problem of feminicide, the killing of women as a result of gender-based violence. Activists throughout Latin America started tabulating cases about it and building databases that were often more thorough than official state records. D’Ignazio has observed the issue and, with colleagues, co-designed AI tools with human rights defenders to support their monitoring work.

In turn, D’Ignazio’s 2024 book on the subject, “Counting Feminicide,” chronicled the entire process and has helped bring the issue to a new audience. Where there was once a data void, now there are substantial databases helping people recognize the reality of the problem on multiple continents, thanks to innovative citizens. The book outlines how grassroots data science and citizen data activism are generally rising forms of civic participation.

“When we talk about innovation, I think: Innovation for whom? And by whom? For me those are key questions,” says D’Ignazio, a faculty member in MIT’s Department of Urban Studies and Planning and director of MIT’s Data and Feminism Lab. For her research and teaching, D’Ignazio was awarded tenure earlier this year.

Out of the grassroots

D’Ignazio has long cultivated an interest in data science, digital design, and global matters. She received her BA in international relations from Tufts University, then became a software developer in the private sector. Returning to her studies, she earned an MFA from the Maine College of Art, and then an MS from the MIT Media Lab, which helped her synthesize her intellectual outlook.

“The Media Lab for me was the place where I was able to converge all those interests I had been thinking about,” D’Ignazio says. “How can we have more creative applications of software and databases? How can we have more socially just applications of AI? And how do we organize our technology and resources for a more participatory and equitable future for all of us?”

To be sure, D’Ignazio did not spend all her time at the Media Lab examining database issues. In 2014 and 2018 she co-organized a feminist hackathon called “Make the Breast Pump Not Suck,” in which hundreds of participants developed innovative technologies and policies to address postpartum health and infant feeding. Still, much of her work has focused on data architecture, data visualization, and the analysis of the relationship between data production and society.

D’Ignazio started her teaching career as a lecturer in the Digital + Media graduate program at Rhode Island School of Design, then became an assistant professor of data visualization and civic media in Emerson College’s journalism department. She joined the MIT faculty as an assistant professor in 2020.

D’Ignazio’s first book, “Data Feminism,” co-authored with Lauren Klein of Emory University and published in 2020, took a wide-ranging look at many ways that everyday data reflects the civic society that it emerges from. The reported rates of sexual assault on college campuses, for instance, could be deceptive because the institutions with the lowest rates might be those with the most problematic reporting climates for survivors.

D’Ignazio’s global outlook — she has lived in France, Argentina, and Uruguay, among other places — has helped her understand the regional and national politics behind these issues, as well as the challenges citizen watchdogs can face in terms of data collection. No one should think such projects are easy.

“So much grassroots labor goes into the production of data,” D’Ignazio says. “One thing that’s really interesting is the huge amount of work it takes on the part of grassroots or citizen science groups to actually make data useful. And oftentimes that’s because of institutional data structures that are really lacking.”

Letting students thrive

Overall, the issue of who participates in data science is, as D’Ignazio and Klein have written, “the elephant in the server room.” As an associate professor, D’Ignazio works to encourage all students to think openly about data science and its social underpinnings. In turn, she also draws inspiration from productive students.

“Part of the joy and privilege of being a professor is you have students who take you in directions you would not have gone in yourself,” D’Ignazio says.

One of D’Ignazio’s graduate students at the moment, Wonyoung So, has been digging into housing data issues. It is fairly simple for property owners to access information about tenants, but less so the other way around; this makes it hard to find out if landlords have abnormally high eviction rates, for example.

“There are all of these technologies that allow landlords to get almost every piece of information about tenants, but there are so few technologies allowing tenants to know anything about landlords,” D’Ignazio explains. The availability of data “often ends up reproducing asymmetries that already exist in the world.” Moreover, even where housing data is published by jurisdictions, she notes, “it’s incredibly fragmented, and published poorly and differently, from place to place. There are massive inequities even in open data.”

In this way housing seems like yet another area where new ideas and better data structures can be developed. It is not a topic she would have focused on by herself, but D’Ignazio also views herself as a facilitator of innovative work by others. There is much progress to be made in the application of data science to society, often by developing new tools for people to use.

“I’m interested in thinking about how information and technology can challenge structural inequalities,” D’Ignazio says. “The question is: How do we design technologies that help communities build power?”

Navigating the 2025 Challenges of Adopting Enterprise AI

The business world has witnessed a phenomenal surge in the adoption of artificial intelligence (AI) — and specifically generative AI (Gen AI). According to Deloitte estimates, enterprise spending on Gen AI in 2024 is poised to increase by 30 percent from the 2023 figure of USD…

Patrick Leung, CTO of Faro Health – Interview Series

Patrick Leung, CTO of Faro Health, drives the company’s AI-enabled platform, which simplifies and speeds up clinical trial protocol design. Faro Health’s tools enhance efficiency, standardization, and accuracy in trial planning, integrating data-driven insights and streamlined processes to reduce trial risks, costs, and patient burden. Faro…

Prescriptive AI: The Smart Decision-Maker for Healthcare, Logistics, and Beyond

Artificial Intelligence (AI) has made significant progress in recent years, transforming how organizations manage complex data and make decisions. With the vast amount of data available, many industries face the critical challenge of acting on real-time insights. This is where prescriptive AI steps in. Unlike traditional…

10 Best AI Phone Platforms & Agents for Call Centers (November 2024)

AI voice agents are an integral part of today’s automated phone communication, enabling businesses to process thousands of concurrent calls through sophisticated speech recognition and natural language processing systems. These platforms combine voice synthesis, real-time transcription, and contextual understanding to handle tasks ranging from appointment scheduling…

When Graph AI Meets Generative AI: A New Era in Scientific Discovery

In recent years, artificial intelligence (AI) has emerged as a key tool in scientific discovery, opening up new avenues for research and accelerating the pace of innovation. Among the various AI technologies, Graph AI and Generative AI are particularly useful for their potential to transform how…