in ,

Accelerating AI computing to the speed of light

Accelerating AI computing to the speed of light
Credit: CC0 Public Domain

Artificial intelligence and machine learning are already an integral part of our everyday lives online. For example, search engines such as Google use intelligent ranking algorithms, and video streaming services such as Netflix use machine learning to personalize movie recommendations.

As the demands for AI online continue to grow, so does the need to speed up AI performance and find ways to reduce its energy consumption.

Now a University of Washington-led team has come up with a system that could help: an optical computing core prototype that uses phase-change material. This system is fast, energy efficient and capable of accelerating the neural networks used in AI and machine learning. The technology is also scalable and directly applicable to cloud computing.

The team published these findings Jan. 4 in Nature Communications.

“The hardware we developed is optimized to run algorithms of an artificial neural network, which is really a backbone algorithm for AI and machine learning,” said senior author Mo Li, a UW associate professor of both electrical and computer engineering and physics. “This research advance will make AI centers and cloud computing more energy efficient and run much faster.”

The team is among the first in the world to use phase-change material in optical computing to enable image recognition by an artificial neural network. Recognizing an image in a photo is something that is easy for humans to do, but it is computationally demanding for AI. Because image recognition is computation-heavy, it is considered a benchmark test of a neural network’s computing speed and precision. The team demonstrated that their optical computing core, running an artificial neural network, could easily pass this test.

“Optical computing first appeared as a concept in the 1980s, but then it faded in the shadow of microelectronics,” said lead author Changming Wu, a UW electrical and computer engineering graduate student. “Now, because of the end of Moore’s law, advances in integrated photonics and the demands of AI computing, it has been revamped. That’s very exciting.”

Provided by University of Washington

What do you think?

20 points
Upvote Downvote

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Diffractive networks improve optical image classification accuracy

Diffractive networks improve optical image classification accuracy

Wearable heart monitors and deep learning can provide early warning of SARS-CoV-2 infection

Wearable heart monitors and deep learning can provide early warning of SARS-CoV-2 infection