A technical paper titled “Hypermultiplexed Integrated Tensor Optical Processor” was published by researchers at the University of Southern California, Massachusetts Institute of Technology (MIT), City University of Hong Kong, and NTT Research.
“The escalating data volume and complexity resulting from the rapid expansion of artificial intelligence (AI), internet of things (IoT) and 5G/6G mobile networks is creating an urgent need for energy-efficient, scalable computing hardware. Here we demonstrate a hypermultiplexed integrated tensor optical processor (HITOP) that performs trillions of operations per second (TeraOPS) at the energy cost of 25 femtojoule per operation (25 fJ/OP). Based on space-time-wavelength three-dimensional (3D) data streaming, HITOP is built with arrays of wafer-fabricated III/V-based micron-scale lasers (spanning ~1 THz) incorporating thin-film Lithium-Niobate electro-optic (EO) photonics. Multidimensional parallelism allows matrix-matrix multiplications (N3 operations) using O(N) devices, facilitating scalable on-chip integration. With each device activating 10 billion parameters per second, the HITOP scalability is validated in machine learning models with 405,000 parameters, which is 25,000 times more than previous integrated optical systems. A combination of high clockrates (10 GS/s), parallel processing and real-time reprogrammability unlocks the full potential of light for next-generation AI accelerators in applications ranging from training with trillions of parameters, real-time decision making in autonomous vehicles and robotics, dynamic optimization in smart manufacturing, to complex simulation for climate modeling and drug discovery.”
Find the technical paper here. Published February 2024 (preprint).
Ou, Shaoyuan, Alexander Sludds, Ryan Hamerly, Ke Zhang, Hanke Feng, Eric Zhong, Cheng Wang, Dirk Englund, Mengjie Yu, and Zaijun Chen. “Hypermultiplexed Integrated Tensor Optical Processor.” arXiv preprint arXiv:2401.18050 (2024).
Related Reading
The Race Toward Quantum Advantage
Enormous amounts of money have been invested into quantum computing, but so far it has not surpassed conventional computers. When will that change?
Data Formats For Inference On The Edge
Choosing the right format may sound like a simple task. It isn’t, but that choice is essential for efficient edge inferencing.
Dealing With Performance Bottlenecks In SoCs
A surge in the amount of data that SoCs need to process is bogging down performance, and memory and communication bandwidth are straining.
Leave a Reply