industry news
Subscribe Now

Lightelligence Revolutionizes Big Data Interconnect with World’s First Optical Network-on-Chip Processor

  • Domain-Specific AI Processor Unlocks New Interconnect Paradigm for Data Centers and Other High-Performance Applications, Breaking the “Memory Wall”
  • Utilizes Advanced Vertically Stacked Packaging, Integrating Optical Network-on-Chip and Electronic Integrated Circuits into One Single Package
  • First Public Demonstration at Hot Chips 2023 

BOSTON –– June 28, 2023 –– Lightelligence, the global leader in photonic computing, today introduced a new big data interconnect paradigm with the launch of Hummingbird™, the world’s first Optical Network-on-Chip (oNOC) processor designed for domain-specific artificial intelligence (AI) workloads.

Hummingbird utilizes advanced vertically stacked packaging technologies to integrate a photonic chip and an electronic chip into one single package serving as the communications network for data centers and other high-performance applications. The first public demonstration of Hummingbird will be at Hot Chips August 27-29 at Stanford University.

“Photonics is the solution to the critical compute scaling problem, which has become pressing as the traditional solutions struggle to keep up with the exponential growth of compute power demand spurred by breakthroughs in the AI industry,” remarks Yichen Shen, CEO of Lightelligence. “Hummingbird demonstrates how the industry can address the scaling problem by incorporating photonic technologies into their next-generation product.” 

“Lightelligence is breaking the memory wall with its proprietary photonics technology that could revolutionize the semiconductor industry,” adds Dylan Patel, Chief Analyst at SemiAnalysis.

Hummingbird is Lightelligence’s second product in its photonic computing portfolio. Its Photonic Arithmetic Computing Engine (PACE) platform released in late-2021 fully integrates photonics and electronics in a small form factor leveraging custom 3D packaging and seamless co-design. 

Introducing Hummingbird

Hummingbird is the first in a family of products that utilize Lightelligence’s oNOC platform, which significantly improves computing performances by enabling innovative interconnect topologies via silicon photonics. Its waveguides propagate signals at the speed of light and utilize an all-to-all data broadcast network to each core on a 64-core domain-specific AI processor chip, giving Hummingbird significant advantages in latency and power reduction over traditional digital interconnect solutions. 

Compute scaling challenges inspired the creation of an optical interconnect solution. Unlike digital networks, Hummingbird’s oNOC technology increases density scaling by enabling interconnect topologies that would be otherwise unrealizable.

In oNOC, power and latency are virtually unaffected by distance, making the technology ideal for developing new and more robust topologies that do not rely on nearest neighbor communication. oNOC topologies like Hummingbird’s enable higher utilization of compute power even in a single electronic IC configuration due to more efficient communication. With oNOC, mapping workloads to hardware becomes easier and provides greater freedom to select the right topology for the computing task. 

In Hummingbird, Lightelligence implemented a low-latency optical all-to-all broadcast network spanning 64 cores. With 64 transmitters and 512 receivers, Hummingbird provides a framework to implement a variety of dense optical network topologies.

Hummingbird’s electronic and photonic ICs are co-packaged and integrated into a PCIe form factor ready for installation in industry-standard servers. Coupled with the Lightelligence Software Development Kit (SDK), machine learning and AI workloads can be optimized to take full advantage of the oNOC. oNOC and Hummingbird IP can also be customized for other unique workloads and applications. 

Future generations of Hummingbird will employ reticle-stitching to support chiplet architectures to enable better scalability, improve energy efficiency, and further reduce bottlenecks.

 

Availability and Pricing

Lightelligence is actively signing development partners to sample Hummingbird-based PCIe add-in cards along with Lightelligence’s SDK in Q3 2023. 

Contact Lightelligence at info@lightelligence.ai for inquiries on pricing and availability. Performance numbers are available to qualified customers upon request.

Lightelligence at Flash Memory Summit and Hot Chips

Lightelligence will exhibit at Flash Memory Summit in Booth #755 August 8-10 at the Santa Clara Convention Center in Santa Clara, Calif.Lightelligence will demonstrate Hummingbird at Hot Chips 2023 at Stanford University in Palo Alto, Calif., August 27-29. Maurice Steinman, Lightelligence’s Vice President of Engineering, will present “Hummingbird Low-Latency Computing Engine” August 29 at 11:30 a.m. P.D.T.

About Lightelligence

Lightelligence is transforming cutting-edge photonic technology into groundbreaking solutions that offer exponential improvements in computing power and dramatically reduce energy consumption. As the global leader in the photonic computing industry, Lightelligence is to date the only company that has publicly demonstrated integrated silicon photonic computing systems working at speed. Founded in 2017, Lightelligence has approximately 200 employees worldwide and has raised more than $220 million in funding.

Leave a Reply

featured blogs
Nov 22, 2024
We're providing every session and keynote from Works With 2024 on-demand. It's the only place wireless IoT developers can access hands-on training for free....
Nov 22, 2024
I just saw a video on YouTube'”it's a few very funny minutes from a show by an engineer who transitioned into being a comedian...

featured video

Introducing FPGAi – Innovations Unlocked by AI-enabled FPGAs

Sponsored by Intel

Altera Innovators Day presentation by Ilya Ganusov showing the advantages of FPGAs for implementing AI-based Systems. See additional videos on AI and other Altera Innovators Day in Altera’s YouTube channel playlists.

Learn more about FPGAs for Artificial Intelligence here

featured paper

Quantized Neural Networks for FPGA Inference

Sponsored by Intel

Implementing a low precision network in FPGA hardware for efficient inferencing provides numerous advantages when it comes to meeting demanding specifications. The increased flexibility allows optimization of throughput, overall power consumption, resource usage, device size, TOPs/watt, and deterministic latency. These are important benefits where scaling and efficiency are inherent requirements of the application.

Click to read more

featured chalk talk

Dependable Power Distribution: Supporting Fail Operational and Highly Available Systems
Sponsored by Infineon
Megatrends in automotive designs have heavily influenced the requirements needed for vehicle architectures and power distribution systems. In this episode of Chalk Talk, Amelia Dalton and Robert Pizuti from Infineon investigate the trends and new use cases required for dependable power systems and how Infineon is advancing innovation in automotive designs with their EiceDRIVER and PROFET devices.
Dec 7, 2023
59,710 views