• Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io
No Result
View All Result
Converge Digest
Wednesday, April 15, 2026
  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io
No Result
View All Result
Converge Digest
No Result
View All Result

Home » Hot Interconnects: Celestial AI’s Photonic Fabric

Hot Interconnects: Celestial AI’s Photonic Fabric

August 25, 2025
in AI Infrastructure, All, Optical
A A

AI workloads are pushing the limits of copper and conventional optics, and Celestial AI believes its Photonic Fabric is the key to breaking through. At this week’s IEEE Hot Interconnects online conference, senior director Ravi Mahatme detailed how the company’s technology embeds optical I/O directly into the middle of silicon dies, bypassing shoreline constraints to deliver nanosecond-class latency, 2.4 pJ/bit efficiency, and terabits of bandwidth per link. The approach, he argued, enables scale-up architectures and unified memory access that are critical for accelerated computing.

Celestial AI’s Photonic Fabric is built as a platform, spanning chiplets, IP blocks, NICs, and system-level modules. The Photonic Fabric Link (PFLink) provides compact, thermally stable optical modulators integrated into advanced 5nm/4nm CMOS packaging. Unlike traditional optical approaches, the link operates without DSP, relying on linear drive optics to maintain high SNR and low BER. The Photonic Fabric chiplet supports 4.4 Tbps of bandwidth—greater than a single HBM3E stack—and is protocol adaptive, supporting AXI, HBM/DDR, UAL, and CXL.

At the system level, the Photonic Fabric Module (PFM) integrates two 36 GB HBM3E stacks, eight DDR5 channels (2 TB capacity), and an 8 Tbps scale-up switch. Sixteen such modules can be assembled into an appliance that offers 115 Tbps of backend switching and 32 TB of unified DDR5 memory, with HBM acting as a write-through cache. Latency from server to memory is ~300 ns, compared with tens of microseconds for conventional networks. Celestial says this architecture reduces sharding, improves throughput, and enables collective communications across XPUs with deterministic tail latencies at p99 and p99.9.

Applications include XPU-to-XPU scale-up, KV cache offload, vector database and embeddings serving, recommendation systems, and long context-length inference for large AI models. For developers, Celestial offers PF-NIC add-in cards supporting CXL 3.0/2.0 over PCIe Gen6, delivering 2 Tbps per card and enabling low-latency server access to PF appliances. Software complements the hardware with link management, telemetry, and predictive insights for fleets of interconnected modules.

• Photonic Fabric efficiency: ~2.4 pJ/bit, nanosecond-class latency

• PFLink: compact, thermally stable optical modulator integrated in-die

• No DSP: linear drive optics maintain SNR/BER without power-hungry DSP blocks

• Photonic Fabric Chiplet: 4.4 Tbps bandwidth, UCIe/proprietary PHY support

• Photonic Fabric Module: 72 GB HBM3E, 2 TB DDR5, 8 Tbps switching, 7.2 Tbps bandwidth per module

• PF Appliance: 16 modules = 115 Tbps switching, 32 TB DDR5 shared memory

• PF-NIC: CXL 3.0/2.0 support, 2 Tbps over PCIe Gen6 per card

• Use cases: XPU scale-up, MoE inference, KV cache offload, embeddings serving, long context AI, checkpointing/weight streaming

• Benefits: reduced sharding, deterministic latency, throughput gains, new application enablement

“Photonic Fabric delivers transformative benefits by reducing tail latency and enabling new AI applications, especially those with long context lengths that are difficult or impossible today,” Mahatme said.

🌐 Analysis: Founded in 2020, Celestial AI is headquartered in Santa Clara, California, specializing in optical interconnect technology to enhance AI infrastructure. The company’s leadership includes co-founder and CEO David Lazovsky, previously the founder and CEO of Intermolecular Inc. and a venture partner at Khosla Ventures; co-founder and COO Preet Virk, who has extensive semiconductor industry experience in operational and business development roles; and CTO Philip Winterbottom, whose background includes leadership roles in developing advanced computing systems. In October 2024, Celestial AI strengthened its technology portfolio by acquiring a significant silicon photonics patent collection from Rockley Photonics.

In March 2025, Celestial AI announced it secured $250 million in Series C1 funding led by Fidelity Management & Research Company, increasing its total funding to over $515 million. The funding round included new investors such as BlackRock, Maverick Silicon, Tiger Global Management, and Lip-Bu Tan, alongside existing investors AMD Ventures, Koch Disruptive Technologies, Temasek, Xora Innovation, Porsche Automobil Holding SE, and The Engine Ventures.

🌐 We’re tracking the latest developments in networking silicon. Follow our ongoing coverage at: https://convergedigest.com/category/semiconductors/

Tags: Celestial AIHOTI
ShareTweetShare
Previous Post

FS Launches 800G LPO Module 

Next Post

Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency

Jim Carroll

Jim Carroll

Editor and Publisher, Converge! Network Digest, Optical Networks Daily - Covering the full stack of network convergence from Silicon Valley

Related Posts

Hot Chips 2025: Celestial AI CTO Details In-Die Optical I/O
All

Hot Chips 2025: Celestial AI CTO Details In-Die Optical I/O

August 29, 2025
Hot Interconnects: UALink for Rack-Scale AI Interconnects
All

Hot Interconnects: UALink for Rack-Scale AI Interconnects

August 25, 2025
Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency
AI Infrastructure

Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency

August 25, 2025
Hot Interconnects: Avicena’s High-Yield MicroLED Arrays for Scale-Up AI Clusters
AI Infrastructure

Hot Interconnects: Avicena’s High-Yield MicroLED Arrays for Scale-Up AI Clusters

August 22, 2025
Hot Interconnects: GigaIO Showcases SuperNODE Scale-Up Fabric for AI
AI Infrastructure

Hot Interconnects: GigaIO Showcases SuperNODE Scale-Up Fabric for AI

August 22, 2025
Hot Interconnects: UCIe Brings On-Package Memory to the Forefront
All

Hot Interconnects: UCIe Brings On-Package Memory to the Forefront

August 22, 2025
Next Post
Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency

Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency

Categories

  • 5G / 6G / Wi-Fi
  • AI Infrastructure
  • All
  • Automotive Networking
  • Blueprints
  • Clouds and Carriers
  • Data Centers
  • Enterprise
  • Explainer
  • Feature
  • Financials
  • Last Mile / Middle Mile
  • Legal / Regulatory
  • Optical
  • Quantum
  • Research
  • Security
  • Semiconductors
  • Space
  • Start-ups
  • Subsea
  • Sustainability
  • Video
  • Webinars

Archives

Tags

5G All AT&T Australia AWS Blueprint columns BroadbandWireless Broadcom China Ciena Cisco Data Centers Dell'Oro Ericsson FCC Financial Financials Huawei Infinera Intel Japan Juniper Last Mile Last Mille LTE Mergers and Acquisitions Mobile NFV Nokia Optical Packet Systems PacketVoice People Regulatory Satellite SDN Service Providers Silicon Silicon Valley StandardsWatch Storage TTP UK Verizon Wi-Fi
Converge Digest

A private dossier for networking and telecoms

Follow Us

  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io

© 2025 Converge Digest - A private dossier for networking and telecoms.

No Result
View All Result
  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io

© 2025 Converge Digest - A private dossier for networking and telecoms.

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy and Cookie Policy.
Go to mobile version