• Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io
No Result
View All Result
Converge Digest
Friday, April 10, 2026
  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io
No Result
View All Result
Converge Digest
No Result
View All Result

Home » Hot Interconnects: Meta Highlights 600K GPUs, Custom Chips, Next-Gen Fabrics

Hot Interconnects: Meta Highlights 600K GPUs, Custom Chips, Next-Gen Fabrics

August 21, 2025
in AI Infrastructure, All
A A

At this week’s IEEE Hot Interconnects conference, Meta outlined the extraordinary scale and technical challenges of building one of the world’s largest AI infrastructures. Wes Bland, Research Scientist at Meta AI, described how the company is deploying GPU clusters numbering in the hundreds of thousands, custom silicon for inference, and advanced networking fabrics to power workloads ranging from recommendation engines to generative AI models such as Llama.

Meta’s AI infrastructure already spans dozens of global data centers, with facilities now reaching into the multi-gigawatt range. A flagship example is the new Prometheus supercluster in Ohio, which will rival the footprint of Manhattan in compute density. Bland noted that Meta’s GPU fleet now exceeds 600,000 H100-equivalents and is supplemented by the company’s internally designed MTIA accelerator, which delivers more than three times the compute density of earlier inference chips.

Networking remains a critical bottleneck at this scale. Meta has developed its own Disaggregated Scheduled Fabric (DSF) to provide near-optimal load balancing and credit-based traffic control, addressing the elephant flows and oscillations that strain conventional approaches. Physical scale also presents unique challenges: GPUs must be spread across large distances within a data center due to power limitations, creating latency and ordering issues that require transports more flexible than InfiniBand.

The scale of training continues to surge. Llama 4, for example, was trained on 30 trillion tokens—double the volume of Llama 3. Meta’s Tectonic data pipeline supports exabyte-scale flows, while topology-aware scheduling ensures efficient use of compute clusters.

  • Meta’s GPU fleet now exceeds 600,000 H100-equivalents
  • MTIA inference chip delivers 3.5× denser compute, 7× sparse compute efficiency
  • New Prometheus supercluster in Ohio will be among the largest data centers ever built
  • Disaggregated Scheduled Fabric addresses congestion, low-entropy traffic, and elephant flows
  • Power efficiency pursued through GPU optimization, optical interconnects, and silicon co-design

Bland stressed that power is the scarcest resource in this environment, requiring co-optimization across hardware, software, and infrastructure. The company is actively hiring engineers and researchers to push these limits further.

Tags: HOTIMeta
ShareTweetShare
Previous Post

Hot Interconnects: Arista Outlines Pathways to Energy-Efficient Optics and Liquid-Cooled Racks

Next Post

Hot Interconnects: Microsoft Maps Out $100B AI Networking Fabric

Jim Carroll

Jim Carroll

Editor and Publisher, Converge! Network Digest, Optical Networks Daily - Covering the full stack of network convergence from Silicon Valley

Related Posts

Meta’s AI Buildout Drives Record $19B Quarterly CapEx
Financials

Meta’s AI Buildout Drives Record $19B Quarterly CapEx

October 30, 2025
Microsoft signs largest corporate solar deal in U.S. at 315 MW
AI Infrastructure

ENGIE Secures New 600 MW Solar PPA with Meta for Texas Data Centers

October 28, 2025
PECC Summit: NVIDIA’s Ashkan Seyedi on AI Networking
All

PECC Summit: Meta’s Drew Alduino on AI Networking Reliability Walls

October 23, 2025
Candle Submarine Cable to Deliver 24 Fiber Pairs Across Asia-Pacific by 2028
Subsea

Candle Submarine Cable to Deliver 24 Fiber Pairs Across Asia-Pacific by 2028

September 22, 2025
Hot Interconnects: UALink for Rack-Scale AI Interconnects
All

Hot Interconnects: UALink for Rack-Scale AI Interconnects

August 25, 2025
Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency
AI Infrastructure

Hot Interconnects: Avicena’s MicroLED Links Promise Sub-1 pJ/bit Energy Efficiency

August 25, 2025
Next Post
Hot Interconnects: Microsoft Maps Out $100B AI Networking Fabric

Hot Interconnects: Microsoft Maps Out $100B AI Networking Fabric

Categories

  • 5G / 6G / Wi-Fi
  • AI Infrastructure
  • All
  • Automotive Networking
  • Blueprints
  • Clouds and Carriers
  • Data Centers
  • Enterprise
  • Explainer
  • Feature
  • Financials
  • Last Mile / Middle Mile
  • Legal / Regulatory
  • Optical
  • Quantum
  • Research
  • Security
  • Semiconductors
  • Space
  • Start-ups
  • Subsea
  • Sustainability
  • Video
  • Webinars

Archives

Tags

5G All AT&T Australia AWS Blueprint columns BroadbandWireless Broadcom China Ciena Cisco Data Centers Dell'Oro Ericsson FCC Financial Financials Huawei Infinera Intel Japan Juniper Last Mile Last Mille LTE Mergers and Acquisitions Mobile NFV Nokia Optical Packet Systems PacketVoice People Regulatory Satellite SDN Service Providers Silicon Silicon Valley StandardsWatch Storage TTP UK Verizon Wi-Fi
Converge Digest

A private dossier for networking and telecoms

Follow Us

  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io

© 2025 Converge Digest - A private dossier for networking and telecoms.

No Result
View All Result
  • Home
  • Events Calendar
  • Blueprint Guidelines
  • Privacy Policy
  • Subscribe to Daily Newsletter
  • NextGenInfra.io

© 2025 Converge Digest - A private dossier for networking and telecoms.

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy and Cookie Policy.
Go to mobile version