Converge Digest

AMD Lays Out Full-Stack Vision for AI Infrastructure

At its “Advancing AI 2025” event, AMD unveiled a complete roadmap for next-generation AI infrastructure, emphasizing open networking standards, scalable system architectures, and developer-accessible software. The announcement included the launch of the Instinct™ MI350 Series GPUs, ROCm 7 software stack, and a preview of “Helios,” a fully integrated rack-scale AI platform set to arrive in 2026. These developments underscore AMD’s commitment to disaggregated, standards-based architectures as AI workloads become increasingly distributed and network-bound.

The Instinct MI355X GPU, based on AMD’s CDNA 4 architecture, delivers up to 20 PFLOPS FP4, backed by 288GB of HBM3E memory and 8TB/s bandwidth. Systems scale up to 128 GPUs per rack in liquid-cooled configurations, achieving 2.6 exaFLOPS of AI compute and supporting models with over 500 billion parameters. ROCm 7, AMD’s open-source AI software stack, brings full support for FP4, advanced inference optimization, and turnkey MLOps tools. Together, these technologies form the foundation of AMD’s open rack-scale AI strategy.

At the heart of this infrastructure is a strong emphasis on networking innovation. AMD is a founding member of the UALink™ Consortium, which is establishing an open interconnect standard for GPU-to-GPU communications across servers and racks. The MI355X and upcoming MI400 Series GPUs are UALink-enabled, allowing up to 72 GPUs per Helios rack to operate as a unified compute domain. UALink offers 260TB/s of intra-rack bandwidth, surpassing proprietary fabrics like NVLink in scalability and openness. AMD’s roadmap also includes support for tunneling UALink over Ultra Ethernet, blending the performance of scale-up fabrics with the flexibility of Ethernet.

AMD’s Pensando™ Pollara 400 and upcoming Vulcano AI NICs are fully compliant with the Ultra Ethernet Consortium (UEC) specification, offering programmable congestion control, path-aware routing, and support for up to 800G network throughput. The Vulcano NIC, launching with the Helios rack in 2026, will deliver 8x greater scale-out bandwidth per GPU than current generation NICs. These networking components enable distributed inference and Mixture of Experts (MoE) workloads to scale with low latency and maximum throughput—critical for the future of agentic AI. AMD’s approach also leverages CXL 3.0 in its “Venice” EPYC CPUs for coherent memory sharing between CPUs and accelerators in heterogeneous deployments.

“AMD is driving AI innovation at an unprecedented pace, highlighted by the launch of our MI350 Series accelerators, our expanding ROCm software ecosystem, and the preview of our Helios rack platform,” said Dr. Lisa Su, Chair and CEO of AMD. “We are building the most open, most performant, and most scalable AI infrastructure portfolio in the industry—one that enables our customers and partners to unlock the full potential of generative and agentic AI at every level of deployment. From silicon to systems to networking, AMD is empowering a new era of open, rack-scale computing that redefines what’s possible in AI.”


Exit mobile version