Converge Digest

Intel Debuts its Xeon Scalable Platform

In what it called its “biggest data center launch in a decade”, Intel officially unveiled its Xeon Scalable platform, a new line of server CPUs based codenamed Skylake and specifically designed for evolving data center and network infrastructure.

The new silicon, which Intel has been refining for the past five years, promises the highest core and system-level performance averaging 1.65x higher performance over the prior generation.  First shipments went out several months ago and are now in commercial use at over 30 customers worldwide, including AT&T, Amazon Web Services and Google.  Intel says every aspect of Xeon has been improved or redesigned: brand new core, cache, on-die interconnects, memory controller and hardware accelerators.

Intel’s new processors scale up to 28 cores and will be offered in four classes: Platinum, Gold, Silver, and Bronze. The design boasts six memory channels versus four memory channels of previous generation for memory-intensive workloads. Up to three Intel Ultra Path Interconnect (Intel UPI) channels provide increase scalability of the platform to as many as eight sockets.

Intel claims 4.2X greater VM capacity than its previous generation and a 65% lower total cost of ownership over a 4-year old server.  Potentially you might need only one quarter of the number of servers. For communication service providers, the claim is that the new Xeon Gold will deliver a 2.7X performance boost for DPDK L3 forwarding applications over a 4-year old server.

Key innovations in Xeon Scalable Platform

Aiming for the megatrends

In a webcast presentation, Navin Shenoy, Exec Vice President & General Manager, Intel’s Data Center Group, said that as traditional industries turn to technology to reinvent themselves, there are three megatrends that Intel is pursuing: Cloud, AI & Analytics, and 5G.  The new Xeon Scalable Platform addresses the performance, security and agility challenges for each of these megatrends.

AT&T’s John Donovan testifies, performance boost about 30%

During the big Xeon Scalable unveiling, Intel invited AT&T’s John Donovan on stage to talk about the new processors/ AT&T gained access to the new processors a few months ago and has already deployed Xeon Scalable servers which are carrying production traffic.  Donovan reported about at 30% performance boost for its applications over the previous Xeon generation. The net effect he said should be a 25% reduction in the number of servers it will need to deploy.  Intel has been seeding the process with other top customers as well.

This 30% performance boost is certainly good, but it is probably a stretch to call this upgrade “the biggest data center announcement in a decade.” For other applications, perhaps the claim is better justified. One such area is machine learning, which Intel identifies as one of the key megatrends for the industry. There are some interesting developments for Xeons in this domain.

A strong market position

Google Cloud Platform (GCP) is the first public cloud to put the Intel Xeon Scalable Platform into commercial operation. A partnership between Google and Intel was announced earlier this year at a Google event where the companies said they are collaborating in other areas as well, including hybrid cloud orchestration, security, machine and deep learning, and IoT edge-to-cloud solutions. Intel is also a backer of Google’s Tensor Flow and Kubernetes open source initiatives.

In May 2016, Google announced the development of a custom ASIC for Tensor Flow processing. These TPUs are already in service in Google data centres where they “deliver an order of magnitude better-optimized performance per watt for machine learning.” For Intel, this poses a long-term strategic threat.  With this announcement, Intel said Xeon’s onboard advanced Vector Extensions 512 (Intel AVX-512) can increase machine learning inference performance by over 100x – a huge boost for AI developers.

The data centre server market is currently dominated by Intel.  Over the years, there have been several attempts by ARM to gain at least a toe-hold of market share in data centre servers, but so far, the impact has been very limited.  AMD recently announced its EPYC processor for data centre servers, but no shipment date has been stated and the current market position is zero. NVIDIA has been gaining traction in AI applications as well as in public cloud acceleration for GPU intensive applications – but these are specialized use cases.

Exit mobile version