AMD Announces 7nm Data Center GPUs for AI, Cloud and HPC

 

AMDAMD has announced the launch of its AMD Radeon Instinct MI60 and MI50 accelerators “with supercharged compute performance, high-speed connectivity, fast memory bandwidth and updated ROCm open software platform. The 7nm data center GPUs are designed to power the most demanding deep learning, HPC, cloud and rendering applications.

The AMD Radeon Instinct MI60 and MI50 accelerators feature flexible mixed-precision capabilities, powered by high-performance compute units that expand the types of workloads these accelerators can address, including a range of HPC and deep learning applications. The new AMD Radeon Instinct MI60 and MI50 accelerators were designed to efficiently process workloads such as rapidly training complex neural networks, delivering higher levels of floating-point performance, greater efficiencies and new features for data center and departmental deployments.

AMD David Wang
“Legacy GPU architectures limit IT managers from effectively addressing the constantly evolving demands of processing and analyzing huge datasets for modern cloud data center workloads,” said David Wang, senior vice president of engineering, Radeon Technologies Group at AMD.

The AMD Radeon Instinct MI60 and MI50 accelerators would provide ultra-fast floating-point performance and hyper-fast HBM2 (second-generation High-Bandwidth Memory) with up to 1 TB/s memory bandwidth speeds. The GPUs are also capable of supporting “next-generation” PCIe 4.0 interconnect, and feature AMD Infinity Fabric Link GPU interconnect technology that enables GPU-to-GPU communications.

Cloud Workloads

AMD also announced a new version of the ROCm open software platform for accelerated computing that supports the architectural features of the new accelerators, including optimized deep learning operations (DLOPS) and the AMD Infinity Fabric Link GPU interconnect technology. Designed for scale, ROCm would allow customers to deploy “high-performance, energy-efficient” heterogeneous computing systems in an open environment.

“Legacy GPU architectures limit IT managers from effectively addressing the constantly evolving demands of processing and analyzing huge datasets for modern cloud data center workloads,” said David Wang, senior vice president of engineering, Radeon Technologies Group at AMD. “Combining world-class performance and a flexible architecture with a robust software platform and the industry’s leading-edge ROCm open software ecosystem, the new AMD Radeon Instinct accelerators provide the critical components needed to solve the most difficult cloud computing challenges today and into the future.”

Key features of the AMD Radeon Instinct MI60 and MI50 accelerators include:

  • Optimized Deep Learning Operations – Provides flexible mixed-precision FP16, FP32 and INT4/INT8 capabilities to meet growing demand for dynamic and ever-changing workloads, from training complex neural networks to running inference against those trained networks.
  • World’s Fastest Double Precision PCIe Accelerator – The AMD Radeon Instinct MI60 is a double precision PCIe 4.0 capable accelerator, delivering “up to 7.4 TFLOPS peak FP64 performance” allowing scientists and researchers to more efficiently process HPC applications across a range of industries including life sciences, energy, finance, automotive, aerospace, academics, government, defense and more. The AMD Radeon Instinct MI50 would deliver up to 6.7 TFLOPS FP64 peak performance, while providing “an efficient, cost-effective” solution for a variety of deep learning workloads, as well as enabling high reuse in Virtual Desktop Infrastructure (VDI), Desktop-as-a-Service (DaaS) and cloud environments.
  • Up to 6X Faster Data Transfer – Two Infinity Fabric Links per GPU deliver up to 200 GB/s of peer-to-peer bandwidth while enabling the connection of up to 4 GPUs in a hive ring configuration (2 hives in 8 GPU servers).
  • Ultra-Fast HBM2 Memory – The AMD Radeon Instinct MI60 provides 32GB of HBM2 Error-correcting code (ECC) memory, while the Radeon Instinct MI50 provides 16GB of HBM2 ECC memory. Both GPUs provide full-chip ECC and Reliability, Accessibility and Serviceability (RAS) technologies, which are critical to deliver more accurate compute results for large-scale HPC deployments.
  • Secured Virtualized Workload Support – As a hardware-based GPU virtualization solution, which is based on the industry-standard SR-IOV (Single Root I/O Virtualization) technology, AMD MxGPU would make it difficult for hackers to attack at the hardware level, helping provide security for virtualized cloud deployments.

New Version ROCm

AMD has also announced a new version of its ROCm open software platform designed to speed development of high-performance, energy-efficient heterogeneous computing systems.

In addition to support for the new Radeon Instinct accelerators, ROCm software version 2.0 would provide updated math libraries for the new DLOPS; support for 64-bit Linux operating systems including CentOS, RHEL and Ubuntu; optimizations of existing components; and support for the latest versions of the most popular deep learning frameworks, including TensorFlow 1.11, PyTorch (Caffe2) and others.

Availability

The AMD Radeon Instinct MI60 accelerator is expected to ship to data center customers by the end of 2018.

The AMD Radeon Instinct MI50 accelerator is expected to begin shipping to data center customers by the end of Q1 2019.

The ROCm 2.0 open software platform is expected to be available by the end of 2018.

Cloudfest 2019