AMD Announces Availability of Instinct MI300X Accelerator and MI300A APU for High-Performance Computing and Generative AI

The FPS Review may receive a commission if you purchase something after clicking a link in this article.

Image: AMD

AMD has announced that the Instinct MI300X accelerator and Instinct MI300A accelerated processing unit (APU) are now available as part of the company’s new Instinct MI300 Series, aimed at high-performance computing, generative AI, and similar applications.

Here’s what AMD had to say about the Instinct MI300X, which is said to feature industry-leading memory bandwidth for generative AI and leadership performance for large language model (LLM) training and inferencing, as well as increased throughput versus the NVIDIA H100 HGX when running inference for certain LLMs:

AMD Instinct MI300X accelerators are powered by the new AMD CDNA 3 architecture. When compared to previous generation AMD Instinct MI250X accelerators, MI300X delivers nearly 40% more compute units, 1.5x more memory capacity, 1.7x more peak theoretical memory bandwidth3 as well as support for new math formats such as FP8 and sparsity; all geared towards AI and HPC workloads.

Today’s LLMs continue to increase in size and complexity, requiring massive amounts of memory and compute. AMD Instinct MI300X accelerators feature a best-in-class 192 GB of HBM3 memory capacity as well as 5.3 TB/s peak memory bandwidth to deliver the performance needed for increasingly demanding AI workloads. The AMD Instinct Platform is a leadership generative AI platform built on an industry standard OCP design with eight MI300X accelerators to offer an industry leading 1.5TB of HBM3 memory capacity. The AMD Instinct Platform’s industry standard design allows OEM partners to design-in MI300X accelerators into existing AI offerings and simplify deployment and accelerate adoption of AMD Instinct accelerator-based servers.

Compared to the Nvidia H100 HGX, the AMD Instinct Platform can offer a throughput increase of up to 1.6x when running inference on LLMs like BLOOM 176B and is the only option on the market capable of running inference for a 70B parameter model, like Llama, on a single MI300X accelerator; simplifying enterprise-class LLM deployments and enabling outstanding TCO.

And here’s the word on the Instinct MI300A, which AMD has noted as being one of the first products of its kind:

The AMD Instinct MI300A APUs, the world’s first data center APU for HPC and AI, leverage 3D packaging and the 4th Gen AMD Infinity Architecture to deliver leadership performance on critical workloads sitting at the convergence of HPC and AI. MI300A APUs combine high-performance AMD CDNA 3 GPU cores, the latest AMD “Zen 4” x86-based CPU cores and 128GB of next-generation HBM3 memory, to deliver ~1.9x the performance-per-watt on FP32 HPC and AI workloads, compared to previous gen AMD Instinct MI250X.

Energy efficiency is of utmost importance for the HPC and AI communities, however these workloads are extremely data- and resource-intensive. AMD Instinct MI300A APUs benefit from integrating CPU and GPU cores on a single package delivering a highly efficient platform while also providing the compute performance to accelerate training the latest AI models. AMD is setting the pace of innovation in energy efficiency with the company’s 30×25 goal, aiming to deliver a 30x energy efficiency improvement in server processors and accelerators for AI-training and HPC from 2020-2025.

The APU advantage means that AMD Instinct MI300A APUs feature unified memory and cache resources giving customers an easily programmable GPU platform, highly performant compute, fast AI training and impressive energy efficiency to power the most demanding HPC and AI workloads.

AMD Instinct MI300 Series Specifications

AMD InstinctArchitectureGPU CUsCPU CoresMemoryMemory Bandwidth
(Peak theoretical)
Process Node3D Packaging w/ 4th Gen AMD Infinity Architecture
MI300AAMD CDNA 322824 “Zen 4”128GB HBM35.3 TB/s5nm / 6nmYes
MI300XAMD CDNA 3304N/A192GB HBM35.3 TB/s5nm / 6nmYes
PlatformAMD CDNA 32,432N/A1.5 TB HMB35.3 TB/s per OAM5nm / 6nmYes

“AMD Instinct MI300 Series accelerators are designed with our most advanced technologies, delivering leadership performance, and will be in large scale cloud and enterprise deployments,” said Victor Peng, president, AMD. “By leveraging our leadership hardware, software and open ecosystem approach, cloud providers, OEMs and ODMs are bringing to market technologies that empower enterprises to adopt and deploy AI-powered solutions.”

Join the discussion in our forums...

Tsing Mui
News poster at The FPS Review.

Recent News