AMD Instinct™ accelerator
of the MI300 series

Leading generative AI accelerators and data center APUs

Exponential data growth and the increasing need for real-time results are the main drivers behind advances in the industry, which is constantly reaching new frontiers of innovation to meet the diverse and complex demands of AI and HPC applications. Both AI and HPC require an architecture with high computing power, large memory and high bandwidth, but their utilization differs slightly.

For AI applications, high computational speed with less accuracy and large memory is critical to train models, make predictions and scale generative AI. In contrast, HPC applications require precise calculations and resources for processing large amounts of data and complex simulations, which are essential for scientific discovery.

The all-new AMD Instinct™ MI300 series with AMD CDNA™ 3 architecture offers customized solutions for the specific needs of AI and HPC. With the updated AMD ROCm™ 6 software platform, partners can quickly develop scalable solutions for a wide range of modern workloads, tackling some of the world's most pressing challenges in the most efficient way.

Platform of the AMD Instinct™ MI300 series

One data center APU (MI300A) and one separate GPU (MI300X)

AMD Instinct™ MI300A APU:
specially developed HPC APU

The AMD Instinct MI300A APU accelerator is optimized for processing large data sets and is ideal for compute-intensive modeling and analytics. It utilizes an innovative 3D chiplet design that combines 3D stacked "Zen 4" x86 CPUs and AMD CDNA 3 GPU XCDs with high-bandwidth memory (HBM). With 24 CPU cores and 14,592 GPU stream processors, this advanced architecture delivers breakthrough performance, density and energy efficiency for accelerated HPC and AI applications. The MI300A will be deployed in large and scalable data centers and supercomputers worldwide, including the two-exaflop El Capitan supercomputer, which will be among the fastest in the world. This will enable advances in various fields such as healthcare, energy, climate science, transportation and scientific research.

AMD Instinct™ MI300X accelerator:
developed for state-of-the-art AI

The AMD Instinct MI300X accelerator is specifically designed for large-scale language models and other advanced AI applications that require training and inference on large datasets. Compared to the MI300A, the MI300X replaces three "Zen 4" CPU chiplets with two additional AMD CDNA 3 XCD chiplets and adds 64 GB of HBM3 memory, increasing memory up to 192 GB. These optimizations enable improved execution of larger AI models directly in memory, allowing cloud providers and enterprises to run more inference jobs per GPU, reduce the number of GPUs required, increase inference performance and reduce total cost of ownership.

Our recommendation

sysGen/SUPERMICRO Universal 8U Dual AMD EPYC™ System
# AS-8125GS-TNMR2
​​​​​​​

o Barebone AS-8125GS-TNMR2
o 2x AMD EPYC™ BERGAMO 9754, 128-Core
   128C/256T, 2.25-3.10 GHz, 256MB L3 Cache,
o 3TB (24x 128GB) DDR5 4800MHz ECC reg.
o 4x 3.84TB Gen 4. NVME SSDs
o 1x AMD Instinct™ MI300X – 192 GB GPU:
   Universal baseboard (UBB) module with 8 AMD
   Instinct™ MI300X OAM GPUs - 1536 GB total memory
o 2x 10GbE RJ45 AIOM Card
   1 RJ45 Dedicated IPMI LAN port
o 3 years warranty included
Preis auf Anfrage
​​​​​​​

Anfrage

AMD CDNA™ 3 architecture

The AMD Instinct MI300 series accelerates based on the AMD CDNA™ 3 architecture and integrates matrix core technologies. This series supports various accuracies, from highly efficient INT8 and FP8 to demanding FP64 for HPC applications. The architecture enables heterogeneous integration and delivers outstanding performance through advanced packaging. A coherent programming model seamlessly connects CPUs and GPUs to provide customers with the density and energy efficiency to tackle complex problems.

The architecture is based on a heterogeneous package with up to eight vertically stacked compute chips and four I/O chips connected by the fourth-generation Infinity architecture. The integration of eight high-bandwidth stacks enables improved performance, efficiency and programmability.

The AMD Instinct MI300 series offers CPU and GPU chiplet variants to accommodate different workloads. The AMD Instinct MI300A APU combines three "Zen 4" x86 CPUs with six 3rd generation AMD CDNA architecture GPU accelerator compute chips. This APU utilizes a shared memory pool and AMD Infinity Cache™, making it the world's first high-performance hyper-scale APU.

series overview

Discover AMD Instinct MI300X accelerators, the AMD Instinct MI300X platform and AMD Instinct MI300A APUs.

AMD Instinct MI300X accelerator

AMD Instinct MI300X Series accelerators are designed to deliver leading performance for generative AI workloads and HPC applications.

View specifications

AMD Instinct MI300X platform

The AMD Instinct MI300X platform integrates 8 fully connected MI300X GPU OAM modules into an industry standard OCP design via 4th generation AMD Infinity Fabric™ interconnects, delivering up to 1.5TB of HBM3 capacity for low latency AI processing. This out-of-the-box platform can accelerate time-to-market and reduce development costs when MI300X accelerators are added to an existing AI rack and server infrastructure.

View specifications

AMD Instinct MI300A APUs

AMD Instinct MI300A accelerated processing units (APUs) combine the power of AMD Instinct accelerators and AMD EPYC™ processors with shared memory to enable advanced efficiency, flexibility and programmability. They are designed to accelerate the convergence of AI and HPC to advance research and new discoveries.

View specifications

Proven performance on a large scale

New optimized data center solutions with AMD Instinct MI300 series accelerators

Supermicro introduces three new GPU systems with AMD Instinct™ MI300 Series accelerators to drive AI and HPC infrastructure at scale. Based on Supermicro's proven AI building block system architecture, the new 8U 8-GPU system with MI300X accelerators optimizes large-scale deployment for the largest AI models and shortens lead time. In addition, Supermicro's 4U and liquid-cooled 2U 4-Way systems supporting MI300A APUs that combine CPUs and GPUs leverage Supermicro's expertise in multiprocessor system architecture and cooling design that are precisely tuned for the convergence of AI and HPC.

Industry-proven system designs

8U high performance 8 GPU fabric system utilizing the industry standard OCP Accelerator Module (OAM) to support AMD's MI300X accelerators. 4U and 2U multiprocessor APU systems integrate 4 AMD Instinct™ MI300A accelerators.

Specially developed for AI and HPC

Maximized features and optimized power consumption, support for up to 16 dedicated hot-swap NVMe drives, high-performance GPUs, CPUs and memory, and high-speed networking for large-scale cluster deployments.

Extended cooling options

Flexible cooling options for air-cooled and liquid-cooled racks, with liquid-cooled solutions offering exceptional TCO with over 51% energy cost savings for data centers.

Scalable design

Designed with full scalability in mind, the system supports 8 high-speed 400G network cards, providing a direct connection to each GPU for massive AI clusters.

ServerOverview

Configuration
Form factor
CPU
GPU
Memory
Drives
Networking
Connect
Electricity
Cooling

AS -8125GS-TNMR2 (8U)

AI training and inference on a large scale

8U 8-GPU system with AMD Instinct MI300X accelerators

8U 8-GPU system with AMD Instinct MI300X accelerators (air-cooled)
Two AMD EPYC 9004 series processors with up to 128 cores/256 threads per socket
8x AMD Instinct MI300X accelerators with 192 GB HBM3 memory per GPU, interconnected on AMD Universal Base Board (UBB 2.0)
Up to 6TB (24x 256GB DRAM) 4800MT/s ECC DDR5 RDIMM/LRDIMM
16x hot-swap PCIe 5.0 U.2 NVMe, 1x onboard M.2 NVMe, 2x 2.5" SATA
8x PCIe 5.0 high-performance network cards, up to 400G with Ethernet or InfiniBand
AMD Infinity Fabric™ links, up to 896 GB/s peer-to-peer connectivity between the GPUs
6x or 8x 3000W redundant Titanium Level power supply units
Air cooling

LIKE -2145GH-TNMR (2U)

HPC and supercomputing for companies

Liquid-cooled 2U Quad-APU system with AMD Instinct MI300A accelerators

2U 4-GPU system with AMD Instinct MI300A accelerators (liquid-cooled)
Four AMD Instinct MI300A APUs with a total of 96 CPU cores (4x 24 AMD "Zen4" cores)
4x AMD Instinct MI300A with 228 AMD CDNA 3 GPU compute units per APU
512 GB standardized HBM3 memory with up to 5.3 TB/s bandwidth
2x onboard 2280 or 22110 M.2 NVMe and 8x hot-swap 2.5" U.2 NVMe
2x AIOM (OCP 3.0) with up to 400G and additional 4x PCIe 5.0 (x8) slots
AMD Infinity Fabric Links, up to 384 GB/s peer-to-peer connections between the GPUs
4x 1600W redundant Titanium Level power supply units
Liquid cooling

LIKE -4145GH-TNMR (4U)

Converged HPC-AI and scientific computing
Air-cooled 4U Quad-APU system with AMD Instinct MI300A accelerators

4U 4-GPU system with AMD Instinct MI300A accelerators (air-cooled)
Four AMD Instinct MI300A APUs with a total of 96 CPU cores (4x 24 AMD "Zen4" cores)
4x AMD Instinct MI300A with 228 AMD CDNA 3 GPU compute units per APU
512 GB standardized HBM3 memory with up to 5.3 TB/s bandwidth
2x onboard 2280 or 22110 M.2 NVMe and 8x hot-swap 2.5" U.2 NVMe or 24x 2.5" SAS/SATA
2x AIOM (OCP 3.0) with up to 400G and additional 4x PCIe 5.0 (x8) slots
AMD Infinity Fabric Links, up to 384 GB/s peer-to-peer connections between the GPUs
4x 1600W redundant Titanium Level power supply units
Air cooling

AMD ROCm™ open source software platform

The AMD ROCm™ 6 open source platform is designed to maximize the performance of AMD Instinct MI300 accelerators for HPC and AI workloads while maintaining compatibility with popular software frameworks. ROCm provides a variety of drivers, development tools and APIs for GPU programming from low-level kernels to end-user applications and can be customized to meet individual requirements. Developers can create, test and deploy applications in a free, secure open source ecosystem. The software is portable and enables seamless switching between different accelerators and inter-GPU connectivity architectures. ROCm is ideal for HPC, AI, scientific computing and CAD. Advanced GPU software containers and deployment guides for HPC, AI and machine learning are available on the Infinity Hub to accelerate system deployments and shorten time-to-insight

Supercharging AI and HPC

AMD Instinct™ MI300 series accelerators are ideal for the most demanding AI and HPC workloads. They offer exceptional compute performance, high memory density, high bandwidth memory and support for specialized data formats.

Conclusion

The AMD Instinct™ MI300 series marks a significant advancement in delivering powerful solutions for the needs of AI and HPC applications.

The AMD Instinct™ MI300 series is equipped with the third generation Compute DNA architecture (AMD CDNA™ 3) and offers two different variants specifically tailored to the needs of AI and HPC applications. These accelerators deliver the compute power, memory capacity and memory bandwidth needed to train complex models, make accurate predictions and process large amounts of data for scientific discovery.

sysGen's expertise and commitment to innovation enable us to provide customized solutions that meet your needs and help you achieve your goals efficiently. Contact us today to learn more about the AMD Instinct™ MI300 series and how it can benefit your applications.



Learn more
Learn more about the 3rd generation AMD CDNA™ architecture.
Learn more about the latest AMD Instinct™ MI300 series accelerators
Visit the AMD Infinity Hub to learn more about AMD Instinct™ supported containers.
Learn more about the AMD ROCm™ open software platform .