AMD Instinct™ MI300 Series Accelerators
Leadership Generative AI GPUs and APUs

Enhancing AI and HPC Capabilities

The AMD Instinct™ MI300 Series accelerators including the MI300X and MI300A, are optimally designed to handle the most challenging AI and HPC workloads. They provide outstanding computing power, substantial memory capacity, high-speed memory bandwidth, and compatibility with specialized data formats.

AMAX MI300 Series Accelerators

AceleMax DGS-424AI

3U with 4x AMD Instinct MI300A Accelerators and 4x FHFL PCIe slot for additional GPU

APUs4x AMD MI300A
Memory512 GB HBM3
Storage8x 2.5″ hot-swap NVMe drive bays

AceleMax DGS-428A

8U air cool with 8x AMD Instinct MI300X Accelerators

CPUsDual AMD EPYC 9004
GPUs8x AMD MI300X
Memory6TB 4800MT/s DDR5
Storage18x 2.5″ hot-swap NVMe drives

AceleMax DGS-214A

Liquid cooled 2U with 4x AMD Instinct MI300A Accelerators and 96 Zen4 cores

APUs4x AMD MI300A
Memory512 GB HBM3
Storage8x 2.5″ hot-swap NVMe / SAS3 / SATA3 drives

AMD Instinct MI300X Platform

The AMD Instinct MI300X Platform integrates eight MI300X GPU OAM modules using 4th-Gen AMD Infinity Fabric™ links into a standard OCP design. It offers up to 1.5TB HBM3 capacity, enabling low-latency AI processing. This plug-and-play platform streamlines market deployment and minimizes development expenses for incorporating MI300X accelerators into current AI rack and server setups.

304 CUs

304 GPU Compute Units

192 GB

192 GB HBM3 Memory

5.3 TB/s

5.3 TB/s Peak Theoretical Memory Bandwidth


Outpacing H100 AI Performance

Experience next-generation AI capabilities with the AMD Instinct MI300X. Outperforming NVIDIA's H100, the MI300X boasts superior memory capacity and bandwidth, essential for complex AI models and data-intensive tasks. With up to 1.5TB of HBM3 memory and enhanced compute performance, the MI300X is designed for efficiency in AI computations, offering a significant advantage in memory-bound applications like Large Language Models (LLMs). Embrace the future of AI processing with AMD's advanced technology, setting new benchmarks in AI acceleration.

Graph courtesy of AMD

Accelerating HPC Applications

The AMD Instinct MI300 Series accelerators, powered by the AMD CDNA™ 3 architecture, include Matrix Core Technologies and support a diverse range of precision capabilities. These accelerators excel from INT8 and FP8, with sparsity support for AI, to FP64 for more demanding tasks, significantly accelerating high-performance computing (HPC) applications with their advanced computational abilities.

Graph courtesy of AMD


AMD Instinct MI300A APUs

The AMD Instinct MI300A APUs merge AMD Instinct accelerators with AMD EPYC™ processors, featuring shared memory. This fusion enhances efficiency, flexibility, and programmability. These APUs are tailored to accelerate the merging of AI and HPC, fostering advancements in research and driving new discoveries.

228 CUs

304 GPU Compute Units

24

24 “Zen 4” x86 CPU Cores

128 GB

192 GB HBM3 Memory

5.3 TB/s

5.3 TB/s Peak Theoretical Memory Bandwidth

AMAX as your MI300X Solution Provider

At AMAX, our engineering excellence and specialization in AI and advanced liquid cooling technologies uniquely position us to maximize the potential of the AMD MI300X AI GPU. Our team's ability to craft custom computing solutions tailored to specific AI applications ensures the exceptional use of the MI300X. We maximize the enhanced memory capacity and advanced performance in FP8 and FP16 operations to create an optimized AI training environment.

Speak to one of our solution experts.Start engineering your ideal data center today.

© 2023 AMAX Engineering. All rights reserved.