AMD Expands AI/HPC Product Lineup With Flagship GPU-only Intuition MI300X with 192GB Reminiscence
Alongside their EPYC server CPU updates, as a part of right this moment’s AMD Knowledge Heart occasion, the corporate can also be providing an replace on the standing of their nearly-finished AMD Intuition MI300 accelerator household. The corporate’s next-generation HPC-class processors, which use each Zen 4 CPU cores and CDNA 3 GPU cores on a single bundle, have now turn out to be a multi-SKU household of XPUs.
Becoming a member of the beforehand introduced 128GB MI300 APU, which is now being referred to as the MI300A, AMD can also be producing a pure GPU half utilizing the identical design. This chip, dubbed the MI300X, makes use of simply CDNA 3 GPU tiles slightly than a mixture of CPU and GPU tiles within the MI300A, making it a pure, high-performance GPU that will get paired with 192GB of HBM3 reminiscence. Aimed squarely on the giant language mannequin market, the MI300X is designed for patrons who want all of the reminiscence capability they will get to run the most important of fashions.
First announced back in June of last year, and detailed in greater depth back at CES 2023, the AMD Intuition MI300 is AMD’s huge play into the AI and HPC market. The distinctive, server-grade APU packs each Zen 4 CPU cores and CDNA 3 GPU cores on to a single, chiplet-based chip. None of AMD’s rivals have (or may have) a mixed CPU+GPU product just like the MI300 sequence this yr, so it offers AMD an fascinating answer with a very united reminiscence structure, and loads of bandwidth between the CPU and GPU tiles.
MI300 additionally consists of on-chip reminiscence by way of HBM3, utilizing 8 stacks of the stuff. On the time of the CES reveal, the best capability HBM3 stacks have been 16GB, yielding a chip design with a most native reminiscence pool of 128GB. Nonetheless, because of the latest introduction of 24GB HBM3 stacks, AMD is now going to have the ability to provide a model of the MI300 with 50% extra reminiscence – or 192GB. Which, together with the extra GPU chiplets discovered on the MI300X, are meant to make it a powerhouse for processing the most important and most advanced of LLMs.
Underneath the hood, MI300X is definitely a barely easier chip than MI300A. AMD has changed MI300A’s trio of CPU chiplets with simply two CDNA 3 GPU chiplets, leading to a 12 chiplet design general – 8 GPU chiplets and what seems to be one other 4 IO reminiscence chiplets. In any other case, regardless of excising the CPU cores (and de-APUing the APU), the GPU-only MI300X appears so much just like the MI300A. And clearly, AMD is aiming to make the most of the synergy in providing each an APU and a flagship CPU in the identical bundle.
Uncooked GPU efficiency apart (we have no onerous numbers to talk of proper now), a bit a part of AMD’s story with the MI300X goes to be reminiscence capability. Simply providing a 192GB chip by itself is a giant deal, on condition that reminiscence capability is the constraining issue for the present era of enormous language fashions (LLMs) for AI. As we’ve seen with latest developments from NVIDIA and others, AI prospects are snapping up GPUs and different accelerators as shortly as they will get them, all of the whereas demanding extra reminiscence to run even bigger fashions. So with the ability to provide a large, 192GB GPU that makes use of 8 channels of HBM3 reminiscence goes to be a large benefit for AMD within the present market – at the least, as soon as MI300X begins delivery.
The MI300 household stays on monitor to ship in some unspecified time in the future later this yr. Based on AMD, the 128GB MI300A APU is already sampling to prospects now. In the meantime the 192GB MI300X GPU will probably be sampling to prospects in Q3 of this yr.
It additionally goes with out saying that, with this announcement, AMD has solidified that they are doing a versatile XPU design at the least 3 years earlier than rival Intel. Whereas Intel scrapped their mixed CPU+GPU Falcon Shores product for a pure GPU Falcon Shores, AMD is now slated to supply a versatile CPU+GPU/GPU-only product as quickly as the top of this yr. On this timeframe, it will likely be going up in opposition to merchandise akin to NVIDIA’s Grace Hopper superchip, which though is not an APU/XPU both, comes very shut by linking up NVIDIA’s Grace CPU with a Hopper GPU by way of a excessive bandwidth NVLink. So whereas we’re ready on additional particulars on MI300X, it ought to make for a really fascinating battle between the 2 GPU titans.
Total, the stress on AMD close to the MI300 household is important. Demand for AI accelerators has been by means of the roof for a lot of the previous yr, and MI300 will probably be AMD’s first alternative to make a major play for the market. MI300 is not going to fairly be a make-or-break product for the corporate, however in addition to getting the technical benefit of being the primary to ship a single-chip server APU (and the bragging rights that include it), it’s going to additionally give them a recent product to promote right into a market that’s shopping for up all of the {hardware} it might probably get. In brief, MI300 is predicted to be AMD’s license to print cash (ala NVIDIA’s H100), or so AMD’s keen traders hope.
AMD Infinity Structure Platform
Alongside right this moment’s 192GB MI300X information, AMD can also be briefly saying what they’re calling the AMD Infinity Structure Platform. That is an 8-way MI300X design, permitting for as much as 8 of AMD’s top-end GPUs to be interlinked collectively to work on bigger workloads.
As we’ve seen with NVIDIA’s 8-way HGX boards and Intel’s own x8 UBB for Ponte Vecchio, an 8-way processor configuration is at present the candy spot for high-end servers. That is each for bodily design causes – room to put the chips and room to route cooling by means of them – in addition to the very best topologies which are obtainable to hyperlink up numerous chips with out placing too many hops between them. If AMD is to go toe-to-toe with NVIDIA and to seize a part of the HPC GPU market, then that is yet another space the place they’re going to want to match NVIDIA’s {hardware} choices
AMD is looking the Infinity Structure Platform an “industry-standard” design. Accoding to AMD, they’re utilizing an OCP server platform as their base right here; and whereas this means that MI300X is utilizing an OAM kind issue, we’re nonetheless ready to get specific affirmation of this.