AMD Instinct AI Accelerator Lineup Gets MI325X Refresh In Q4, 3nm MI350 “CDNA 4” In 2025, CDNA MI400 “CDNA Next” In 2026

AMD Instinct AI Accelerator Lineup Gets MI325X Refresh In Q4, 3nm MI350 “CDNA 4” In 2025, CDNA MI400 “CDNA Next” In 2026

 0
AMD Instinct AI Accelerator Lineup Gets MI325X Refresh In Q4, 3nm MI350 “CDNA 4” In 2025, CDNA MI400 “CDNA Next” In 2026

AMD has announced its brand new AI Accelerators including the Instinct MI325X "CDNA 3", MI350X "CDNA 4" & MI400 "CDNA Next" for data centers and the cloud.

It looks like AMD is accelerating its AI Instinct accelerator roadmap quite aggressively following the recent and similar announcements by NVIDIA during their keynote. The company now plans to offer a new AI accelerator, either a refresh or a completely new product, each respective year.

Starting with the first product, we have the AMD Instinct MI325X AI accelerator which will be using the same CDNA 3 architecture as the existing MI300 series. This accelerator will feature 288 GB of HBM3E memory, 6 TB/s memory bandwidth, 1.3 PFLOPs of FP16, and 2.6 PFLOPs of FP8 compute performance and will be able to handle up to 1 trillion parameters per server. Versus the NVIDIA H200, the AI accelerator will offer:

  • 2x Memory
  • 1.3x Memory Bandwidth
  • 1.3x Peak Theoretical FP16
  • 1.3x Peak Theoretical FP8
  • 2x Model Size per Server
  • The AMD Instinct MI325X AI accelerator can be seen as a beefed-up refresh of the MI300X series and it is being previewed today at Computex 2024 with a launch planned for Q4 2024. It also uses the same chiplet housing structure as the existing MI300X series but we can expect 12-Hi HBM3E sites, allowing for increased capacities.

    But AMD is also looking into the future and announced its next-gen Instinct MI350 series which will be available in 2025. AMD states that the Instinct MI350 series will be based on a 3nm process node, also offer up to 288 GB HBM3E memory, and support FP4/FP6 data types which are also supported by NVIDIA's Blackwell GPUs. These chips will be based on the next-gen CDNA 4 architecture and will be arriving with OAM compatibility.

    Lastly, AMD shares an update on its Instinct AI roadmap which now goes for an annual cadence as we mentioned above. In 2026, AMD is planning to introduce its next-gen Instinct MI400 series based on the next-gen CDNA architecture simply called "CDNA Next".

    In terms of performance, the Instinct CDNA 3 architecture is expected to bring an 8x increase over CDNA 2 while the CDNA 4 architecture is expected to offer around a 35x increase over CDNA 3GPUs. AMD Is also sharing some comparison figures against NVIDIA's Blackwell B200 GPUs. The MI350 series is expected to offer 50% more memory and 20% more compute TFLOPs than the B200 offering. NVIDIA also announced its Blackwell Ultra GPU for 2025 which should push things further so it's going to be a heated battleground within the high-end AI accelerator segment.

  • The AMD Instinct MI325X accelerator, is on track to be available in Q4 2024, which features 288GB of HBM3E memory and 6TB/s peak memory bandwidth. It will use the same Universal Baseboard infrastructure as the MI300 series accelerators.  The accelerator will have industry-leading memory capacity and bandwidth, 2x and 1.3x better than the competition respectively and 1.3x better AI compute performance than the competition
  • The accelerator will have industry-leading memory capacity and bandwidth, 2x and 1.3x better than the competition respectively and 1.3x better AI compute performance than the competition
  • The AMD CDNA 4 architecture, planned for 2025, which will power the AMD Instinct MI350 series accelerators, and deliver up to a 35x increase in AI inference performance compared to CDNA 3. The new architecture will have support for FP4 and FP6 data types and include up to 288 GB of HBM3E memory The AMD Instinct MI350x accelerator will be the first product to use this new architecture
  • The new architecture will have support for FP4 and FP6 data types and include up to 288 GB of HBM3E memory
  • The AMD Instinct MI350x accelerator will be the first product to use this new architecture
  • The AMD CDNA “Next” architecture, planned for 2026, which will power the AMD Instinct MI400 series accelerators.
  • Lastly, AMD reiterates the recent UALink (Ultra Accelerator Link) announcement from last week which is a new high-performance, open and salable AI interconnect infrastructure that is being worked upon by several vendors including Microsoft, Intel, CISCO, Broadcom, META, HPE & more. There's also the Ultra Ethernet Consortium which is being referred to as the answer for scale out AI infrastructure.

    With that said, AMD looks to have a very solid foundational roadmap for its AI endeavors as it competes against the might of NVIDIA. AMD is also calling MI300 its fastest ramping product in history with several partners and vendors currently offering them in their servers.

    What's Your Reaction?

    like

    dislike

    love

    funny

    angry

    sad

    wow