Get the latest tech news

AMD's MI300X Outperforms Nvidia's H100 for LLM Inference


Discover if AMD's MI300X accelerator can outperform NVIDIA's H100 in real-world AI workloads. Early results are in!

To start, we focused on Mixture of Expert (MoE) architectures due to their compute efficiency and popularity – notably used by Mistral, Meta, Databricks, and X.ai for their most powerful open-source LLMs. The initial results are impressive: using MK1’s inference software, the MI300X achieves 33% higher throughput compared to the H100 SXM running vLLM on Mixtral 8x7B for a real-world chat use case. Given its impressive performance, competitive cost, and hardware availability, the MI300X with MK1 software is an excellent choice for enterprises looking to scale their AI inference capabilities.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Nvidia

Nvidia

Photo of AMD

AMD

Photo of H100

H100

Related news:

News photo

Nvidia CEO Jensen Huang has a no one-on-one meetings rule for his 55 direct reports

News photo

AMD's DC chief happy to work with Intel and others to chip away at Nvidia's AI empire

News photo

Nvidia Eclipsing Apple Threatens Radical Shakeup of $67 Billion ETF