Sign Up
Stories
AMD's MI300X Surpasses Nvidia's H100
Share
AI Innovation and Partnerships Unveiled
AI Innovation and Quantum Growth
AI Performance Race: Nvidia Leads
AI Chip Battle at CES
Desktop Conversion of Intel Arc A370M
Efficient LLM Inference Platform Launch
Overview
API
AMD's MI300X outperforms Nvidia's H100 by 30% in benchmarks, with a focus on FP16 using vLLM versus FP8 using TensorRT-LLM. AMD refutes Nvidia's use of selective workloads and highlights improved performance and reduced latency with TensorRT-LLM.
Ask a question
How could this performance difference affect the choice of GPUs for specific applications or industries?
How might this performance comparison impact the competition between AMD and Nvidia in the GPU market?
What implications does this have for the adoption of FP16 versus FP8 in inferencing workloads?
Article Frequency
0.2
0.4
0.6
0.8
1.0
Sep 2023
Oct 2023
Nov 2023
Coverage