Published: Jun 12, 2024

AMD’s MI300X Outperforms NVIDIA’s H100 for LLM Inference