Share This Article
Alvin Lang Nov 14, 2024 10:15
NVIDIA's Blackwell platform showcases exceptional performance in MLPerf Training 4.1 benchmarks, enhancing AI training capabilities with significant per-GPU improvements.
In a remarkable achievement, NVIDIA’s Blackwell platform has set new standards in the MLPerf Training 4.1 industry benchmarks, delivering outstanding results across a variety of workloads, according to NVIDIA’s blog. The platform demonstrated up to 2.2x more performance per GPU on large language model (LLM) benchmarks, notably excelling in Llama 2 70B fine-tuning and GPT-3 175B pretraining.
The Blackwell architecture’s first submission to the MLCommons Consortium emphasizes its role in advancing generative AI training performance. Key to this achievement are new kernels that optimize the use of Tensor Cores, the fundamental math operations behind many deep learning algorithms. This optimization allows Blackwell to achieve higher compute throughput per GPU while utilizing significantly larger and faster high-bandwidth memory.
Notably, the platform’s efficiency was highlighted by its ability to run the GPT-3 LLM benchmark with just 64 GPUs, maintaining exceptional per-GPU performance. In contrast, the same task required 256 GPUs on the Hopper platform, underlining Blackwell’s superior efficiency and capability.
NVIDIA continues to enhance its platforms through ongoing software development, improving performance and features for a wide range of frameworks and applications. The latest MLPerf training submissions showcased a 1.3x improvement in GPT-3 175B per-GPU training performance on Hopper since the benchmark’s introduction.
Additionally, large-scale results were achieved using 11,616 Hopper GPUs, connected via NVIDIA NVLink and NVSwitch for high-bandwidth communication, alongside NVIDIA Quantum-2 InfiniBand networking. This setup has more than tripled scale and performance on the GPT-3 175B benchmark compared to the previous year.
NVIDIA’s success is also reflected in the contributions of its partners, including major system makers and cloud service providers such as ASUSTek, Azure, Cisco, Dell, Fujitsu, and others, who submitted impressive results to MLPerf. As a founding member of MLCommons, NVIDIA emphasizes the importance of industry-standard benchmarks in AI computing, providing crucial data for companies to make informed platform investment decisions.
Through continuous advancements and optimizations, NVIDIA’s accelerated computing platforms are setting new benchmarks in AI training, offering enhanced performance and greater returns on investment for partners and customers alike.
10/27/2024 3:16:20 AM
10/27/2024 2:17:05 AM
10/27/2024 12:30:12 AM
10/26/2024 8:00:00 PM
10/26/2024 6:49:36 PM
Email us at info@blockchain.news
Welcome to your premier source for the latest in AI, cryptocurrency, blockchain, and AI search tools—driving tomorrow’s innovations today.
Disclaimer: Blockchain.news provides content for informational purposes only. In no event shall blockchain.news be responsible for any direct, indirect, incidental, or consequential damages arising from the use of, or inability to use, the information provided. This includes, but is not limited to, any loss or damage resulting from decisions made based on the content. Readers should conduct their own research and consult professionals before making financial decisions.