In the ever-evolving world of artificial intelligence, performance and efficiency are paramount. The ability to train and deploy AI models quickly and cost-effectively has become a competitive advantage for organizations across various industries. Intel, a pioneer in the field of semiconductor technology, continues to push the boundaries of AI performance with its Intel Gaudi2 accelerator and 4th Gen Intel Xeon Scalable processors. In a recent development, Intel has achieved a remarkable 2x performance leap on the GPT-3 benchmark by implementing FP8 software. This achievement, validated through the industry-standard MLPerf training v3.1 benchmark, underscores Intel’s commitment to providing competitive AI solutions that can be deployed anywhere.
The Milestone Announcement
On November 8, 2023, Intel announced the groundbreaking results of its MLPerf training v3.1 benchmark for training AI models. These results encompassed Intel’s Gaudi2 accelerators and 4th Gen Intel Xeon Scalable processors equipped with Intel Advanced Matrix Extensions (Intel AMX). The standout performance came from Intel Gaudi2, which demonstrated an impressive 2x performance improvement thanks to the implementation of the FP8 data type on the v3.1 training GPT-3 benchmark. This accomplishment reaffirms Intel’s dedication to making AI accessible and efficient for a wide range of applications.
Sandra Rivera, Intel’s Executive Vice President and General Manager of the Data Center and AI Group, highlighted the significance of this achievement, stating, “We continue to innovate with our AI portfolio and raise the bar with our MLPerf performance results in consecutive MLCommons AI benchmarks. Intel Gaudi and 4th Gen Xeon processors deliver a significant price-performance benefit for customers and are ready to deploy today. Our breadth of AI hardware and software configuration offers customers comprehensive solutions and choice tailored for their AI workloads.”