skip to main content

Lenovo Leads the Pack in MLPerf Inference: Datacenter 5.0 Benchmarks with Industry-Best AI Performance

Article

Home
Top
Published
23 Jun 2025
Form Number
LP2240
PDF size
6 pages, 600 KB

Abstract

MLCommons® announced new results for its industry-standard MLPerf® Inference v5.0 benchmark suite, which provides insights into machine learning (ML) system performance benchmarking. This article explores Lenovo’s contributions to MLPerf inferencing 5.0. The combination of recent hardware and software advances optimized for generative AI have led to dramatic performance improvements over the past year.

Introduction

The release of MLPerf Inference: Datacenter 5.0 benchmark suite by MLCommons marks another significant milestone in the evaluation of machine learning performance. By measuring how fast systems can process inputs and produce results using a trained model, these comprehensive benchmarks provide insights into datacenter AI capabilities, focusing on inference across diverse hardware configurations. We are proud to participate and are excited to share our outstanding achievements from the latest Lenovo ThinkSystem portfolio including SR650a V4, SR680a V3 and the SR780a V3.

With groundbreaking performance results, Lenovo shines as a global leader in AI infrastructure. In the latest MLPerf Inference benchmarks, three of Lenovo’s powerful ThinkSystem servers—SR650a V4, SR680a V3, and SR780a V3—dominated across a wide spectrum of AI workloads. These systems, equipped with cutting-edge NVIDIA GPUs, delivered top-tier results in critical benchmarks using models such as GPT-J, Llama 2, Mixtral, Stable Diffusion, ResNet50, and more. Notably, the ThinkSystem SR650a V4 emerged as the top-performing general-purpose AI inferencing system, delivering unmatched versatility for both datacenter and colocation AI deployments.

Highlights from MLPerf Inference: Data Center Performance

Key highlights from the MLPerf results include benchmarks using these servers:

  • SR650a V4 - 2U server with for 4x NVIDIA H100 NVL PCIe GPUs
  • SR680a V4 - 8U air-cooled server with 8x NVIDIA H200 SXM GPUs
  • SR780a V4 - 5U water-cooled server with 8x NVIDIA H200 SXM GPUs

ThinkSystem SR650a V4 with 4x NVIDIA H100-NVL-94GB

Tested with GPT-J, Mixtral, Stable Diffusion and others takes 1st place as general purpose for AI inferencing systems and use cases. Key tests included:

  • Vision Tasks: ResNet50 and RetinaNet (both server & offline)
  • Medical Imaging: 3D-UNet (99 and 99.9 offline)
  • NLP & Generative AI: GPT-J (99.9 both server & offline; 99 offline)
  • Diffusion Models: Stable Diffusion XL (both server & offline)
  • Mixture of Experts: Mixtral-8x7B (both server & offline)

Its consistent top-tier performance across diverse use cases demonstrates its versatility and efficiency, especially for enterprises seeking robust inferencing capabilities.

Lenovo ThinkSystem SR650a V4
Figure 1. Lenovo ThinkSystem SR650a V4

ThinkSystem SR680a V3 with 8x NVIDIA H200-SXM-141GB

One of Lenovo’s high-performance computing servers excelled in large language model workloads, taking 1st place in:

  • LLMs: Llama 2-70B (99 and 99.9 offline)
  • Vision: ResNet50 offline

Designed to maximize GPU performance, SR680a achieves 2nd and 3rd places in interactive LLM inferencing, Stable Diffusion XL, Mixtral-8x7B, and RGAT. This balance of top-end performance and flexibility makes the SR680a V3 ideal for demanding AI-heavy enterprises and research environments.

Lenovo ThinkSystem SR680a V3
Figure 2. Lenovo ThinkSystem SR680a V3

ThinkSystem SR780a V3 with 8x NVIDIA H200-SXM-141GB

With dual 5th Gen Intel® Xeon® Scalable processors, the ThinkSystem SR780a V3 offers the performance needed for compute-demanding AI and HPC workloads, proudly achieves first place as most advanced AI and HPC Inferencing systems, the SR780a V3 also impressed across a wide range of benchmarks, with 1st place finishes in:

  • NLP & LLMs: GPT-J and Llama 2-70B (interactive offline)
  • Mixture of Experts: Mixtral-8x7B offline
  • Graph Neural Networks: RGAT offline

Strongly performing for vision and diffusion model workloads, achieving 2nd and 3rd place, underlining its strength as a general-purpose inferencing powerhouse for advanced AI applications.

Lenovo ThinkSystem SR780a V3
Figure 3. Lenovo ThinkSystem SR780a V3

Lenovo Hybrid AI Advantage with NVIDIA

Lenovo Hybrid AI Advantage™ with NVIDIA help organizations improve productivity, increase agility, and innovate with trust through standardized and accelerated development and deployment of AI use case solutions. Lenovo Hybrid AI Advantage bring the power of Lenovo AI library and validated, tested hybrid AI factories (hybrid AI platforms, workstations, servers, storage, network, software, models, services, partner ecosystem) to the enterprises.

The hybrid AI factory is designed to support hybrid deployments at the Edge, data centers, Colos, and business locations with cloud integration. It offers flexibility of model, infrastructure choice, enables a wide range of AI applications, agentic and machine learning workflows, and real-time data analysis. Lenovo’s hybrid AI platforms power the hybrid AI factory, and they can scale from a single server with just four GPUs as starter environment to a rack scalable unit (SU) as a turnkey infrastructure solution with partner technology choice.

Future-focused AI with Lenovo

As generative AI, large language models, and inferencing continue to evolve, Lenovo is an established industry leader. These benchmark-topping results not only validate current capabilities but also demonstrate our commitment to AI readiness today and in the future—whether in the datacenter or cloud. Our partnership ecosystem with industry leaders like NVIDIA Lenovo is building the foundation for a more intelligent, connected, and autonomous future.

Lenovo’s leadership in the latest MLPerf Inference results showcases more than just benchmark dominance—it emphasizes our vision for AI that is inclusive, scalable, and enterprise-ready. The ThinkSystem SR650a V4, SR680a V3, and SR780a V3 exemplify Lenovo’s commitment to delivering innovative industry leading infrastructure for every stage of the AI journey. Whether you're developing the next breakthrough in generative AI or deploying edge inferencing at scale, Lenovo offers future-focused solutions you can trust.

For more information

For more information, see the following resources:

Author

Aaron Gilbert is the Worldwide AI Solutions Marketing at Lenovo . He specializes in Lenovo Hybrid AI Platform with NVIDIA.

Related product families

Product families related to this document are the following:

Trademarks

Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.

The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
Lenovo Hybrid AI Advantage
ThinkSystem®

The following terms are trademarks of other companies:

Intel® and Xeon® are trademarks of Intel Corporation or its subsidiaries.

Other company, product, or service names may be trademarks or service marks of others.