skip to main content

Breaking Barriers in AI Inference: Lenovo ThinkSystem Servers Shine in MLPerf v4.1 - Unleashing the Power of AI-Ready Infrastructure

Article

Home
Top
Published
11 Oct 2024
Form Number
LP2036
PDF size
5 pages, 53 KB

Abstract

The latest results from the Machine Learning Performance Foundation (MLPerf) v4.1 benchmarking initiative demonstrate the impressive capabilities of the Lenovo ThinkSystem SR680a V3 and SR685a V3 servers. These systems, equipped with Intel and AMD processors respectively, and powered by NVIDIA GPUs, achieved outstanding performance in various artificial intelligence (AI) inference scenarios.

Lenovo ThinkSystem and ThinkEdge servers won over two-thirds of the benchmarks where Lenovo submitted results. Additionally, the Lenovo ThinkSystem SR650 V3 rack server with 3x NVIDIA L40S GPUs, the ThinkSystem SR675 V3 server with 8x NVIDIA H100-NVL GPUs, the ThinkEdge SE360 V2 with 2x NVIDIA L4 GPUs, and the ThinkEdge SE455 V3 with 2x NVIDIA L40S GPUs, achieved perfect scores in their respective categories and in five categories acquired world record benchmarks.

This achievement highlights the reliability and performance of Lenovo's ThinkSystem and ThinkEdge servers for AI inference workloads, making them a valuable solution for organizations seeking to accelerate their AI initiatives.

Introduction

In a world where AI is revolutionizing industries and transforming businesses, Lenovo is proud to announce that the ThinkSystem servers have broken barriers and taken the top spot in an impressive 54 out of 79 MLPerf v4.1 benchmarks! This remarkable achievement solidifies our position as a leader in the AI infrastructure market, empowering organizations to unlock the full potential of their AI initiatives.

With this groundbreaking performance, Lenovo ThinkSystem servers have demonstrated their ability to handle complex AI workloads with ease and efficiency, making them an ideal choice for organizations looking to accelerate their AI projects. Whether you're a leading financial institution, a cutting-edge tech firm, or a forward-thinking healthcare provider, Lenovo ThinkSystem servers are designed to help you stay ahead of the curve in the rapidly evolving world of AI.

Lenovo dominates GenAI benchmarks

Our ThinkSystem SR680a V3 and SR685a V3 systems have taken center stage by competing against each other in several generative AI benchmarks, showcasing the incredible power of Lenovo's server configurations. The results are impressive:

  • GPT-99 Champion: The ThinkSystem SR680a V3 (Intel) with 8x NVIDIA H200 Tensor Core SXM GPUs, each with 141GB, took top honors, leveraging its Intel 8568Y 48-core processor and advanced memory architecture to deliver exceptional performance.
  • Llama 2 Leaderboard: We secured victory again, this time with the Lenovo ThinkSystem SR685a V3 (AMD) with 8x NVIDIA H200 SXM GPUs with 141GB, highlighting the versatility of Lenovo's server configurations across different architectures.
  • Stable Diffusion XL Speedster: Lenovo ThinkSystem SR680a V3 (Intel) with 8x NVIDIA H200 SXM GPUs with 141GB came out on top in this highly competitive category, demonstrating its ability to handle complex AI workloads and scale performance for demanding applications.

Outstanding MLPerf results

With finishes at or near the top of the pack across numerous MLPerf Inference tests, these results showcase the capabilities of Lenovo ThinkSystem servers in various AI inference scenarios. Whether you're developing cutting-edge AI models or processing large datasets, Lenovo's server configurations provide the scalability and performance you need to drive innovation forward.

The following table provides a breakdown of our results.

Table 1. MLPerf results
System Total Categories Submitted First Place Finishes Second Place Finishes Third Place Finishes
ThinkSystem SR675 V3 16 16 0 0
ThinkEdge SE455 V3 12 12 0 0
ThinkEdge SE360 V2 10 10 0 0
ThinkSystem SR650 V3 9 9 0 0
ThinkSystem SR680a V3 (Intel) 16 6 7 2
ThinkSystem SR685a V3 (AMD) 16 1 1 6

Our benchmarks resulted in world records against all tested systems showcasing Lenovo’s consistent improvement to achieve best-in-class results for our customers. These key results, in particular, show how powerful our systems are in those specific categories:

  • ThinkSystem SR685a V3(8x H200-SXM-141GB, NVIDIA TensorRT) - 1st place on retinanet offline
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) - 1st place on bert-99.9 server
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 1st place on gptj-99 offline
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 1st place on gptj-99.9 offline
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 1st place on stable-diffusion-xl server
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 2nd place on stable-diffusion-xl offline
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) - 2nd place on bert-99.9 server
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 3rd place on gptj-99 server
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) - 3rd place on 3d-unet-99.9 offline
  • ThinkSystem SR680a V3 (8x H200-SXM-141GB, NVIDIA TensorRT) – 3rd place on gptj-99.9 server

Conclusion

The insights from the latest MLPerf benchmarks are critical for stakeholders in the generative AI and machine learning ecosystem, from system architects to application developers. They provide a quantitative foundation for hardware selection and optimization, crucial for deploying scalable and efficient AI/ML systems. Future developments in hardware and software are anticipated to further influence these benchmarks, continuing the cycle of innovation and evaluation in the field of machine learning.

Professionals in the field are encouraged to consider these results in their future hardware procurement and system design strategies. For further discussion or consultation on leveraging these insights in specific use cases, engage with our expert team at aidiscover@lenovo.com.

For more information

For more information, see the following resources:

Author

David Ellison is the Chief Data Scientist for Lenovo ISG. Through Lenovo’s US and European AI Discover Centers, he leads a team that uses cutting-edge AI techniques to deliver solutions for external customers while internally supporting the overall AI strategy for the World Wide Infrastructure Solutions Group. Before joining Lenovo, he ran an international scientific analysis and equipment company and worked as a Data Scientist for the US Postal Service. Previous to that, he received a PhD in Biomedical Engineering from Johns Hopkins University. He has numerous publications in top tier journals including two in the Proceedings of the National Academy of the Sciences.

Related product families

Product families related to this document are the following:

Trademarks

Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.

The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
ThinkEdge®
ThinkSystem®

The following terms are trademarks of other companies:

AMD is a trademark of Advanced Micro Devices, Inc.

Intel® is a trademark of Intel Corporation or its subsidiaries.

Other company, product, or service names may be trademarks or service marks of others.