
Nvidia Benchmark Recipes Bring Deep Insights In AI Performance
Nvidia DGX Cloud Data Center
As AI workloads and accelerated applications grow in sophistication and complexity, businesses and developers need better tools to assess their infrastructure's ability to handle the demands of both training and inference more efficiently. To that end, Nvidia has been working on a set of performance testing tools, called DGX Cloud Benchmark Recipes, that are designed to help organizations evaluate how their hardware and cloud infrastructure perform when running the most advanced AI models available today. Our team at HotTech had a chance to kick the tires on a few of these recipes recently, and found the data they can capture to be extremely insightful.
Nvidia's toolkit also offers a database and calculator of performance results for GPU-compute workloads on various configurations, including the number of Nvidia H100 GPUs and cloud service providers, while the recipes allow businesses to run realistic performance evaluations on their own infrastructure. The results can help guide decisions on whether to invest in more powerful hardware, cloud provider service levels, or tweak configurations to better meet machine learning demands. These tools also take a holistic approach that incorporates network technologies for optimal throughput.
Nvidia DGX Cloud Benchmarking Recipes are a set of pre-configured containers and scripts that users can download and run on their own infrastructure. These containers are optimized for testing the performance of various AI models under different configurations, making them very valuable for companies looking to benchmark systems, whether on prem or in the cloud, before committing to larger-scale AI workloads or infrastructure deployments.
Nvidia DGX Cloud Time To Train Performance Report
In addition to offering static performance data, time to train and efficiency calculated from its database, Nvidia has recipes readily available for download that let businesses run real-world tests on their own hardware or cloud infrastructure, helping them understand the performance impact of different configurations. The recipes include benchmarks for training models like Meta's Llama 3.1 and Nvidia's own Llama 3.1 branch, called Nemotron, across several cloud providers (AWS, Google Cloud, and Azure), with options for adjusting factors like model size, GPU usage, and precision. The database is broad enough to cover popular AI models, but it is primarily designed for testing large-scale pre-training tasks, rather than inference on smaller models. The benchmarking process also allows for flexibility. Users can tailor the tests to their specific infrastructure by adjusting parameters such as the number of GPUs and the size of the model being trained.
The default hardware configuration in Nvidia's database of results uses the company's high-end H100 80GB GPUs, but it is designed to be adaptable. Although currently, it does not include consumer or prosumer-grade GPUs (e.g., RTX A4000 or RTX 50) or the company's latest Blackwell GPU family, these options could be added in the future.
Running the DGX Cloud Benchmarking Recipes is straightforward, assuming a few prerequisites are met. The process is well-documented, with clear instructions on setting up, running the benchmarks, and interpreting the results. Once a benchmark is completed, users can review the performance data, which includes key metrics like training time, GPU usage, and throughput. This allows businesses to make data-driven decisions about which configurations deliver the best performance and efficiency for their AI workloads. This could also go a long way in helping companies maintain green initiatives in terms of meeting power consumption and efficiency budgets.
While the DGX Cloud Benchmarking Recipes offer valuable insights, there are a few areas where Nvidia's tools could be expanded. First, benchmarking recipes are currently focused primarily on pre-training large models, not on real-time inference performance. Inference tasks, such as token generation or running smaller AI models, are equally important in many business applications. Expanding the toolset to include more detailed inference benchmarks would provide a fuller picture of how different hardware configurations handle these real-time demands. Additionally, by expanding the recipe selection to include lower-end or even higher-end GPUs (like Blackwell or even competitive offerings), Nvidia could cater to a broader audience, particularly businesses that don't require the massive compute power of a Hopper H100 80GB cluster for every workload.
Regardless, Nvidia's new DGX Cloud Benchmarking Recipes look like a very helpful resource for evaluating the performance of AI compute infrastructure, before making major investment decisions. They offer a practical way to understand how different configurations—whether cloud-based or on-premises—handle complex AI workloads. This is especially valuable for organizations exploring which cloud provider best meets their needs, or if the company is looking for new ways to optimize existing infrastructure.
As AI's role in business and our everyday lives continues to grow, tools like this will become essential for guiding infrastructure decisions, balancing performance versus cost and power consumption, and optimizing AI applications to meet real-world demands. As Nvidia expands these recipes to include more inference-focused benchmarks and potentially expands its reference data with a wider range of GPU options, these tools could become even more indispensable to businesses and developers of all sizes.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Bloomberg
3 hours ago
- Bloomberg
Meta in Talks for Scale AI Investment That Could Top $10 Billion
Meta Platforms Inc. is in talks to make a multibillion-dollar investment into artificial intelligence startup Scale AI, according to people familiar with the matter. The financing could exceed $10 billion in value, some of the people said, making it one of the largest private company funding events of all time.
Yahoo
8 hours ago
- Yahoo
Why this key chip technology is crucial to the AI race between the US and China
In the largest single foreign investment in US history, Taiwan Semiconductor Manufacturing Company has unveiled a $100 billion investment, drawing global attention and prompting concern in Taiwan. TSMC, which produces more than 90% of the world's advanced semiconductor chips that power everything from smartphones and artificial intelligence (AI) applications to weapons, will build two new advanced packaging facilities in Arizona, among others. Here's everything you need to know about advanced packaging technology, which has seen exponential demand growth along with the global AI frenzy, and what that means for the struggle between the US and China for AI dominance. While the two countries have announced a temporary truce that rolled back disruptive three-digit tariffs for 90 days, the relationship remains tense because of ongoing feuding over chip restrictions imposed by the US and other issues. Last month at Computex, an annual trade show in Taipei that has been thrust under the limelight because of the AI boom, the CEO of chipmaker Nvidia, Jensen Huang, told reporters that 'the importance of advanced packaging for AI is very high,' proclaiming that 'no one has pushed advanced packaging harder than me.' Packaging generally refers to one of the manufacturing processes of semiconductor chips, which means sealing a chip inside a protective casing and mounting it to the motherboard that goes into an electronic device. Advanced packaging, specifically, refers to techniques that allow more chips — such as graphic processing units (GPU), central processing units (CPU) or high bandwidth memory (HBM) — to be placed closer together, leading to better overall performance, faster transmission of data and lower energy consumption. Think of these chips as different departments within a company. The closer these departments are to each other, the easier it is, and less time it takes, for people to travel between them and exchange ideas, and the more efficient the operation becomes. 'You're trying to put the chips as close together as possible, and you're also putting in different solutions to make the connection between the chips very easy,' Dan Nystedt, vice president of Asia-based private investment firm TrioOrient, told CNN. In a way, advanced packaging keeps afloat Moore's Law, the idea that the number of transistors on microchips would double every two years, as breakthroughs in the chip fabrication process become increasingly costly and more difficult. While there are many types of advanced packaging technologies, CoWoS, short for Chips-on-Wafer-on-Substrate and invented by TSMC, is arguably the best known that was thrown under the limelight since the debut of OpenAI's ChatGPT, which sparked the AI frenzy. It has even become a household name in Taiwan, prompting Lisa Su, CEO of Advanced Micro Devices (AMD), to say that the island is the 'only place that you can say CoWoS and everybody would understand.' Advanced packaging has become a big deal in the tech world because it ensures AI applications, which require a lot of complex computing, run without delays or glitches. CoWoS is indispensable to producing AI processors, such as the GPUs produced by Nvidia and AMD that are used in AI servers or data centers. 'You could call it the Nvidia packaging process if you want to. Almost anyone making AI chips is using the CoWoS process,' said Nystedt. That is why demand for CoWoS technology has skyrocketed. As a result, TSMC is scrambling to ramp up production capacity. In a visit to Taiwan in January, Huang told reporters that the amount of advanced packaging capacity currently available was 'probably four times' what it was less than two years ago. 'The technology of packaging is very important to the future of computing,' he said. 'We now need to have very complicated advanced packaging to put many chips together into one giant chip.' If advanced fabrication is one piece of the puzzle in terms of chip manufacturing, advanced packaging is another. Analysts say having both pieces of that jigsaw in Arizona means the US will have a 'one-stop shop' for chip production and a strengthened position for its AI arsenal, benefitting Apple, Nvidia, AMD, Qualcomm and Broadcom, some of TSMC's top clients. 'It ensures that the US has a complete supply chain from advanced manufacturing to advanced packaging, which would benefit the US' competitiveness in AI chips,' Eric Chen, an analyst with market research firm Digitimes Research, told CNN. Because advanced packaging technologies key to AI are currently only produced in Taiwan, having it in Arizona also reduces potential supply chain risks. 'Instead of having all eggs in one basket, CoWoS would be in Taiwan and also the US, and that makes you feel more safe and secure,' said Nystedt. While CoWoS got its moment recently, the technology has actually existed for at least 15 years. It was the brainchild of a team of engineers led by Chiang Shang-yi, who served two stints at TSMC and retired from the company as its co-chief operating officer. Chiang first proposed developing the technology in 2009 in an attempt to fit more transistors into chips and solve bottlenecks in performance. But when it was developed, few companies took up the technology because of the high cost associated with it. 'I only had one customer … I really became a joke (in the company), and there was so much pressure on me,' he recalled in a 2022 oral history project recorded for the Computer History Museum in Mountain View, California. But the AI boom turned CoWoS around, making it one of the most popular technologies. 'The result was beyond our original expectation,' Chiang said. In the global semiconductor supply chain, companies that specialize in packaging and testing services are referred to as outsourced semiconductor assembly and test (OSAT) firms. In addition to TSMC, South Korea's Samsung and America's Intel, as well as OSAT firms including China's JCET Group, America's Amkor and Taiwan's ASE Group and SPIL are all key players in advanced packaging technologies.
Yahoo
9 hours ago
- Yahoo
Why this key chip technology is crucial to the AI race between the US and China
In the largest single foreign investment in US history, Taiwan Semiconductor Manufacturing Company has unveiled a $100 billion investment, drawing global attention and prompting concern in Taiwan. TSMC, which produces more than 90% of the world's advanced semiconductor chips that power everything from smartphones and artificial intelligence (AI) applications to weapons, will build two new advanced packaging facilities in Arizona, among others. Here's everything you need to know about advanced packaging technology, which has seen exponential demand growth along with the global AI frenzy, and what that means for the struggle between the US and China for AI dominance. While the two countries have announced a temporary truce that rolled back disruptive three-digit tariffs for 90 days, the relationship remains tense because of ongoing feuding over chip restrictions imposed by the US and other issues. Last month at Computex, an annual trade show in Taipei that has been thrust under the limelight because of the AI boom, the CEO of chipmaker Nvidia, Jensen Huang, told reporters that 'the importance of advanced packaging for AI is very high,' proclaiming that 'no one has pushed advanced packaging harder than me.' Packaging generally refers to one of the manufacturing processes of semiconductor chips, which means sealing a chip inside a protective casing and mounting it to the motherboard that goes into an electronic device. Advanced packaging, specifically, refers to techniques that allow more chips — such as graphic processing units (GPU), central processing units (CPU) or high bandwidth memory (HBM) — to be placed closer together, leading to better overall performance, faster transmission of data and lower energy consumption. Think of these chips as different departments within a company. The closer these departments are to each other, the easier it is, and less time it takes, for people to travel between them and exchange ideas, and the more efficient the operation becomes. 'You're trying to put the chips as close together as possible, and you're also putting in different solutions to make the connection between the chips very easy,' Dan Nystedt, vice president of Asia-based private investment firm TrioOrient, told CNN. In a way, advanced packaging keeps afloat Moore's Law, the idea that the number of transistors on microchips would double every two years, as breakthroughs in the chip fabrication process become increasingly costly and more difficult. While there are many types of advanced packaging technologies, CoWoS, short for Chips-on-Wafer-on-Substrate and invented by TSMC, is arguably the best known that was thrown under the limelight since the debut of OpenAI's ChatGPT, which sparked the AI frenzy. It has even become a household name in Taiwan, prompting Lisa Su, CEO of Advanced Micro Devices (AMD), to say that the island is the 'only place that you can say CoWoS and everybody would understand.' Advanced packaging has become a big deal in the tech world because it ensures AI applications, which require a lot of complex computing, run without delays or glitches. CoWoS is indispensable to producing AI processors, such as the GPUs produced by Nvidia and AMD that are used in AI servers or data centers. 'You could call it the Nvidia packaging process if you want to. Almost anyone making AI chips is using the CoWoS process,' said Nystedt. That is why demand for CoWoS technology has skyrocketed. As a result, TSMC is scrambling to ramp up production capacity. In a visit to Taiwan in January, Huang told reporters that the amount of advanced packaging capacity currently available was 'probably four times' what it was less than two years ago. 'The technology of packaging is very important to the future of computing,' he said. 'We now need to have very complicated advanced packaging to put many chips together into one giant chip.' If advanced fabrication is one piece of the puzzle in terms of chip manufacturing, advanced packaging is another. Analysts say having both pieces of that jigsaw in Arizona means the US will have a 'one-stop shop' for chip production and a strengthened position for its AI arsenal, benefitting Apple, Nvidia, AMD, Qualcomm and Broadcom, some of TSMC's top clients. 'It ensures that the US has a complete supply chain from advanced manufacturing to advanced packaging, which would benefit the US' competitiveness in AI chips,' Eric Chen, an analyst with market research firm Digitimes Research, told CNN. Because advanced packaging technologies key to AI are currently only produced in Taiwan, having it in Arizona also reduces potential supply chain risks. 'Instead of having all eggs in one basket, CoWoS would be in Taiwan and also the US, and that makes you feel more safe and secure,' said Nystedt. While CoWoS got its moment recently, the technology has actually existed for at least 15 years. It was the brainchild of a team of engineers led by Chiang Shang-yi, who served two stints at TSMC and retired from the company as its co-chief operating officer. Chiang first proposed developing the technology in 2009 in an attempt to fit more transistors into chips and solve bottlenecks in performance. But when it was developed, few companies took up the technology because of the high cost associated with it. 'I only had one customer … I really became a joke (in the company), and there was so much pressure on me,' he recalled in a 2022 oral history project recorded for the Computer History Museum in Mountain View, California. But the AI boom turned CoWoS around, making it one of the most popular technologies. 'The result was beyond our original expectation,' Chiang said. In the global semiconductor supply chain, companies that specialize in packaging and testing services are referred to as outsourced semiconductor assembly and test (OSAT) firms. In addition to TSMC, South Korea's Samsung and America's Intel, as well as OSAT firms including China's JCET Group, America's Amkor and Taiwan's ASE Group and SPIL are all key players in advanced packaging technologies. Error while retrieving data Sign in to access your portfolio Error while retrieving data Error while retrieving data Error while retrieving data Error while retrieving data