logo
UPDATE – New MLPerf Storage v2.0 Benchmark Results Demonstrate the Critical Role of Storage Performance in AI Training Systems

UPDATE – New MLPerf Storage v2.0 Benchmark Results Demonstrate the Critical Role of Storage Performance in AI Training Systems

Business Upturn10 hours ago
SAN FRANCISCO, Aug. 04, 2025 (GLOBE NEWSWIRE) — Today, MLCommons® announced results for its industry-standard MLPerf® Storage v2.0 benchmark suite, which is designed to measure the performance of storage systems for machine learning (ML) workloads in an architecture-neutral, representative, and reproducible manner. This round of the benchmark saw dramatically increased participation, more geographic representation from submitting organizations, and greater diversity of the systems submitted for testing.
The benchmark results show that storage systems performance continues to improve rapidly, with tested systems serving roughly twice the number of accelerators than in the v1.0 benchmark round.
Additionally, the v2.0 benchmark adds new tests that replicate real-world checkpointing for AI training systems. The benchmark results provide essential information for stakeholders who need to configure the frequency of checkpoints to optimize for high performance – particularly at scale.
Version 2.0 adds checkpointing tasks, delivers essential insights
As AI training systems have continued to scale up to billions and even trillions of parameters, and the largest clusters of processors have reached one hundred thousand accelerators or more, system failures have become a prominent technical challenge. Because data centers tend to run accelerators at near-maximum utilization for their entire lifecycle, both the accelerators themselves and the supporting hardware (power supplies, memory, cooling systems, etc.) are heavily burdened, minimizing their expected lifetime. This is a chronic issue, especially in large clusters: if the mean time to failure for an accelerator is 50,000 hours, then a 100,000-accelerator cluster running for extended periods at full utilization will likely experience a failure every half-hour. A cluster with one million accelerators would expect to see a failure every three minutes. Worse, because AI training usually involves massively parallel computation where all the accelerators are moving in lockstep on the same iteration of training, a failure of one processor can grind an entire cluster to a halt.
It is now broadly accepted that saving checkpoints of intermediate training results at regular intervals is essential to keep AI training systems running at high performance. The AI training community has developed mathematical models that can optimize cluster performance and utilization by trading off the overhead of regular checkpoints against the expected frequency and cost of failure recovery (rolling back the computation, restoring the most recent checkpoint, restarting the training from that point, and duplicating the lost work). Those models, however, require accurate data on the scale and performance of the storage systems that are used to implement the checkpointing system.
The MLPerf Storage v2.0 checkpoint benchmark tests provide precisely that data, and the results from this round suggest that stakeholders procuring AI training systems need to carefully consider the performance of the storage systems they buy, to ensure that they can store and retrieve a cluster's checkpoints without slowing the system down to an unacceptable level. For a deeper understanding of the issues around storage systems and checkpointing, as well as of the design of the checkpointing benchmarks, we encourage you to read this post from Wes Vaske, a member of the MLPerf Storage working group.
'At the scale of computation being implemented for training large AI models, regular component failures are simply a fact of life,' said Curtis Anderson, MLPerf Storage working group co-chair. 'Checkpointing is now a standard practice in these systems to mitigate failures, and we are proud to be providing critical benchmark data on storage systems to allow stakeholders to optimize their training performance. This initial round of checkpoint benchmark results shows us that current storage systems offer a wide range of performance specifications, and not all systems are well-matched to every checkpointing scenario. It also highlights the critical role of software frameworks such as PyTorch and TensorFlow in coordinating training, checkpointing, and failure recovery, as well as some opportunities for enhancing those frameworks to further improve overall system performance.'
Workload benchmarks show rapid innovation in support of larger-scale training systems
Continuing from the v1.0 benchmark suite, the v2.0 suite measures storage performance in a diverse set of ML training scenarios. It emulates the storage demands across several scenarios and system configurations covering a range of accelerators, models, and workloads. By simulating the accelerators' 'think time' the benchmark can generate accurate storage patterns without the need to run the actual training, making it more accessible to all. The benchmark focuses the test on a given storage system's ability to keep pace, as it requires the simulated accelerators to maintain a required level of utilization.
The v2.0 results show that submitted storage systems have substantially increased the number of accelerators they can simultaneously support, roughly twice the number compared to the systems in the v1.0 benchmark.
'Everything is scaling up: models, parameters, training datasets, clusters, and accelerators. It's no surprise to see that storage system providers are innovating to support ever larger scale systems,' said Oana Balmau, MLPerf Storage working group co-chair.
The v2.0 submissions also included a much more diverse set of technical approaches to delivering high-performance storage for AI training, including: 6 local storage solutions;
2 solutions using in-storage accelerators;
13 software-defined solutions;
12 block systems;
16 on-prem shared storage solutions;
2 object stores.
'Necessity continues to be the mother of invention: faced with the need to deliver storage solutions that are both high-performance and at unprecedented scale, the technical community has stepped up once again and is innovating at a furious pace,' said Balmau.
MLPerf Storage v2.0: skyrocketing participation and diversity of submitters
The MLPerf Storage benchmark was created through a collaborative engineering process by 35 leading storage solution providers and academic research groups across 3 years. The open-source and peer-reviewed benchmark suite offers a level playing field for competition that drives innovation, performance, and energy efficiency for the entire industry. It also provides critical technical information for customers who are procuring and tuning AI training systems.
The v2.0 benchmark results, from a broad set of technology providers, reflect the industry's recognition of the importance of high-performance storage solutions. MLPerf Storage v2.0 includes >200 performance results from 26 submitting organizations: Alluxio, Argonne National Lab, DDN, ExponTech, FarmGPU, H3C, Hammerspace, HPE, JNIST/Huawei, Juicedata, Kingston, KIOXIA, Lightbits Labs, MangoBoost, Micron, Nutanix, Oracle, Quanta Computer, Samsung, Sandisk, Simplyblock, TTA, UBIX, IBM, WDC, and YanRong. The submitters represent seven different countries, demonstrating the value of the MLPerf Storage benchmark to the global community of stakeholders.
'The MLPerf Storage benchmark has set new records for an MLPerf benchmark, both for the number of organizations participating and the total number of submissions,' said David Kanter, Head of MLPerf at MLCommons. The AI community clearly sees the importance of our work in publishing accurate, reliable, unbiased performance data on storage systems, and it has stepped up globally to be a part of it. I would especially like to welcome first-time submitters Alluxio, ExponTech, FarmGPU, H3C, Kingston, KIOXIA, Oracle, Quanta Cloud Technology, Samsung, Sandisk, TTA, UBIX, IBM, and WDC.'
'This level of participation is a game-changer for benchmarking: it enables us to openly publish more accurate and more representative data on real-world systems,' Kanter continued. 'That, in turn, gives the stakeholders on the front lines the information and tools they need to succeed at their jobs. The checkpoint benchmark results are an excellent case in point: now that we can measure checkpoint performance, we can think about optimizing it.'
We invite stakeholders to join the MLPerf Storage working group and help us continue to evolve the benchmark suite.
View the Results
To view the results for MLPerf Storage v2.0, please visit the Storage benchmark results .
About MLCommons
MLCommons is the world's leader in AI benchmarking. An open engineering consortium supported by over 125 members and affiliates, MLCommons has a proven record of bringing together academia, industry, and civil society to measure and improve AI. The foundation for MLCommons began with the MLPerf benchmarks in 2018, which rapidly scaled as a set of industry metrics to measure machine learning performance and promote transparency of machine learning techniques. Since then, MLCommons has continued using collective engineering to build the benchmarks and metrics required for better AI – ultimately helping to evaluate and improve AI technologies' accuracy, safety, speed, and efficiency.
For additional information on MLCommons and details on becoming a member, please visit MLCommons.org or email [email protected] .
Disclaimer: The above press release comes to you under an arrangement with GlobeNewswire. Business Upturn takes no editorial responsibility for the same.
Ahmedabad Plane Crash
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Unpacking Fashion's New AI Marketing Toolkit
Unpacking Fashion's New AI Marketing Toolkit

Business of Fashion

time20 minutes ago

  • Business of Fashion

Unpacking Fashion's New AI Marketing Toolkit

Ever since Mango started to use Smartly, an AI-driven performance marketing tool, for social media ads, its revenue generated by these ads has quadrupled. Using the platform, Mango's design team is able create brand templates that can be easily toggled between different versions of a photo of a dress, for example. Whereas before, creating different images required taking photos of different outfits in multiple settings. The womenswear brand is far from the only fashion player to have figured out how to use artificial intelligence to optimise and boost the efficiency of advertising. Companies including sneaker retailer Foot Locker and German e-commerce giant Zalando are using AI for everything from product image backgrounds to quick-turnaround campaigns that align with social media microtrends. Consumers go through content — and trends — so quickly today that advertisers struggle to keep up with analog creative production alone. AI's rapid generation capabilities help marketers produce more varied ads so viewers are less likely to tire of them as they scroll social media's endless feeds. But figuring out the right formula to use AI is no small feat. New tools seem to pop up weekly and while some platforms excel at automatically generating images, they may lack the editing tools marketers need to ensure visuals stay on-brand. Brands must consider their unique needs — which can range from visualising creative ideas for a campaign and enhancing their product imagery with dynamic backgrounds to automatically targeting different audiences with different versions of the same ad — in order to determine the ways in which AI could amplify their creative production. The most accessible AI tools are ChatGPT and Adobe's Firefly, which many agencies already use for easy image generation. Others favour more advanced platforms like Midjourney and Leonardo AI for greater creative control and experimentation. Meanwhile, AI agency Maison Meta and Zalando aggregate multiple AI technologies into their own platforms, curating the best tools for specific tasks like background generation and editing under one interface — Maison Meta also hosts workshops for clients like Mango to teach them how to use its toolkit. More advanced tools are able to connect the dots between image generation and ad optimisation on social media. Meta, for instance, has its own suite of AI tools called Advantage+, which includes creative enhancements that allow brands to customise their ads by animating imagery, adding music or enhancing copy. Platforms like Smartly and Pixis are also dedicated to optimising the creative content in ads, allowing marketers to create visual templates that can be automatically updated to cater to different customers across platforms like Meta, TikTok, Amazon and more. Footwear retailer JustFab, for example, was able to create a template for its broad product catalog by removing the backgrounds across all images and generating new seasonal scenes across using Smartly. 'It's the consumption happening now through social media that requires a volume that the marketing industry wasn't prepared for,' said PJ Pereira, creative chairman of advertising agency Pereira O'Dell. Lead With Human Creativity Regardless of the AI tools, human creativity and artistic direction must come first. 'This is not the end of the creative director or the creative teams,' said Jason Widup, senior vice president of marketing at AI advertising company Pixis. 'What we're seeing is the creative teams want to create new, interesting, fun concepts. They want to be pushing the envelope. And AI still doesn't have human taste.' What this looks like in practice is using traditional brainpower to come up with core campaign concepts, and then using AI to see how far the idea can be pushed — testing out how an image would look on Mars, for instance, or with a giraffe alongside a model, said Pereira. 'You can think and see in real time,' he said. While in the past, campaign generation could be costly, AI enables marketers to test multiple ideas at one time and see what works best. 'The biggest leap currently as a creative industry is the fact that the risk of trying something is infinitely smaller,' said Pereira. Turn One Asset into Countless Variations Proponents of the technology still don't trust AI to fully take over ad production, said Pixis's Widup. Instead, many feed product assets and brand guidelines into tools like Smartly and Pixis, and then use the platforms to iterate on existing ideas and imagery to change backgrounds, colours or placement to create different versions of an ad that can be personalised and targeted across audiences. For Zalando, using its own AI tools to showcase products in a relevant context — like a hiking boot on a mountain, or even to turn static imagery into video — makes all the difference in driving customer engagement. This, in turn, helps them better understand what their customers want to see, which leads to fewer returns and triple the conversions. 'We do not have endless hands, and we do not have endless time and endless budget,' said Matthias Haase, vice president of content solutions at Zalando. Localise in Real Time Beyond creative content, AI can be leveraged to reach the right users at the right time. Whereas in the analog days of digital acquisition, consumers were repeatedly served the same ads, brands can now automate when viewers are served a piece of content — whether for a complementary product to one they already own or a specific handbag colour they've been searching for. Most of Zalando's AI-generated content is market-specific to cater to local moments like Oktoberfest in Germany or regional running events, for example. These quick-turnaround, targeted campaigns typically take four days to produce, compared to six to eight weeks for traditional ad campaigns, and made up 40 percent of Zalando's campaigns in 2025. Its product description pages will eventually be customised to each geographical market using AI, showing different outfit combinations for French versus Polish shoppers, for example, and swapping out models that are recognised in different regions using digital twins. Build Content that Responds to Context For platforms that connect AI image generation with targeting such as Smartly, brands can reach customers even more dynamically by making each element of an image changeable depending on the customer who is interfacing with it. 'Every element can effectively change based on the time of day, [and the] specific audience that is being targeted,' said Oliver Marlow-Thomas, chief innovation officer at Smartly. 'We can dynamically map specific audiences to show specific variants … The dress the model is wearing, the shoe that's in the image, the colour of the background, or the colour of the hat, all of that is mapped and changeable.' Pixis, on its end, draws on data sources from Shopify to Google to macroeconomic data and weather channels to inform its decisions around which ad imagery is shown to who and when, said Widup. The platform has tested these dynamic ads for fashion brands on Meta, tailoring visuals and messaging to current weather patterns to make them feel relevant — showing rain jackets during storms or breezy outfits during heatwaves, for example — and boosting engagement and conversion rates. Optimise Targeting with Platform Data Brands that advertise on Meta can use its native AI tools to enhance existing content and optimise targeting. After Meta's targeting capabilities were reduced with Apple iOS changes in 2021, the platform began to build out Advantage+ in order to enhance advertising effectiveness on the platform by testing different ad types across users — and the more versions of creative content it has to test on customers, the better. 'Platforms like Meta are now saying, 'We've reduced your targeting capabilities, but just give us a lot of creative ads to use, and we'll refine it,'' said Widup. Meta's own targeting only goes so far, however, because it solely relies on data from its own ecosystem of platforms and first-party customer data provided by advertisers. Tools like Pixis and Smartly, which are integrated across social, search engines and shopping platforms including TikTok, YouTube, Google and Amazon, can use data from various sources to expand their targeting capabilities. Regardless of the optimisation tool used, the creative idea at the root of any campaign — and the creativity that underlies the different AI-made iterations — ultimately remains the most important piece of the puzzle. 'The original thought is the thing that will capture the human,' said Smartly's Marlow-Thomas. 'The artificial thought is the thing that will scale it and create lots of different iterations.'

Arteris up 57% after announcing licensing with AMD
Arteris up 57% after announcing licensing with AMD

Business Insider

time26 minutes ago

  • Business Insider

Arteris up 57% after announcing licensing with AMD

The company stated: 'In a market reshaped by the compute demands of AI, Arteris (AIP) announced AMD (AMD), a global leader in high-performance and adaptive computing, has licensed FlexGen network-on-chip interconnect IP for its next generation of AI chiplet design. FlexGen, Arteris' smart NoC IP technology, will provide high-performance data transport in AMD chiplets powering AI across the company's broad portfolio which spans from data centers to edge and end devices. The strategic combination and interoperability of Arteris' FlexGen NoC IP with the AMD Infinity Fabric interconnect underscores the increasing complexity of modern SoCs and chiplet-based architectures, which now require multiple highly specialized interconnects or NoCs to efficiently meet the demands of modern electronic systems.' Shares of Arteris are up 57% at $14.84 afterhours. Elevate Your Investing Strategy: Take advantage of TipRanks Premium at 50% off! Unlock powerful investing tools, advanced data, and expert analyst insights to help you invest with confidence.

Advanced Micro Devices (AMD) Is About to Report Q2 Earnings Tomorrow. Here Is What to Expect
Advanced Micro Devices (AMD) Is About to Report Q2 Earnings Tomorrow. Here Is What to Expect

Business Insider

timean hour ago

  • Business Insider

Advanced Micro Devices (AMD) Is About to Report Q2 Earnings Tomorrow. Here Is What to Expect

Advanced Micro Devices (AMD) is set to report its Q2 2025 earnings on Tuesday, August 5, after the market closes. AMD stock has gained 42% in 2025 so far, driven by investors' growing confidence in the company's AI chip strategy. The stock rally picked up pace after the company unveiled its upcoming MI350 chip at its AI event in June. Positive sentiment was further boosted by reports that AMD intends to raise prices on its MI350 chip, indicating strong demand. Wall Street analysts expect the company to post revenues of $7.41 billion, up 27% from the year-ago quarter, according to data from the TipRanks Forecast page. Elevate Your Investing Strategy: Take advantage of TipRanks Premium at 50% off! Unlock powerful investing tools, advanced data, and expert analyst insights to help you invest with confidence. However, earnings are expected to decrease by about 30% from the year-ago quarter to $0.48 per share. Notably, AMD has an encouraging earnings surprise history. The company missed earnings estimates just twice out of the previous nine quarters. Analysts' Views Remain Divided Ahead of Q2 Earnings Wall Street is divided on AMD's short-term outlook. While some analysts expect strong growth driven by AI demand, others remain cautious due to lofty expectations and execution risks. For instance, Erste Group's Stephan Lingnau upgraded AMD to Buy, citing strong demand for high-performance chips and rising margins. He expects profit growth to pick up in 2025 and sees 'good growth prospects' for the stock. Meanwhile, DZ Bank analyst Ingo Wermann downgraded the stock from Buy to Sell but increased the price target to $150.00 from $118 per share. Also, Bernstein Top analyst Stacy Rasgon prefers to remain on the sidelines. While he raised his price target on AMD to $140, he kept a neutral rating, pointing to high valuations and near-term risks like ' tariff pull-forwards.' Rasgon sees near-term upside from China AI sales and the MI350 launch but warned that investor expectations may already be 'elevated.' Options Traders Anticipate a Large Move Using TipRanks' Options tool, we can see what options traders are expecting from the stock immediately after its earnings report. The expected earnings move is determined by calculating the at-the-money straddle of the options closest to expiration after the earnings announcement. If this sounds complicated, don't worry, the Options tool does this for you. Indeed, it currently says that options traders are expecting an 8.97% move in either direction. Is AMD Stock a Good Buy? Overall, Wall Street has a Moderate Buy consensus rating on Advanced Micro Devices stock based on 24 Buys, 10 Holds, and 1 Sell rating. The average AMD stock price target of $156.47 indicates a possible downside of 8.87% from current levels.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store