logo
Dell launches new AI innovations for enterprise & research

Dell launches new AI innovations for enterprise & research

Techday NZ20-05-2025
Dell has announced a range of advancements in enterprise AI infrastructure, solutions and services to support organisations seeking to adopt and scale artificial intelligence.
The company reported that 75% of organisations view AI as central to their strategy, but high costs and data security concerns remain significant obstacles. Dell aims to address these challenges by simplifying deployment, reducing expenses and enabling secure, scalable AI adoption through its expanded Dell AI Factory, enhanced infrastructure and an expanding partner ecosystem.
The Dell AI Factory, launched a year ago, has received more than 200 updates and now supports AI workloads at any scale with new infrastructure, software improvements and collaborations with partners such as NVIDIA, Meta and Google. The company states its approach to on-premises AI inference can be up to 62% more cost effective for large language models compared to public cloud options.
Among the notable product introductions is the Dell Pro Max Plus laptop, equipped with the Qualcomm AI 100 PC Inference Card, which the company states is the world's first mobile workstation to include an enterprise-grade discrete NPU. This platform is intended to provide rapid, secure on-device inferencing for large AI models, facilitating edge deployments outside traditional data centres. The Qualcomm AI 100 PC Inference Card offers 32 AI-cores and 64GB of memory to support engineers and scientists working with sizeable data models.
Addressing the energy demands of AI workloads, Dell introduced the PowerCool Enclosed Rear Door Heat Exchanger (eRDHx), designed to capture 100% of IT-generated heat and reduce cooling costs by up to 60% compared to current solutions. This innovation supports water temperatures between 32°C and 36°C and enables increased data centre density, allowing organisations to deploy up to 16% more racks of dense compute without raising power consumption, and provides advanced leak detection and unified management features.
For high performance computing and AI, Dell's PowerEdge XE9785 and XE9785L servers will support AMD Instinct MI350 series GPUs, promising up to 35 times greater inferencing performance. Both liquid- and air-cooled versions will be available to further reduce facility cooling costs.
In terms of data management, Dell's AI Data Platform now includes updates designed to improve access to structured and unstructured data, with Project Lightning, a parallel file system, reported to deliver up to two times greater throughput than alternatives. Enhancements to the Data Lakehouse further streamline AI workflows for use cases like recommendation engines and semantic search, and the introduction of Linear Pluggable Optics aims to lower power use and boost networking efficiency.
Dr Paul Calleja, Director of the Cambridge Open Zettascale Lab and Research Computing Services at the University of Cambridge, commented: "We're excited to work with Dell to support our cutting-edge AI initiatives, and we expect Project Lightning to be a critical storage technology for our AI innovations."
Dell has also broadened its partner ecosystem to include on-premises deployments with platforms such as Cohere North, Google Gemini, Glean's Work AI platform and Meta's Llama Stack, as well as joint solutions with Mistral AI. The company is providing enhancements to its AI platform with AMD and Intel technologies, including upgraded networking, software stack improvements, container support and integration with Intel Gaudi 3 AI accelerators.
Updates to the Dell AI Factory with NVIDIA include new PowerEdge servers supporting up to 192 NVIDIA Blackwell Ultra GPUs per standard configuration and up to 256 per Dell IR7000 rack with direct to chip liquid cooling. These advancements aim to simplify data centre integration, speed up rack-scale AI deployment and are reported to deliver up to four times faster large language model training compared to the previous generation.
The PowerEdge XE9712, featuring NVIDIA GB300 NVL72, targets efficiency at rack scale for training and is said to offer up to 50 times more inference output and five times improvement in throughput, with new PowerCool technology supporting power efficiency in high-demand environments. The company intends to support the NVIDIA Vera CPU and Vera Rubin platform in future server offerings.
In networking, Dell has extended its portfolio with new PowerSwitch and InfiniBand switches, deliver up to 800 Gbps of throughput, and are now supported by ProSupport and Deployment Services. Further software platform updates include direct availability of NVIDIA NIM, NeMo microservices and Blueprints, plus Red Hat OpenShift integration on the Dell AI Factory with NVIDIA.
To streamline AI operations, Dell has introduced Managed Services for the AI Factory with NVIDIA, providing 24/7 monitoring, reporting, upgrades and patching for the stack, supported by Dell's technical teams.
Michael Dell, Chairman and Chief Executive Officer, Dell Technologies, said: "We're on a mission to bring AI to millions of customers around the world. Our job is to make AI more accessible. With the Dell AI Factory with NVIDIA, enterprises can manage the entire AI lifecycle across use cases, from training to deployment, at any scale."
Jensen Huang, Founder and Chief Executive Officer, NVIDIA, added: "AI factories are the infrastructure of modern industry, generating intelligence to power work across healthcare, finance and manufacturing. With Dell Technologies, we're offering the broadest line of Blackwell AI systems to serve AI factories in clouds, enterprises and at the edge."
Jeff Clarke, Chief Operating Officer, Dell Technologies, stated: "It has been a non-stop year of innovating for enterprises, and we're not slowing down. We have introduced more than 200 updates to the Dell AI Factory since last year. Our latest AI advancements — from groundbreaking AI PCs to cutting-edge data centre solutions — are designed to help organisations of every size to seamlessly adopt AI, drive faster insights, improve efficiency and accelerate their results."
Christopher M. Sullivan, Director of Research and Academic Computing for the College of Earth, Ocean and Atmospheric Sciences at Oregon State University, said: "We leverage the Dell AI Factory for our oceanic research at Oregon State University to revolutionise and address some of the planet's most critical challenges. Through advanced AI solutions, we're accelerating insights that empower global decision-makers to tackle climate change, safeguard marine ecosystems and drive meaningful progress for humanity."
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

HPE expands ProLiant servers & AI cloud with new NVIDIA GPUs
HPE expands ProLiant servers & AI cloud with new NVIDIA GPUs

Techday NZ

timea day ago

  • Techday NZ

HPE expands ProLiant servers & AI cloud with new NVIDIA GPUs

HPE has announced a series of updates to its NVIDIA AI Computing by HPE portfolio, emphasising expanded server capabilities and deeper integration with NVIDIA AI Enterprise solutions. New server models The HPE ProLiant Compute range will soon include servers featuring the NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, available in a 2U form factor. Two main configurations will be available: the HPE ProLiant DL385 Gen11 server, supporting up to two of the new GPUs in a 2U chassis, and the previously announced HPE ProLiant Compute DL380a Gen12 server, capable of using up to eight GPUs in a 4U form factor. According to HPE, the latter configuration will be shipping in September. These servers are designed for broad enterprise applications such as generative and agentic AI, robotics and industrial AI, visual computing (including autonomous vehicles and quality control monitoring), simulation, 3D modelling, digital twins, and enterprise applications. The HPE ProLiant Compute Gen12 servers also include hardware features aimed at enhancing security and operational efficiency, such as the HPE Integrated Lights Out (iLO) 7 Silicon Root of Trust and a secure enclave for tamper-resistant protection and quantum-resistant firmware signing. HPE estimates that its Compute Ops Management, a cloud-native tool for managing server lifecycles, can decrease IT hours for server management by up to 75 percent and reduce downtime by approximately 4.8 hours per server each year. HPE Private Cloud AI HPE's Private Cloud AI offering, which has been co-developed with NVIDIA, is being updated to support the latest NVIDIA GPU technologies and AI models. The next version will offer compatibility for NVIDIA RTX PRO 6000 GPUs on Gen12 servers and aims to provide seamless scalability across different GPU generations. Features will include air-gapped management for security and support for enterprise multi-tenancy. The new release of HPE Private Cloud AI will integrate recent NVIDIA AI models, including the NVIDIA Nemotron models focused on agentic AI, Cosmos Reason vision language model (VLM) for physical AI and robotics, and the NVIDIA Blueprint for Video Search and Summarization (VSS 2.4), intended to build video analytics AI agents that can process large volumes of video data. Customers will have access to these developments through the HPE AI Essentials platform, enabling the quick deployment of NVIDIA NIM microservices and other AI tools. Through the continued collaboration, HPE Private Cloud AI is designed to deliver an integrated solution that leverages NVIDIA's portfolio in AI accelerated computing, networking, and software. This enables businesses to address increasing demand for AI inferencing and to accelerate the development and deployment of AI systems, maintaining high security and control over enterprise data. Collaboration and customer impact "HPE is committed to empowering enterprises with the tools they need to succeed in the age of AI," said Cheri Williams, senior vice president and general manager for private cloud and flex solutions at HPE. "Our collaboration with NVIDIA continues to push the boundaries of innovation, delivering solutions that unlock the value of generative, agentic and physical AI while addressing the unique demands of enterprise workloads. With the combination of HPE ProLiant servers and expanded capabilities in HPE Private Cloud AI, we're enabling organizations to embrace the future of AI with confidence and agility." Justin Boitano, vice president of enterprise AI at NVIDIA, commented, "Enterprises need flexible, efficient infrastructure to keep pace with the demands of modern AI. With NVIDIA RTX PRO 6000 Blackwell GPUs in HPE's 2U ProLiant servers, enterprises can accelerate virtually every workload on a single, unified, enterprise-ready platform." Availability According to HPE, the HPE ProLiant DL385 Gen11 and HPE ProLiant Compute DL380a Gen12 servers featuring the NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs are now orderable and are set for distribution worldwide beginning 2 September 2025. Support within HPE Private Cloud AI for the latest NVIDIA Nemotron models, Cosmos Reason, and the NVIDIA Blueprint for VSS 2.4 is scheduled for release in the second half of 2025. The next generation of HPE Private Cloud AI with the updated GPU capabilities is also expected to be available during this period.

How optimisation is helping to tackle the data centre efficiency challenge
How optimisation is helping to tackle the data centre efficiency challenge

Techday NZ

time3 days ago

  • Techday NZ

How optimisation is helping to tackle the data centre efficiency challenge

In the era of cloud adoption and AI, the demand for data centre bandwidth has skyrocketed, leading to the exponential sprawl of data centres worldwide. However, new data centres are running up against sustainability, space and budget constraints. Policymakers recognise the benefits of data centres to productivity, economic growth and research, but there is still a tension over their impact on local communities, water and electricity use. The best solution is in optimising the data centre infrastructure we have already to unlock more performance while still being mindful of the limits we have. Our cities, our consumer products and our world is going to become more digital and we need more compute to keep up. Optimising the data centre infrastructure we have already to unlock more performance is the best way data centres can turn constraints into an opportunity for a competitive advantage. Why data centre optimisation matters CIOs and IT leaders increasingly face calls to provide a high-performance foundational compute infrastructure across their businesses and handle new, more demanding, use cases while balancing sustainability commitments, space and budget constraints. Many have sought to build new data centres outright to meet demand and pair them with energy efficient technologies to minimise their environmental impact. For example, the LUMI (Large Unified Modern Infrastructure) Supercomputer, one of the most powerful in Europe uses 100% carbon-free hydroelectric energy for its operations and its waste heat is reused to heat homes in the nearby town of Kajanni, Finland. There are many other examples like LUMI showing the considerable progress the data centre industry have made in addressing the need for energy efficiency. Yet energy efficiency alone won't be enough to power the growing demands of AI which is expected to plump up data centre storage capacity. AI's greater energy requirements will also require more energy efficient designs to help ensure scalability and address environmental goals and with data centre square footage, land and power grids nearing capacity, one way to optimise design is to upgrade from old servers. Data centres are expensive investments, and some CIOs and IT leaders try to recoup costs by running their hardware for as long as possible. As a result, most data centres are still using hardware that is 10 years old (Dell) and only expand compute when absolutely necessary. While building new data centres might be necessary for some, there are significant opportunities to upgrade existing infrastructure. Upgrading to newer systems means data centres can achieve the same tasks more efficiently. Global IT data centre capacity will grow from 180 Gigawatts (GW) in 2024 to 296 GW in 2028, representing a 12.3% CAGR, while electricity consumption will grow at a higher rate 23.3% from 397 Terawatt hours (TWh) to 915 TWh in 2028. For the ageing data centres, that can translate to fewer racks and systems to manage, while still maintaining the same bandwidth. It can leave significant room for future IT needs but also makes room for experimentation which is absolutely necessary in AI workloads at the moment. They can use the space to build less expensive proof of concept half racks before it leads to bigger build outs and use new hyper-efficient chips to help reduce energy consumption and cooling requirements, recouping investment back more quickly. What to look for in an upgrade There are many factors to consider in a server upgrade and there isn't a one size fits all solution to data centre needs. It's not just about buying the most powerful chip that can be afforded. Yes, the significance of a good chip on energy efficiency cannot be overstated, but each data centre has different needs that will shape the hardware and software stack they need to operate most efficiently. Leading South Korean cloud provider, Kakao Enterprise, needed servers that can deliver high performance across a wide range of workloads to support its expansive range of offerings. By deploying a mixed fleet of 3rd and 4th Gen AMD EPYC processors, the company was able to reduce the server required for its total workload to 40 percent of its original fleet, while achieving increased performance by 30 percent, with a 50 percent reduction in total cost of ownership. Much like Kakao Enterprise, IT decision makers should look for providers that can deliver end-to-end data centre Infrastructure at scale combining high performance chips, networking, software and systems design expertise. For example, the right physical racks make it easy to swap in new kit as needs evolve, and having open software is equally important for getting the different pieces of the software stack from different providers talking with each other. In addition, providers that are continually investing in world class systems design and AI systems capabilities will be best positioned to accelerate enterprise AI hardware and software roadmaps. AMD, for example, recently achieved a 38× improvement in node-level energy efficiency for AI training and HPC over just five years. This translates to a 97% reduction in energy for the same performance, empowering providers and end-users alike to innovate more sustainably and at scale. Advancing the Data Centre As our reliance on digital technologies continues to grow, so too does our need for computing power. It is important to balance the need for more compute real estate with sustainability goals, and the way forward is in making the most out of the existing real estate we have. This is a big opportunity to think smartly about this and turn an apparent tension into a massive advantage. By using the right computational architecture, data centres can achieve the same tasks more efficiently, making room for the future technologies that will transform businesses and lives.

HPE expands AI server range with NVIDIA Blackwell GPU solutions
HPE expands AI server range with NVIDIA Blackwell GPU solutions

Techday NZ

time3 days ago

  • Techday NZ

HPE expands AI server range with NVIDIA Blackwell GPU solutions

Hewlett Packard Enterprise has introduced several updates to its NVIDIA AI Computing by HPE portfolio, aimed at supporting enterprise clients seeking to accelerate agentic and physical AI deployment across a variety of use cases. Server advancements Among the headline updates, HPE has confirmed it will ship new HPE ProLiant Compute servers equipped with NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs. This includes a new 2U RTX PRO Server form factor in the DL385 Gen11 model, as well as an 8-GPU 4U configuration with the DL380a Gen12 model. According to HPE, the DL385 Gen11 supports up to two of the Blackwell Server Edition GPUs, providing an air-cooled solution suitable for datacentres coping with increasing artificial intelligence workloads. Meanwhile, the DL380a Gen12 can accommodate up to eight GPUs in a larger form factor, with shipments scheduled to begin in September 2025. HPE highlighted that the ProLiant Compute servers are purpose-built for handling a variety of tasks, including generative and agentic AI, robotics, industrial automation, visual computing, simulation, 3D modelling, digital twins, and autonomous systems. Security features on the Gen12 models include HPE Integrated Lights Out 7 Silicon Root of Trust and a secure enclave for tamper-resistant protection and quantum-resistant firmware signing. The company states that its server management platform, HPE Compute Ops Management, can reduce IT hours spent on server management by up to 75% and lower downtime by an average of 4.8 hours per server annually. HPE has also indicated that these servers are designed to be flexible and scalable, able to support a growing range of GPU-accelerated workloads across the enterprise. AI development platform HPE Private Cloud AI, a collaborative development with NVIDIA, will incorporate support for the latest NVIDIA AI models. This includes the NVIDIA Nemotron agentic AI model, Cosmos Reason vision language model for robotics and physical AI, and the NVIDIA Blueprint for Video Search and Summarization (VSS 2.4). These additions will allow customers to build and deploy video analytics AI agents that can process extensive volumes of video data and extract actionable insights. The new release promises seamless scalability across GPU generations, air-gapped management, and enterprise multi-tenancy. Continuous integration with NVIDIA technologies will also allow HPE Private Cloud AI to deliver rapid deployment of NVIDIA NIM microservices, with access provided via HPE AI Essentials. The platform is positioned to help enterprises handle increasing AI inferencing workloads while retaining control over their data, supporting high performance and security requirements in demanding sectors. Regional and industry response "Asia Pacific is one of the fastest-growing AI markets, and enterprises face the imperative to transform ambition into results, with agility and security at the core," said Joseph Yang, General Manager, HPC, AI & NonStop, at HPE APAC and India. "With NVIDIA Blackwell GPUs in our HPE ProLiant servers and the latest NVIDIA AI models in HPE Private Cloud AI, we're enabling customers across APAC to accelerate agentic and physical AI, powering everything from advanced manufacturing to smart cities, while safeguarding data sovereignty and maximizing operational efficiency." Data sovereignty and operational efficiency were also cited as important capabilities for regional customers working in sectors such as advanced manufacturing and public infrastructure. "HPE is committed to empowering enterprises with the tools they need to succeed in the age of AI," said Cheri Williams, Senior Vice President and General Manager for Private Cloud and Flex Solutions at HPE. "Our collaboration with NVIDIA continues to push the boundaries of innovation, delivering solutions that unlock the value of generative, agentic and physical AI while addressing the unique demands of enterprise workloads. With the combination of HPE ProLiant servers and expanded capabilities in HPE Private Cloud AI, we're enabling organizations to embrace the future of AI with confidence and agility." The collaboration between HPE and NVIDIA is expected to support customers managing large-scale enterprise AI workloads, with the infrastructure designed to be as flexible and scalable as present and emerging tasks require. "Enterprises need flexible, efficient infrastructure to keep pace with the demands of modern AI," said Justin Boitano, Vice President of Enterprise AI at NVIDIA. "With NVIDIA RTX PRO 6000 Blackwell GPUs in HPE's 2U ProLiant servers, enterprises can accelerate virtually every workload on a single, unified, enterprise-ready platform." Availability The HPE ProLiant DL385 Gen11 and DL380a Gen12 servers equipped with NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs are currently open for orders, with first shipments expected from September 2025. HPE intends to roll out support for the newest NVIDIA AI models, the Cosmos Reason VLM, and the VSS 2.4 blueprint in HPE Private Cloud AI during the latter half of 2025. The next generation of HPE Private Cloud AI, with Blackwell GPU support, is also slated for release in the same period.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store