logo
Hyperscalers Form ASIC Coalition to Challenge NVIDIA Dominance

Hyperscalers Form ASIC Coalition to Challenge NVIDIA Dominance

Arabian Post2 days ago

Cloud computing giants AWS, Google, Microsoft, Meta and OpenAI are accelerating in-house development of custom application‑specific integrated circuits, aiming to erode NVIDIA's dominance in high‑performance AI datacentres. Industry reports highlight a projected annual growth rate of around 50% for ASIC purchases by hyperscalers, marking a strategic pivot in the AI hardware landscape.
NVIDIA's premium-priced solutions—including Blackwell GPUs—have placed pressure on hyperscalers to secure more cost‑efficient, scalable systems. With single GPUs ranging from $70,000 to $80,000 and fully configured servers tallying up to $3 million, these companies are betting on internal design to manage costs and supply risks.
Amazon Web Services has notably moved ahead with its in‑house chips—Trainium for training and Inferentia for inference—reporting 30 – 40% greater cost efficiency compared with NVIDIA hardware. AWS is also collaborating with Marvell and Taiwan's Alchip on next‑generation Trainium versions. Internal indications suggest AWS may deploy as many as half‑a‑million ASIC units in its data centres, an expansive scale‑up that could rival NVIDIA's installed base.
ADVERTISEMENT
Google, meanwhile, has scaled its TPU v6 Trillium chips, transitioning from single‑supplier to dual‑supplier design by partnering with MediaTek. With deployments reportedly hitting 100,000‑unit clusters to support Gemini 2.0 workloads, Google claims competitive cost-performance metrics relative to NVIDIA GPUs. Microsoft's forthcoming Maia 200 chip, co‑designed with GUC using TSMC's 3 nm process, is scheduled for commercial release in 2026.
Meta's Meta Training and Inference Accelerator, developed alongside Broadcom, Socionext and GUC, is expected in early 2026 on TSMC's 3 nm node, featuring HBM3e memory—another step towards self‑sufficiency in AI compute. OpenAI has also announced a proprietary training processor, with mass production anticipated at TSMC by 2026.
Market projections reflect this tectonic shift. ASICs are poised to claim between $100 billion and $130 billion of custom AI accelerator spend by 2030, with Broadcom estimating a market of $60 billion to $90 billion by 2027. Traditional ASIC powerhouses—Broadcom, Marvell, MediaTek, Alchip and GUC—are experiencing surging demand as they support hyperscaler transitions.
Despite these developments, hyperscalers continue to reserve capacity for NVIDIA chips, recognising the GPU giant's entrenched ecosystem—especially its CUDA software stack—and the steep technical barriers to immediate elimination of GPU dependencies.
The trend resembles historical transitions in specialised compute. Just as cryptocurrency mining moved from GPUs to ASICs for lower costs and greater efficiency, hyperscalers now aim to fragment the AI compute supply chain and diversify their hardware portfolios.
ADVERTISEMENT
TSMC stands to benefit significantly, serving as the foundry for both NVIDIA's mass‑market GPUs and hyperscaler ASICs. Its chairman emphasises that the competition between NVIDIA and cloud‑designed chips is ultimately beneficial to TSMC, ensuring a broad customer base.
Broadcom has emerged as a frontrunner, with its ASIC and networking chipset revenues soaring 220% to $12.2 billion in 2024. Hyperscalers are investing in clusters featuring up to one million custom XPUs over open‑Ethernet networks—an architecture that places Broadcom and Marvell in strategic positions. Networking ASICs are expected to account for 15–20% of AI data‑centre silicon budgets, rising from the 5–10% range.
Revenue trends reflect these structural shifts. Marvell has secured a multi‑year AI chip deal with AWS and anticipates its AI silicon revenue jumping from $550 million in 2024 to over $2.5 billion in 2026. Broadcom, similarly, is redirecting significant investment toward hyperscaler ASIC demand.
Nevertheless, NVIDIA retains a commanding lead in AI training and general‑purpose GPU compute. Its end‑to‑end platform—from hardware to software—remains deeply embedded in the AI ecosystem. Custom ASICs, by contrast, offer task‑specific gains but lack the breadth of software compatibility that NVIDIA enables.
Analysts caution that the AI compute landscape is evolving toward a more fragmented, mixed‑architecture model combining GPUs and ASICs. Hyperscalers' shift signals strategic recognition of rising costs, supply constraints, and performance demands. Yet, they also underscore persistent obstacles: software ecosystem maturity, long development cycles, and the complexity of large‑scale deployment.
Questions remain regarding the timeframe in which hyperscalers can meaningfully shift workloads away from NVIDIA GPUs. Industry roadmaps project new ASIC deployments through 2026–27. Analysts expect GPU market share erosion may begin toward the end of the decade, provided in-house ASICs deliver consistent performance and efficiency.
The stage is set for a multi‑year contest in datacentre compute. NVIDIA faces increasing pressure from hyperscalers building bespoke chips to optimise workloads and control supply. The next evolution of AI infrastructure may look less like a GPU‑centric world and more like a diverse ecosystem of specialised, interlocking processors.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

HPE Unveils New AI Factory Solutions for Enterprises
HPE Unveils New AI Factory Solutions for Enterprises

TECHx

time25 minutes ago

  • TECHx

HPE Unveils New AI Factory Solutions for Enterprises

Home » Emerging technologies » Artificial Intelligence » HPE Unveils New AI Factory Solutions for Enterprises Hewlett Packard Enterprise (HPE) has announced new solutions aimed at accelerating the creation, adoption, and management of AI factories across all organizational types and the entire AI lifecycle. HPE revealed the expansion of its NVIDIA AI Computing by HPE portfolio, now including NVIDIA Blackwell GPUs. The update introduces new composable solutions designed for service providers, model builders, and sovereign entities. It also includes the next-generation HPE Private Cloud AI, a turnkey AI factory for enterprises. The new end-to-end solutions eliminate the need for customers to compile their own AI tech stack when building AI-ready data centers. According to HPE President and CEO Antonio Neri, achieving AI potential requires strong infrastructure and the right IT foundation. He stated that HPE and NVIDIA offer a comprehensive approach to support organizations in realizing sustainable business value through AI. Jensen Huang, founder and CEO of NVIDIA, reported that HPE and NVIDIA are jointly delivering full-stack AI factory infrastructure to help businesses innovate at scale with speed and precision. HPE's Private Cloud AI offers a fully integrated solution featuring NVIDIA accelerated computing, networking, and software. It supports: NVIDIA Blackwell GPUs and HPE ProLiant Compute Gen12 servers Investment protection and seamless GPU scalability Air-gapped management for data privacy and multi-tenancy for collaboration It also includes NVIDIA AI Blueprints and a 'try and buy' program at Equinix data centers. HPE introduced new validated AI factory solutions leveraging five decades of liquid cooling expertise and HPE Morpheus Enterprise Software. These modular stacks offer a unified control plane and faster deployment. HPE OpsRamp now provides full-stack observability and is validated for the NVIDIA Enterprise AI Factory. Additional AI factory models include: A large-scale design for service providers and model builders using HPE ProLiant XD, NVIDIA AI Enterprise, and advanced cooling A solution for sovereign entities with enhanced privacy and sovereignty features The HPE Compute XD690, which supports eight NVIDIA Blackwell Ultra GPUs, has also been added to the portfolio. It includes the HPE Performance Cluster Manager for managing complex AI environments. To support data-hungry AI workloads, HPE Alletra Storage MP X10000 now supports Model Context Protocol (MCP) servers. The system accelerates AI data pipelines and supports the NVIDIA AI Data Platform reference design. HPE's Unleash AI ecosystem has expanded to 75+ use cases and added 26 new partners. These use cases span agentic AI, smart cities, data governance, and cybersecurity. Additionally, HPE and Accenture are co-developing agentic AI solutions for the financial sector. The collaboration uses Accenture's AI Refinery on HPE Private Cloud AI to explore applications in sourcing, spend management, and contract analysis. To support customer adoption, HPE has introduced new services to design, finance, deploy, and manage AI factories. These offerings aim to simplify AI journeys from planning to long-term operation. HPE Financial Services is also offering flexible financing, including lower initial payments for Private Cloud AI and options to fund new AI projects using existing tech assets. HPE continues to position itself as a leader in enterprise AI by delivering complete, integrated solutions that support innovation and scale.

HPE unveils new AI factory solutions built with NVIDIA to accelerate AI adoption at global scale
HPE unveils new AI factory solutions built with NVIDIA to accelerate AI adoption at global scale

Zawya

time3 hours ago

  • Zawya

HPE unveils new AI factory solutions built with NVIDIA to accelerate AI adoption at global scale

HPE drives AI innovation with modular AI factory solutions powered by NVIDIA Blackwell, fueled by HPE Alletra Storage MP, and optimized to deploy and scale across enterprises, service providers, sovereign entities, and model builders HPE Private Cloud AI adds investment protection, air-gapped management, and HPE ProLiant Compute Gen12 servers with the NVIDIA RTX PRO 6000 Blackwell GPUs The new HPE Compute XD690 with eight NVIDIA Blackwell Ultra GPUs is the latest entry into NVIDIA AI Computing by HPE portfolio HPE Alletra Storage MP X10000 fuels AI factories with AI-ready data through full integration of Model Context Protocol (MCP) HPE OpsRamp Software is a validated observability solution for NVIDIA Enterprise AI Factory HPE fast tracks AI value creation with support for more than 75 use cases through 26 new Unleash AI partners and the latest NVIDIA AI Blueprints HPE and Accenture are collaborating to bring to market new agentic AI solutions for financial services using Accenture's AI Refinery platform with NVIDIA on HPE Private Cloud AI​​​​ Dubai, United Arab Emirates – HPE Discover Las Vegas 2025– HPE (NYSE: HPE) today announced new solutions to supercharge the creation, adoption and management of AI factories across the entire AI lifecycle and for every type of organization. HPE is expanding its NVIDIA AI Computing by HPE portfolio of AI factory solutions with NVIDIA Blackwell GPUs, including new composable solutions optimized for service providers, model builders and sovereign entities, as well as the next generation of HPE Private Cloud AI, the turnkey AI factory for enterprises. The integrated end-to-end solutions and services for AI factories remove the complexity of customers having to compile a full AI tech stack on their own when building a modern AI-ready data center. 'Generative, agentic and physical AI have the potential to transform global productivity and create lasting societal change, but AI is only as good as the infrastructure and data behind it. Organizations need the data, intelligence and vision to capture the AI opportunity and this makes getting the right IT foundation essential,' said Antonio Neri, president and CEO, at HPE. 'HPE and NVIDIA are delivering the most comprehensive approach, joining industry-leading AI infrastructure and services to enable organizations to realize their ambitions and deliver sustainable business value.' 'We are entering a new industrial era — one defined by the ability to generate intelligence at scale,' said Jensen Huang, founder and CEO of NVIDIA. 'Together, HPE and NVIDIA are delivering full-stack AI factory infrastructure to drive this transformation, empowering enterprises to harness their data and accelerate innovation with unprecedented speed and precision.' HPE speeds AI factory time to value with next-generation HPE Private Cloud AI The signature offering in the NVIDIA AI Computing by HPE portfolio, Private Cloud AI is a fully integrated turnkey AI factory offering with NVIDIA accelerated computing, networking and software. Private Cloud AI will provide: Support for NVIDIA Blackwell accelerated computing with HPE ProLiant Compute Gen12 servers. These servers rank #1 in over 23 AI tests [1] and provide secure enclaves to prevent tampering, as well as post-quantum cryptography and trusted supply chain capabilities at rack and server level. Investment protection and seamless scalability from one GPU generation to the next, including NVIDIA H200 NVL and now with NVIDIA RTX PRO 6000 Server Edition GPUs, to support a wide range of enterprise AI workloads including agentic and physical AI use cases. A new federated architecture unifies resource pooling to make new GPUs and resources available to all AI workloads. Air-gapped management for organizations with strict data privacy requirements. Multi-tenancy that enables enterprises to collaborate and partition resources across teams. The latest NVIDIA AI Blueprints including the NVIDIA AI-Q Blueprint for AI agent creation and workflows. A new 'try and buy' program that lets customers test Private Cloud AI across Equinix's global footprint of high-performance data centers before making a purchase. HPE expands AI factory portfolio with new validated solutions built on industry-leading innovation New AI factory solutions combine the latest AI innovations and expertise from HPE, including five decades of industry-leading liquid cooling and HPE Morpheus Enterprise Software for a unified control plane. Unlike most competing solutions, these validated, end-to-end composable solutions are modular, full tech stacks integrated ahead of time for the customer, speeding time to value. Using AI to Observe AI: HPE OpsRamp Software provides full-stack observability across Private Cloud AI and new AI factory solutions. HPE OpsRamp is also now a validated observability solution for the NVIDIA Enterprise AI Factory. New AI factory at scale is designed for organizations operating at large scale such as service providers and model builders. HPE offers HPE ProLiant Compute XD, NVIDIA AI Enterprise software and blueprints, air- and liquid-cooling technologies and a full range of advisory and professional services. New AI factory for sovereigns is for nations, governments and public sector organizations and offers additional specialized capabilities such as air-gapped management and solutions and services that enable data, technological and operational sovereignty. By leveraging the NVIDIA Enterprise AI Factory validated design, AI factory solutions from HPE can be deployed using the latest NVIDIA accelerated computing, NVIDIA Spectrum-X Ethernet networking, NVIDIA BlueField-3 DPUs, and NVIDIA AI Enterprise software to ensure high performance, robust security, efficient storage acceleration, and scalable infrastructure for the next era of AI. HPE Compute XD690 joins NVIDIA AI Computing by HPE portfolio New HPE Compute XD690 supports eight NVIDIA Blackwell Ultra GPUs and reflects HPE's continued commitment to be a time-to-market leader with NVIDIA's latest innovations in accelerated computing for AI. HPE Performance Cluster Manager delivers fully integrated systems management and advanced infrastructure monitoring and alerting across large, complex AI environments that scale thousands of nodes. HPE Alletra Storage MP X10000 fuels AI factories, agents and use cases with AI-ready data To ensure AI factories and applications are continuously fueled with AI-ready unstructured data, HPE Alletra Storage MP X10000 will support Model Context Protocol (MCP) servers. Integrating MCP with the X10000's built-in data intelligence accelerates data pipelines and enables AI factories, applications, and agents to process and act on intelligent unstructured data. Additionally, the X10000 supports the NVIDIA AI Data Platform reference design and offers an SDK to streamline unstructured data pipelines for ingestion, inferencing, training and continuous learning Unleash AI adds 26 new partners, enabling HPE to support more than 75 AI use cases Through new Unleash AI ecosystem partners, the latest NVIDIA AI Blueprints and internal professional services, HPE delivers more than 75 AI use cases to customers. The Unleash AI partner ecosystem now features software solutions for agentic AI, sovereign AI, smart cities, industrial and manufacturing applications, data governance and privacy, responsible AI, video analytics, responsible AI, security and cybersecurity. HPE Private Cloud AI and Accenture AI Refinery simplify agentic AI for financial services HPE and Accenture are collaborating to create agentic AI solutions for financial services and procurement with a new joint go-to-market offering that leverages the Accenture AI Refinery™ platform, which is built on NVIDIA AI Enterprise, and deployed on HPE Private Cloud AI. HPE is applying the solution within its finance organization, exploring uses across category and sourcing strategies, spend management, strategic relationship analysis and contract obligation management. Together, the companies aim to help spur AI adoption and digital innovation in one of the most tightly regulated industries. HPE Services accelerate and simplify AI journey to speed time to value HPE eases the burden of building and adopting AI factories and speeds time to value through new services that assist customers on their AI journey. The new services span designing, financing, building and securing AI factories from Day -1 to Day 2+ with a focus on operating efficiently and sustainably. The services also cover model development and migration, business benefit analysis, deployment, education, management, ongoing support and tech refreshes. HPE Financial Services enables enterprises to start their AI journey faster HPE Financial Services (HPEFS) offers a new program that helps enterprises begin their AI projects sooner with lower payments for Private Cloud AI for the first six months. HPEFS also offers a comprehensive suite of customizable financing and IT lifecycle services to help accelerate the move to AI, including a program that allows customers to use their current tech estate as a source of capital to fund additional AI projects. Availability HPE ProLiant Compute DL380a Gen12 servers are now available to order with NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs. The next generation of HPE Private Cloud AI with NVIDIA RTX PRO 6000 Blackwell GPUs will be released in the second half of 2025. New AI factory solutions are available now. HPE Alletra Storage MP X10000 with Model Context Protocol support is planned for the second half of 2025. HPE Compute XD690 will be available in October 2025. New AI services are available now. About HPE HPE (NYSE: HPE) is a leader in essential enterprise technology, bringing together the power of AI, cloud, and networking to help organizations achieve more. As pioneers of possibility, our innovation and expertise advance the way people live and work. We empower our customers across industries to optimize operational performance, transform data into foresight, and maximize their impact. Unlock your boldest ambitions with HPE. Discover more at Media Contact: Ronak Thakkar, Senior Vice President, FleishmanHillard. Financing and service offerings available through Hewlett-Packard Financial Services Company and its subsidiaries and affiliates (collectively, 'HPFSC') in certain countries and is subject to credit approval and execution of standard HPFSC documentation. Rates and terms are based on customer's credit rating, offering types, services and/or equipment type and options. Not all customers may qualify. Not all services or offers are available in all countries. Other restrictions may apply. HPFSC reserves the right to change or cancel this program at any time without notice [1] ' HPE sets new AI inference world records: Continued excellence in performance,' April 02, 2025.

NVIDIA architecture helps accelerate Giga Computing solutions
NVIDIA architecture helps accelerate Giga Computing solutions

Tahawul Tech

time4 hours ago

  • Tahawul Tech

NVIDIA architecture helps accelerate Giga Computing solutions

In a world increasingly driven by AI, data-intensive workloads, and real-time processing, Giga Computing—a GIGABYTE subsidiary—has introduced a game-changing portfolio of rack-scale solutions that redefine what's possible in modern data centres. Engineered to support the newly launched NVIDIA Blackwell Architecture, these systems are purpose-built for generative AI, LLMs, HPC, and visual computing workloads. Meet GIGAPOD: A Complete AI Data Centre Solution At the heart of Giga Computing's NVIDIA Blackwell-powered compute ecosystem is GIGAPOD—a fully integrated, scalable AI infrastructure designed to meet the demands of tomorrow's compute-heavy workloads today. GIGAPOD brings together: GIGABYTE GPU servers with NVIDIA GPU baseboards The intuitive and powerful GIGABYTE POD Manager software for streamlined control and efficiency Scalable deployment options, from nine air-cooled racks to five liquid-cooled racks This modular approach means organizations of all sizes—from startups building LLMs to hyperscalers managing global AI operations—can scale performance, cooling, and energy efficiency based on their unique infrastructure requirements. One Rack for Liquid. One for Air. Your Choice. Whether your data centre favours the familiarity of air cooling or the performance density of liquid, Giga Computing delivers. Air-Cooled Option : The 8U G893-series supports NVIDIA HGX™ B300 platforms and fits up to 32 GPUs in a single rack—ideal for traditional data centre builds. : The 8U G893-series supports NVIDIA HGX™ B300 platforms and fits up to 32 GPUs in a single rack—ideal for traditional data centre builds. Liquid-Cooled Option: The 4U G4L3-series brings direct liquid cooling (DLC) to the forefront, using cold plates on all GPUs and CPUs to pack up to 64 GPUs per rack—a perfect fit for hyperscalers prioritising energy efficiency. Both platforms are supercharged by NVIDIA Blackwell Ultra GPUs, which deliver breakthrough compute performance and memory optimisation for AI reasoning, agentic AI, and video inference at scale. Introducing Exascale Performance: NVIDIA GB300 NVL72 The NVIDIA GB300 NVL72 represents the pinnacle of AI-first infrastructure—delivering exascale performance within a single rack. Built with NVIDIA Blackwell Ultra GPUs and advanced networking, this system propels inference and training capabilities to unprecedented levels, setting a new standard for the most demanding AI workloads. NVIDIA MGX™ Servers for AI and HPC Versatility Giga Computing's modular server lineup complements its rack-scale offerings with tailored, modular systems that offer unmatched flexibility. Key platforms include: XL44-SX0 : Hosts up to eight NVIDIA RTX PRO™ 6000 Blackwell Server Edition GPUs for AI, rendering, and visualization : Hosts up to eight NVIDIA RTX PRO™ 6000 Blackwell Server Edition GPUs for AI, rendering, and visualization XV23-VC0: Powered by the NVIDIA Grace™ CPU, optimized for AI, cloud-native workloads, and virtualization These servers seamlessly integrate NVIDIA ConnectX®-7, NVIDIA BlueField®-3 DPU and are deeply optimized across the NVIDIA software stack. Built for Visual Computing & Beyond The NVIDIA RTX PRO™ Blackwell Server Edition is purpose-designed for data centres that demand the best in AI acceleration, scientific computing, advanced graphics, and video applications. Whether for enterprise AI development or high-end content creation, this GPU delivers outstanding results in demanding environments. GIGAPOD: More Than Just Hardware What truly sets GIGAPOD apart is its comprehensive approach. It's not just about compute power—it's about managing infrastructure at scale. GIGAPOD includes: Performance-tuned clusters Streamlined deployment models Data storage and networking solutions The full-featured POD Manager for orchestrating AI development and operations Ready to Explore? From hyperscalers building AI factories to research institutions training the next big model, Giga Computing offers a future-proof foundation. Whether you're optimising for performance, energy efficiency, or deployment flexibility, GIGAPOD and NVIDIA Blackwell are ready to meet your needs. Discover the future of rack-scale data centers powered by Blackwell: Visit the GIGABYTE NVIDIA Blackwell Solutions Page For queries or more information, please contact our local partner Ingram Micro sales Image Credit: NVIDIA and GIGABYTE

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store