logo
AMD Stakes Future on Open AI Infrastructure

AMD Stakes Future on Open AI Infrastructure

Arabian Post10 hours ago

Advanced Micro Devices projected bold expectations for its artificial intelligence trajectory during its Advancing AI event in San Jose on 12 June 2025, emphasising system-level openness and ecosystem collaboration. CEO Dr Lisa Su unveiled the Instinct MI350 accelerator series, introduced plans for the Helios rack-scale AI server launching in 2026, and fortified AMD's software stack to challenge incumbent leaders in the sector.
Top-tier AI customers including OpenAI, Meta, Microsoft, Oracle, xAI and Crusoe pledged significant investments. OpenAI's CEO Sam Altman joined Su onstage, confirming the firm's shift to MI400-class chips and collaboration on MI450 design. Crusoe disclosed a $400 million commitment to the platform.
MI350 Series, which includes the MI350X and MI355X, are shipping to hyperscalers now, with a sharp generational performance leap — delivering about four times the compute capacity of prior-generation chips, paired with 288 GB of HBM3e memory and up to 40% better token‑per‑dollar performance than Nvidia's B200 models. Initial deployments are expected in Q3 2025 in both air‑ and liquid‑cooled configurations, with racks supporting up to 128 GPUs, producing some 2.6 exaflops FP4 compute.
ADVERTISEMENT
Looking further ahead, AMD previewed 'Helios'—a fully integrated rack comprising MI400 GPUs, Zen 6‑based EPYC 'Venice' CPUs and Pensando Vulcano NICs, boasting 72 GPUs per rack, up to 50% more HBM memory bandwidth and system‑scale networking improvements compared to current architectures. Helios is poised for market launch in 2026, with an even more advanced MI500‑based variant expected around 2027.
Dr Su underscored openness as AMD's competitive lever. Unlike Nvidia's proprietary NVLink interface, AMD's designs will adhere to open industry standards—extending availability of networking architectures to rivals such as Intel. Su argued this approach would accelerate innovation, citing historical parallels from open Linux and Android ecosystems.
On the software front, the ROCm 7 stack is being upgraded with enterprise AI and MLOps features, including integrated tools from VMware, Red Hat, Canonical and others. ROCm Enterprise AI, launching in Q3 or early Q4, aims to match or exceed Nvidia's CUDA-based offerings in usability and integration.
Strategic acquisitions underpin AMD's infrastructure ambitions. The purchase of ZT Systems in March 2025 brought over 1,000 engineers to accelerate rack-scale system builds. Meanwhile, AMD has onboarded engineering talent from Untether AI and Lamini to enrich its AI software capabilities.
Market reaction was muted; AMD shares fell roughly 1–2% on the event day, with analysts noting that while the announcements are ambitious, immediate market share gains are uncertain.
Financially, AMD projects AI data centre revenues growing from over $5 billion in 2024 to tens of billions annually, anticipating the AI chip market reaching around $500 billion by 2028.
These developments position AMD as a serious contender in the AI infrastructure arena. Its push for rack‑scale systems and open‑standard platforms aligns with the growing trend toward modular, interoperable computing. Competition with Nvidia will intensify through 2026 and 2027, centred on performance per dollar in large‑scale deployments.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Big Tech on a quest for ideal AI device as legacy gizmos seem outdated
Big Tech on a quest for ideal AI device as legacy gizmos seem outdated

Khaleej Times

time6 hours ago

  • Khaleej Times

Big Tech on a quest for ideal AI device as legacy gizmos seem outdated

ChatGPT-maker OpenAI has enlisted the legendary designer behind the iPhone to create an irresistible gadget for using generative artificial intelligence (AI). The ability to engage digital assistants as easily as speaking with friends is being built into eyewear, speakers, computers and smartphones, but some argue that the Age of AI calls for a transformational new gizmo. "The products that we're using to deliver and connect us to unimaginable technology are decades old," former Apple chief design officer Jony Ive said when his alliance with OpenAI was announced. "It's just common sense to at least think, surely there's something beyond these legacy products." Sharing no details, OpenAI chief executive Sam Altman said that a prototype Ive shared with him "is the coolest piece of technology that the world will have ever seen." According to several US media outlets, the device won't have a screen, nor will it be worn like a watch or broach. Kyle Li, a professor at The New School, said that since AI is not yet integrated into people's lives, there is room for a new product tailored to its use. The type of device won't be as important as whether the AI innovators like OpenAI make "pro-human" choices when building the software that will power them, said Rob Howard of consulting firm Innovating with AI Learning from flops The industry is well aware of the spectacular failure of the AI Pin, a square gadget worn like a badge packed with AI features but gone from the market less than a year after its debut in 2024 due to a dearth of buyers. The AI Pin marketed by startup Humane to incredible buzz was priced at $699. Now, Meta and OpenAI are making "big bets" on AI-infused hardware, according to CCS Insight analyst Ben Wood. OpenAI made a multi-billion-dollar deal to bring Ive's startup into the fold. Google announced early this year it is working on mixed-reality glasses with AI smarts, while Amazon continues to ramp up Alexa digital assistant capabilities in its Echo speakers and displays. Apple is being cautious embracing generative AI, slowly integrating it into iPhones even as rivals race ahead with the technology. Plans to soup up its Siri chatbot with generative AI have been indefinitely delayed. The quest for creating an AI interface that people love "is something Apple should have jumped on a long time ago," said Futurum research director Olivier Blanchard. Time to talk Blanchard envisions some kind of hub that lets users tap into AI, most likely by speaking to it and without being connected to the internet. "You can't push it all out in the cloud," Blanchard said, citing concerns about reliability, security, cost, and harm to the environment due to energy demand. "There is not enough energy in the world to do this, so we need to find local solutions," he added. Howard expects a fierce battle over what will be the must-have personal device for AI, since the number of things someone is willing to wear is limited and "people can feel overwhelmed." A new piece of hardware devoted to AI isn't the obvious solution, but OpenAI has the funding and the talent to deliver, according to Julien Codorniou, a partner at venture capital firm 20VC and a former Facebook executive. OpenAI recently hired former Facebook executive and Instacart chief Fidji Simo as head of applications, and her job will be to help answer the hardware question. Voice is expected by many to be a primary way people command AI. Google chief Sundar Pichai has long expressed a vision of "ambient computing" in which technology blends invisibly into the world, waiting to be called upon. "There's no longer any reason to type or touch if you can speak instead," Blanchard said. "Generative AI wants to be increasingly human" so spoken dialogues with the technology "make sense," he added. However, smartphones are too embedded in people's lives to be snubbed any time soon, said Wood. © Agence France-Presse

AMD Stakes Future on Open AI Infrastructure
AMD Stakes Future on Open AI Infrastructure

Arabian Post

time10 hours ago

  • Arabian Post

AMD Stakes Future on Open AI Infrastructure

Advanced Micro Devices projected bold expectations for its artificial intelligence trajectory during its Advancing AI event in San Jose on 12 June 2025, emphasising system-level openness and ecosystem collaboration. CEO Dr Lisa Su unveiled the Instinct MI350 accelerator series, introduced plans for the Helios rack-scale AI server launching in 2026, and fortified AMD's software stack to challenge incumbent leaders in the sector. Top-tier AI customers including OpenAI, Meta, Microsoft, Oracle, xAI and Crusoe pledged significant investments. OpenAI's CEO Sam Altman joined Su onstage, confirming the firm's shift to MI400-class chips and collaboration on MI450 design. Crusoe disclosed a $400 million commitment to the platform. MI350 Series, which includes the MI350X and MI355X, are shipping to hyperscalers now, with a sharp generational performance leap — delivering about four times the compute capacity of prior-generation chips, paired with 288 GB of HBM3e memory and up to 40% better token‑per‑dollar performance than Nvidia's B200 models. Initial deployments are expected in Q3 2025 in both air‑ and liquid‑cooled configurations, with racks supporting up to 128 GPUs, producing some 2.6 exaflops FP4 compute. ADVERTISEMENT Looking further ahead, AMD previewed 'Helios'—a fully integrated rack comprising MI400 GPUs, Zen 6‑based EPYC 'Venice' CPUs and Pensando Vulcano NICs, boasting 72 GPUs per rack, up to 50% more HBM memory bandwidth and system‑scale networking improvements compared to current architectures. Helios is poised for market launch in 2026, with an even more advanced MI500‑based variant expected around 2027. Dr Su underscored openness as AMD's competitive lever. Unlike Nvidia's proprietary NVLink interface, AMD's designs will adhere to open industry standards—extending availability of networking architectures to rivals such as Intel. Su argued this approach would accelerate innovation, citing historical parallels from open Linux and Android ecosystems. On the software front, the ROCm 7 stack is being upgraded with enterprise AI and MLOps features, including integrated tools from VMware, Red Hat, Canonical and others. ROCm Enterprise AI, launching in Q3 or early Q4, aims to match or exceed Nvidia's CUDA-based offerings in usability and integration. Strategic acquisitions underpin AMD's infrastructure ambitions. The purchase of ZT Systems in March 2025 brought over 1,000 engineers to accelerate rack-scale system builds. Meanwhile, AMD has onboarded engineering talent from Untether AI and Lamini to enrich its AI software capabilities. Market reaction was muted; AMD shares fell roughly 1–2% on the event day, with analysts noting that while the announcements are ambitious, immediate market share gains are uncertain. Financially, AMD projects AI data centre revenues growing from over $5 billion in 2024 to tens of billions annually, anticipating the AI chip market reaching around $500 billion by 2028. These developments position AMD as a serious contender in the AI infrastructure arena. Its push for rack‑scale systems and open‑standard platforms aligns with the growing trend toward modular, interoperable computing. Competition with Nvidia will intensify through 2026 and 2027, centred on performance per dollar in large‑scale deployments.

G42 Plants Flag in London to Power Europe's AI Ambition
G42 Plants Flag in London to Power Europe's AI Ambition

Arabian Post

timea day ago

  • Arabian Post

G42 Plants Flag in London to Power Europe's AI Ambition

G42 has unveiled a new London‑based subsidiary, G42 Europe & UK, designed to amplify its AI infrastructure and services across the United Kingdom and continental Europe. The entity, co‑chaired by Omar Mir and Marty Edelman, will deliver end‑to‑end AI offerings—from advisory and model deployment to supercomputing infrastructure—for sectors including finance, healthcare, energy and manufacturing. The launch comes as G42 continues its strategic European rollout, following the roll‑out of AI compute clusters and data‑centre capabilities in France and Italy. Positioned at the heart of London, the new hub aims to serve as a local nexus for AI innovation, bridging the gulf between regulatory frameworks and technological delivery. Omar Mir, an international board member at World Wide Technology with more than two decades of experience in 5G, cloud and AI services across Europe and the Middle East, will co‑lead the initiative. His counterpart, Marty Edelman, G42's group general counsel, brings deep expertise in legal oversight and governance—a reflection of G42's emphasis on secure, sovereign deployment in regulated markets. ADVERTISEMENT Mir emphasised the vision: 'Our goal is to harness G42's proven AI expertise and localise it for European and UK businesses—fuelling digital transformation, enhancing competitiveness, and building resilient, sovereign AI infrastructure in partnership with public and private stakeholders'. Edelman added that the UK and Europe represent dynamic markets with significant scope for AI‑driven innovation, and that a London hub enables the company to align closely with local regulatory regimes. G42 Europe & UK will leverage the group's global computing backbone—comprising supercomputing nodes, data centres and AI‑modeling platforms—to deliver a full spectrum of capabilities. Services will span strategic advisory, AI model development, infrastructure deployment and managed services, all with a focus on local data sovereignty and regulatory compliance. The initiative complements G42's earlier announcements of strategic investments in AI infrastructure in France and Italy. In Italy, G42 is partnering on a €1 billion supercomputing build with iGenius, while in France it has established an AMD‑powered AI facility in Grenoble. The London operation is expected to accelerate regulatory engagement and foster partnerships tailored to the European sovereignty agenda. Analysts suggest the move is part of a broader ambition by the UAE and G42 to position themselves as credible alternatives to US and Chinese cloud and AI providers. By offering sovereign, locally governed infrastructure, G42 aims to appeal to European clients seeking secure, high‑performance computing solutions. G42's broader portfolio, including Khazna data centres, Core42 sovereign‑cloud capabilities, cybersecurity firm CPX, analytics outfit Presight and AI lab Inception, will support the London hub's operations. This integrated ecosystem underpins G42's 'Intelligence Grid', a unified platform delivering enterprise‑grade AI research, analytics and cloud services. London's strategic significance is two‑fold. As a pre‑Brexit financial and regulatory anchor, it provides a gateway to both UK and EU markets. Moreover, it enhances G42's ability to work alongside national and regional authorities in shaping next‑generation AI infrastructure and data governance frameworks. Employee responses on social media underline the positive reception. A post from G42's LinkedIn noted the subsidiary 'will drive localized AI solutions and lead infrastructure build‑outs across the UK and continental Europe,' garnering broad applauds from industry peers. Industry observers acknowledge that G42's European foothold aligns with wider trends in AI geopolitics. Investment in local compute and data sovereignty reflects growing European determination to reduce dependency on US or Chinese technology ecosystems while scaling home‑grown digital capacity.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store