logo
#

Latest news with #TensorProcessingUnits

Red Hat & Google Cloud extend partnership for AI innovation
Red Hat & Google Cloud extend partnership for AI innovation

Techday NZ

time21-05-2025

  • Business
  • Techday NZ

Red Hat & Google Cloud extend partnership for AI innovation

Red Hat and Google Cloud have agreed to extend their partnership to focus on advancing artificial intelligence (AI) for enterprises, specifically with new developments in open and agentic AI solutions. The collaboration will bring together Red Hat's open source technologies and Google Cloud's infrastructure, along with Google's Gemma family of open AI models. This initiative aims to offer cost-effective AI inference and greater hardware choices for businesses deploying generative AI at scale. Brian Stevens, Senior Vice President and Chief Technology Officer – AI, Red Hat said, "With this extended collaboration, Red Hat and Google Cloud are committed to driving groundbreaking AI innovations with our combined expertise and platforms. Bringing the power of vLLM and Red Hat open source technologies to Google Cloud and Google's Gemma equips developers with the resources they need to build more accurate, high-performing AI solutions, powered by optimized inference capabilities." The latest phase of the alliance will see the companies launch the llm-d open source project, with Google acting as a founding contributor. This project is intended to facilitate scalable and efficient AI inference across diverse computing environments. Red Hat is introducing the project as a response to enterprise challenges, such as the growing complexity of AI ecosystems and the need for distributed computing strategies. The companies have also announced that support for vLLM, an open source inference server used to speed up generative AI outputs, will be enabled on Google Cloud's Tensor Processing Units (TPUs) and GPU-based virtual machines. Google Cloud's TPUs, which are already a part of Google's own AI infrastructure, will now be accessible to developers using vLLM, allowing for improved performance and resource efficiency for fast and accurate inference. Red Hat will be among the earliest testers for Google's new open model Gemma 3, and it will provide 'Day 0' support for vLLM on Gemma 3 model distributions. This is part of Red Hat's broader efforts as a commercial contributor to the vLLM project, focusing on more cost-effective and responsive platforms for generative AI applications. The collaboration also includes the availability of Red Hat AI Inference Server on Google Cloud. This enterprise distribution of vLLM helps companies scale and optimise AI model inference within hybrid cloud environments. The integration with Google Cloud enables enterprises to deploy generative AI models that are ready for production and can deliver cost and responsiveness efficiencies at scale. Supporting community-driven AI development, Red Hat will join Google as a contributor to the Agent2Agent (A2A) protocol, an application-level protocol designed to enable communication between agents or end-users across different platforms and cloud environments. Through the A2A ecosystem, Red Hat aims to promote new ways to accelerate innovation and enhance the effectiveness of AI workflows through agentic AI. Mark Lohmeyer, Vice President and General Manager, AI and Computing Infrastructure, Google Cloud, commented, "The deepening of our collaboration with Red Hat is driven by our shared commitment to foster open innovation and bring the full potential of AI to our customers. As we enter a new age of AI inference, together we are paving the way for organisations to more effectively scale AI inference and enable agentic AI with the necessary cost-efficiency and high performance." The llm-d project builds upon the established vLLM community, aiming to create a foundation for generative AI inference that can adapt to the demands of large-scale enterprises while facilitating innovation and cost management. The intention is to enable AI workload scalability across different resource types and enhance workload efficiency. These initiatives highlight the companies' collective effort to offer business users production-ready, scalable, and efficient AI solutions powered by open source technologies and robust infrastructure options.

Google Preparing to Partner with Taiwan's MediaTek on Next AI chip, Information Reports
Google Preparing to Partner with Taiwan's MediaTek on Next AI chip, Information Reports

Asharq Al-Awsat

time17-03-2025

  • Business
  • Asharq Al-Awsat

Google Preparing to Partner with Taiwan's MediaTek on Next AI chip, Information Reports

Alphabet's Google is preparing to partner with Taiwan's MediaTek on the next version of its AI chips, Tensor Processing Units, that will be made next year, the Information reported on Monday, citing people involved in the project. However, Google has not cut ties with Broadcom, the chip designer it has worked with exclusively on the AI chips over the past several years, the report said, citing an employee at the San Jose-based company. Like Nvidia,. Google also designs its own AI server chips, which it uses for internal research and development and also rents out to cloud customers, Reuters reported. This approach gives Google a competitive edge in the AI race by reducing its reliance on Nvidia, even as rivals like Microsoft-backed OpenAI and Meta Platforms have seen a surge in demand for Nvidia chips. Late last year, Google rolled out its sixth-generation TPU in a bid to give itself and its Cloud customers an alternative to Nvidia's chips, which are the most sought-after processors in the industry. Google chose MediaTek partly because the Taiwanese firm has a strong relationship with TSMC and charges Google less per chip compared to Broadcom, the Information report added. Google spent between $6 billion and $9 billion on TPUs last year, according to research firm Omdia, based on Broadcom's target for AI semiconductor revenue last year.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store