logo
#

Latest news with #MarkLohmeyer

Google Cloud Gets More Serious About Infrastructure At Next 2025
Google Cloud Gets More Serious About Infrastructure At Next 2025

Forbes

time24-04-2025

  • Business
  • Forbes

Google Cloud Gets More Serious About Infrastructure At Next 2025

Google Cloud was especially bold in its competitive positioning against AWS at the Google Cloud Next ... More 2025 conference. Here, Mark Lohmeyer, vice president and general manager of AI and computing infrastructure at Google Cloud, presents head-to-head comparisons. This month's Google Cloud Next 2025 event was an excellent reference point for how far Google Cloud has come since CEO Thomas Kurian took the helm of the business at the start of 2019. Back then, Google Cloud had about $6 billion in revenue and was losing a ton of money; six years later, it's nearing a $50 billion annual run rate, and it's profitable. I remember that when Kurian started, early odds were that Google would get out of the cloud service business altogether — yet here we are. Typically for this conference, there was so much announced that I can't cover it all here. (Among the many progress stats that Kurian cited onstage: the business shipped more than 3,000 product advances in 2024.) For deeper dives into specific areas, see the articles from my colleagues Matt Kimball on the new Ironwood TPU chip, Jason Andersen on Google's approach to selling enterprise AI (especially agents) and Melody Brue on the company's approach to the connected future of AI in the workplace. Our colleague Robert Kramer also wrote an excellent preview of the event that still makes good background reading. What I want to focus on here are Next 25's most interesting developments in connectivity, infrastructure and AI. (Note: Google is an advisory client of my firm, Moor Insights & Strategy.) Kurian placed a strong focus on connectivity, specifically with the company's new Cloud WAN and Cloud Interconnect offerings. Cloud WAN makes the most of Google's network, which the company rightly calls 'planet-scale,' to deliver faster performance than the public internet (40% faster, according to the company) that's also significantly cheaper than enterprise WANs (with a claimed 40% lower TCO). Meanwhile, Cloud Interconnect is built to connect your own enterprise network to Google's — or even to your network hosted by a different CSP — with high availability and low latency. Interestingly, in the analyst readout at the conference, Kurian started off with networking, which highlights its importance to Google. This makes sense, as enterprises are all bought into the hybrid multicloud and the growing need to connect all those datacenters, whether public or private cloud. This went hand in hand with a lot of discussion about new infrastructure. For context, all of the hyperscalers have announced extra-large capex investments in infrastructure for this year, with Google weighing it at $75 billion. The presentations at Next 25 showed where a good chunk of that money is going. I'll talk more below about the infrastructure investments specific to AI, starting with the Ironwood TPU chip and AI Hypercomputer. For now I want to note that the infrastructure plays also include networking offload, new storage options, a new CPU . . . It's a long list, all aimed at supporting Google Cloud's strategy of combining hardware and software to enable bigger outputs — especially in AI — at a low price. Make special note of that low price element, which is unusual for Google. I'll come back to that in a minute. Strategically, I think that Google is recognizing that infrastructure as a service is an onramp to PaaS and SaaS services revenue. If you can get people signed on for your IaaS — because, say, you have competitive compute and storage and a planet-scale network that you're allowing them to piggyback on — that opens the door for using a bigger selection of your offerings at the platform level. And while we're at it, why not a PaaS or SaaS approach to handling a bigger slice of your enterprise AI needs? It's a solid move from Google, and I'm intrigued to see how it plays out competitively, especially given that Azure seemed to get serious about IaaS in the past couple of years. It's also notable that Next 25 is the first time I can remember Google Cloud going after AWS on the infrastructure front. As shown in the image accompanying this article, Google touts its Arm-based Axion CPU as outperforming the competing Arm-based processor from AWS, Graviton. In the Mark Lohmeyer breakout session, there was a lot of specific discussion of AWS Trainium chips, too. I'm a fan of stiff competition, so it's refreshing to see Google getting more aggressive with this. It's about time. Considering all the years I spent in the semiconductor industry, it's no surprise that my ears perked up at the announcement of Google's seventh-generation Ironwood tensor processing unit, which comes out later this year. (I wish Google had been more specific about when we can expect it, but so far it's just 'later in 2025.') Google was a pioneer in this area, and this TPU is miles ahead of its predecessors in performance, energy efficiency, interconnect and so on. My colleague Matt Kimball has analyzed Ironwood in detail, so I won't repeat his work here. I will note briefly that Google's Pathways machine-learning runtime can manage distributed workloads across thousands of TPUs, and that Ironwood comes in scale-up pods of 256 chips or 9,216 chips. It also natively supports the vLLM library for inference. vLLM is an accepted abstraction layer that enterprises can comfortably code to for their optionality, and it should allow users to run inference on Ironwood with an appealing price-to-performance profile — yet another instance of combining hardware and software to enable more output at a manageable price. Next 25 was also the enterprise coming-out party for the Gemini 2.5 model, which as I write this is the best AI model in the world according to Hugging Face's Chatbot Arena LLM Leaderboard. The event showcased some impressive visual physics simulations using the model. (Google also put together a modification of The Wizard of Oz for display on the inner surface of The Sphere in Las Vegas. I can be pretty jaded about that kind of thing, but in this case I was genuinely impressed.) I haven't been a big consumer of Google's generative AI products in the past, even though I am a paying customer for Workspace and Gemini. But based on what I saw at the event and what I'm hearing from people in my network about Gemini 2.5, I'm going to give it another try. For now, let's focus on what Google claims for the Gemini 2.0 Flash model, which allows control over how much the model reasons to balance performance and cost. In fact, Google says that Gemini 2.0 Flash achieves intelligence per dollar that's 24x better than GPT-4o and 5x better than DeepSeek-R1. Again, I want to emphasize how unusual the 'per dollar' part is for Google messaging. Assuming the comparison figures are accurate, Google Cloud is able to achieve this by running its own (very smart) models on its new AI Hypercomputer system, which benefits from tailored hardware (including TPUs), software and machine learning frameworks. AI Hypercomputer is designed to allow easy adaptation of hardware so it can make the most of new advances in chips. On a related note, Google says that it will be one of the first adopters of Nvidia's GB200 GPUs. At the keynote, there was also a video of Nvidia CEO Jensen Huang in which he praised the partnership between the two companies and said, 'No company is better at every single layer of computing than Google.' In my view, Google is doing a neat balancing act to reassure the market that it loves Nvidia — while also creating its own wares to deliver better price per outcome. Touting itself for delivering the best intelligence at the lowest cost was not something I expected from Google Cloud. But as I reflect on it, it makes sense. Huang has a point: even though it's a fairly distant third place in the CSP market, Google really is good at every layer of the computing stack. It has the homegrown chips. The performance of its homegrown AI models is outstanding. It understands the (open) software needed to deliver AI for enterprise uses. And it's only getting stronger in infrastructure, as Next 25 emphasized. Now it wants to take this a step further by using Google Distributed Cloud to bring all of that goodness on-premises. Imagine running high-performing Gemini models, Agentspace and so on in your own air-gapped environment to support your enterprise tools and needs. In comparison to this, I thought that the announcements at Next 25 about AI agents were perfectly nice, but not any kind of strategic change or differentiator for the company — at least not yet. To be sure, Google is building out its agent capabilities both internally and with APIs. Its Vertex AI and Agentspace offerings are designed to make it dead-simple for customers to pick models from a massive library, connect to just about any data source and choose from a gallery of agents or roll their own. On top of that, Google's new Agent2Agent open protocol promises to improve agent interoperability, even if the agents are on different frameworks. And as I said during the event, the team deserves credit for its simplicity in communicating about AI agents. So please don't get me wrong: all of this agentic stuff is good. My reservation is that I'm still not convinced that I see any clear differences among any of the horizontal agents offered by Google, AWS or Microsoft. And it's still very early days for agentic AI. I suspect we'll see a lot more changes in this area in the coming year or two. I just haven't seen anything yet that I would describe as an agentic watershed for any of the big CSPs — or as exciting for Google Cloud as the bigger strategic positioning in AI that I'm describing here. At the event, Kurian said that companies work with Google Cloud because it has an open, multi-cloud platform that is fully optimized to help them implement AI. I think that its path forward reflects those strengths. I really like the idea of combining Cloud WAN plus Cloud Interconnect — plus running Gemini on-prem (on high-performing Dell infrastructure) as a managed service. In fact, this may be the embodiment of the true hybrid multicloud vision that I've been talking about for the past 10 years. Why is this so important today? Well, stop me if you've heard me say this before, but something like 70% to 80% of all enterprise data lives on-prem, and the vast majority of it isn't moving to the cloud anytime soon. It doesn't matter if you think it should or if I think it should or if every SaaS vendor in the world thinks it should. What does matter is that for reasons of control, perceived security risks, costs and so on . . . it's just not moving. Yet enterprises still need to activate all that data to get value out of it, and some of the biggest levers available to do that are generative AI and, more and more each day, agentic AI. Google Cloud is in a position to deliver this specific solution — in all its many permutations — for enterprise customers across many industries. It has the hardware, the software and the know-how, and under the direction of Thomas Kurian and his team, it has a track record for smart execution. That's no guarantee of more success against AWS, Microsoft, Oracle and others, but I'll be fascinated to see how it plays out.

AMD Powers Google Cloud's New AI Servers Promising 80% Boost In Speed
AMD Powers Google Cloud's New AI Servers Promising 80% Boost In Speed

Yahoo

time11-04-2025

  • Business
  • Yahoo

AMD Powers Google Cloud's New AI Servers Promising 80% Boost In Speed

Advanced Micro Devices, Inc (NASDAQ:AMD) on Wednesday announced the new Alphabet Inc (NASDAQ:GOOG) (NASDAQ:GOOGL) Google Cloud C4D and H4D virtual machines (VMs), which fifth-generation AMD EPYC processors power. The latest additions to Google Cloud's general-purpose and HPC-optimized VMs deliver leadership performance, scalability, and efficiency, from data analytics and web serving to high-performance computing (HPC) and AI. Based on Google Cloud's testing, leveraging the advancements of the AMD 'Zen 5' architecture allowed C4D to deliver up to 80% higher throughput/vCPU compared to previous generations. Nvidia's Custom GPU Enhances Nintendo Switch 2 With AI Features, Real-Time Ray Tracing, 4K Support AMD SVP Dan McNamara noted that its partnership with Google Cloud enabled it to rapidly adopt the latest AMD EPYC processors and deliver consistent high-performance and cost-efficient instances for its most demanding customers. Google Cloud VP Mark Lohmeyer noted that C4D and H4D instances powered by AMD EPYC processors can help businesses benefit from advanced performance and efficiency tailored to their cloud-native and enterprise applications. AMD stock plunged 52% in the last 12 months as it battled with Nvidia Corp (NASDAQ:NVDA) for market share. Wedbush analyst Daniel Ives cited AMD as among the vulnerable tech companies, including Apple Inc (NASDAQ:AAPL) and Nvidia, after China's finance ministry announced a 34% tariff on all goods imported from the U.S., effective April 10, in response to the U.S. administration's 54% duties imposed by the Trump-led administration. Jefferies analysts expect the tariffs to continue to put downward pressure on semiconductor stocks, including AMD. DA Davidson's Gil Luria highlighted that companies like Microsoft Corp (NASDAQ:MSFT) and Inc (NASDAQ:AMZN) have already begun to scale back their data center investments due to financial uncertainty. President Trump's sweeping new trade tariffs sparked a dramatic sell-off last Thursday. Wall Street witnessed a staggering $2 trillion wiped from market value, with the 10 largest U.S. companies alone accounting for $1 trillion of those losses. Price Action: AMD stock is up 20.10% at $93.96 at the last check on Wednesday. Read Next:Photo via Shutterstock Up Next: Transform your trading with Benzinga Edge's one-of-a-kind market trade ideas and tools. Click now to access unique insights that can set you ahead in today's competitive market. Get the latest stock analysis from Benzinga? ALPHABET (GOOGL): Free Stock Analysis Report ADVANCED MICRO DEVICES (AMD): Free Stock Analysis Report This article AMD Powers Google Cloud's New AI Servers Promising 80% Boost In Speed originally appeared on © 2025 Benzinga does not provide investment advice. All rights reserved. Sign in to access your portfolio

5th Gen AMD EPYC Processors Deliver Leadership Performance for Google Cloud C4D and H4D Virtual Machines
5th Gen AMD EPYC Processors Deliver Leadership Performance for Google Cloud C4D and H4D Virtual Machines

Associated Press

time09-04-2025

  • Business
  • Associated Press

5th Gen AMD EPYC Processors Deliver Leadership Performance for Google Cloud C4D and H4D Virtual Machines

— New instances provide enterprises with high-performance, scalable, and cost-effective cloud computing solutions — SANTA CLARA, Calif., April 09, 2025 (GLOBE NEWSWIRE) -- Today, AMD (NASDAQ: AMD) announced the new Google Cloud C4D and H4D virtual machines (VMs) are powered by 5th Gen AMD EPYC™ processors. The latest additions to Google Cloud's general-purpose and HPC-optimized VMs deliver leadership performance, scalability, and efficiency for demanding cloud workloads; for everything from data analytics and web serving to high-performance computing (HPC) and AI. Google Cloud C4D instances deliver impressive performance, efficiency, and consistency for general-purpose computing workloads and AI inference. Based on Google Cloud's testing, leveraging the advancements of the AMD 'Zen 5' architecture allowed C4D to deliver up to 80% higher throughput/vCPU compared to previous generations. H4D instances, optimized for HPC workloads, feature AMD EPYC CPUs with Cloud RDMA for efficient scaling of up to tens of thousands of cores. 'Since our launch, 5th Gen AMD EPYC solutions have been widely adopted across our OEM partners, enterprise customers, and now we're excited to bring it to the cloud,' said Dan McNamara, senior vice president and general manager, Server Business, AMD. 'Our deep technology partnership with Google Cloud enabled them to rapidly adopt the latest AMD EPYC processors to deliver consistent high performance and cost-efficient instances for their most demanding customers.' 'Google Cloud is committed to delivering high-performance, secure, and scalable compute solutions to our customers,' said Mark Lohmeyer, vice president and general manager, Compute and Machine Learning Infrastructure, Google Cloud. 'With the introduction of C4D and H4D instances powered by AMD EPYC processors, businesses can benefit from cutting-edge performance and efficiency, tailored to their cloud-native and enterprise applications.' Both C4D and H4D virtual machines are available in preview now, with general availability planned for later in the year across multiple global regions. Supporting Resources About AMD For more than 50 years AMD has driven innovation in high-performance computing, graphics, and visualization technologies. Billions of people, leading Fortune 500 businesses, and cutting-edge scientific research institutions around the world rely on AMD technology daily to improve how they live, work, and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit the AMD (NASDAQ: AMD) website, blog, LinkedIn, and Twitter pages. AMD, the AMD Arrow logo, EPYC and combinations thereof are trademarks of Advanced Micro Devices, Inc. Other names are for informational purposes only and may be trademarks of their respective owners.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store