logo
#

Latest news with #Next25

Google Cloud Gets More Serious About Infrastructure At Next 2025
Google Cloud Gets More Serious About Infrastructure At Next 2025

Forbes

time24-04-2025

  • Business
  • Forbes

Google Cloud Gets More Serious About Infrastructure At Next 2025

Google Cloud was especially bold in its competitive positioning against AWS at the Google Cloud Next ... More 2025 conference. Here, Mark Lohmeyer, vice president and general manager of AI and computing infrastructure at Google Cloud, presents head-to-head comparisons. This month's Google Cloud Next 2025 event was an excellent reference point for how far Google Cloud has come since CEO Thomas Kurian took the helm of the business at the start of 2019. Back then, Google Cloud had about $6 billion in revenue and was losing a ton of money; six years later, it's nearing a $50 billion annual run rate, and it's profitable. I remember that when Kurian started, early odds were that Google would get out of the cloud service business altogether — yet here we are. Typically for this conference, there was so much announced that I can't cover it all here. (Among the many progress stats that Kurian cited onstage: the business shipped more than 3,000 product advances in 2024.) For deeper dives into specific areas, see the articles from my colleagues Matt Kimball on the new Ironwood TPU chip, Jason Andersen on Google's approach to selling enterprise AI (especially agents) and Melody Brue on the company's approach to the connected future of AI in the workplace. Our colleague Robert Kramer also wrote an excellent preview of the event that still makes good background reading. What I want to focus on here are Next 25's most interesting developments in connectivity, infrastructure and AI. (Note: Google is an advisory client of my firm, Moor Insights & Strategy.) Kurian placed a strong focus on connectivity, specifically with the company's new Cloud WAN and Cloud Interconnect offerings. Cloud WAN makes the most of Google's network, which the company rightly calls 'planet-scale,' to deliver faster performance than the public internet (40% faster, according to the company) that's also significantly cheaper than enterprise WANs (with a claimed 40% lower TCO). Meanwhile, Cloud Interconnect is built to connect your own enterprise network to Google's — or even to your network hosted by a different CSP — with high availability and low latency. Interestingly, in the analyst readout at the conference, Kurian started off with networking, which highlights its importance to Google. This makes sense, as enterprises are all bought into the hybrid multicloud and the growing need to connect all those datacenters, whether public or private cloud. This went hand in hand with a lot of discussion about new infrastructure. For context, all of the hyperscalers have announced extra-large capex investments in infrastructure for this year, with Google weighing it at $75 billion. The presentations at Next 25 showed where a good chunk of that money is going. I'll talk more below about the infrastructure investments specific to AI, starting with the Ironwood TPU chip and AI Hypercomputer. For now I want to note that the infrastructure plays also include networking offload, new storage options, a new CPU . . . It's a long list, all aimed at supporting Google Cloud's strategy of combining hardware and software to enable bigger outputs — especially in AI — at a low price. Make special note of that low price element, which is unusual for Google. I'll come back to that in a minute. Strategically, I think that Google is recognizing that infrastructure as a service is an onramp to PaaS and SaaS services revenue. If you can get people signed on for your IaaS — because, say, you have competitive compute and storage and a planet-scale network that you're allowing them to piggyback on — that opens the door for using a bigger selection of your offerings at the platform level. And while we're at it, why not a PaaS or SaaS approach to handling a bigger slice of your enterprise AI needs? It's a solid move from Google, and I'm intrigued to see how it plays out competitively, especially given that Azure seemed to get serious about IaaS in the past couple of years. It's also notable that Next 25 is the first time I can remember Google Cloud going after AWS on the infrastructure front. As shown in the image accompanying this article, Google touts its Arm-based Axion CPU as outperforming the competing Arm-based processor from AWS, Graviton. In the Mark Lohmeyer breakout session, there was a lot of specific discussion of AWS Trainium chips, too. I'm a fan of stiff competition, so it's refreshing to see Google getting more aggressive with this. It's about time. Considering all the years I spent in the semiconductor industry, it's no surprise that my ears perked up at the announcement of Google's seventh-generation Ironwood tensor processing unit, which comes out later this year. (I wish Google had been more specific about when we can expect it, but so far it's just 'later in 2025.') Google was a pioneer in this area, and this TPU is miles ahead of its predecessors in performance, energy efficiency, interconnect and so on. My colleague Matt Kimball has analyzed Ironwood in detail, so I won't repeat his work here. I will note briefly that Google's Pathways machine-learning runtime can manage distributed workloads across thousands of TPUs, and that Ironwood comes in scale-up pods of 256 chips or 9,216 chips. It also natively supports the vLLM library for inference. vLLM is an accepted abstraction layer that enterprises can comfortably code to for their optionality, and it should allow users to run inference on Ironwood with an appealing price-to-performance profile — yet another instance of combining hardware and software to enable more output at a manageable price. Next 25 was also the enterprise coming-out party for the Gemini 2.5 model, which as I write this is the best AI model in the world according to Hugging Face's Chatbot Arena LLM Leaderboard. The event showcased some impressive visual physics simulations using the model. (Google also put together a modification of The Wizard of Oz for display on the inner surface of The Sphere in Las Vegas. I can be pretty jaded about that kind of thing, but in this case I was genuinely impressed.) I haven't been a big consumer of Google's generative AI products in the past, even though I am a paying customer for Workspace and Gemini. But based on what I saw at the event and what I'm hearing from people in my network about Gemini 2.5, I'm going to give it another try. For now, let's focus on what Google claims for the Gemini 2.0 Flash model, which allows control over how much the model reasons to balance performance and cost. In fact, Google says that Gemini 2.0 Flash achieves intelligence per dollar that's 24x better than GPT-4o and 5x better than DeepSeek-R1. Again, I want to emphasize how unusual the 'per dollar' part is for Google messaging. Assuming the comparison figures are accurate, Google Cloud is able to achieve this by running its own (very smart) models on its new AI Hypercomputer system, which benefits from tailored hardware (including TPUs), software and machine learning frameworks. AI Hypercomputer is designed to allow easy adaptation of hardware so it can make the most of new advances in chips. On a related note, Google says that it will be one of the first adopters of Nvidia's GB200 GPUs. At the keynote, there was also a video of Nvidia CEO Jensen Huang in which he praised the partnership between the two companies and said, 'No company is better at every single layer of computing than Google.' In my view, Google is doing a neat balancing act to reassure the market that it loves Nvidia — while also creating its own wares to deliver better price per outcome. Touting itself for delivering the best intelligence at the lowest cost was not something I expected from Google Cloud. But as I reflect on it, it makes sense. Huang has a point: even though it's a fairly distant third place in the CSP market, Google really is good at every layer of the computing stack. It has the homegrown chips. The performance of its homegrown AI models is outstanding. It understands the (open) software needed to deliver AI for enterprise uses. And it's only getting stronger in infrastructure, as Next 25 emphasized. Now it wants to take this a step further by using Google Distributed Cloud to bring all of that goodness on-premises. Imagine running high-performing Gemini models, Agentspace and so on in your own air-gapped environment to support your enterprise tools and needs. In comparison to this, I thought that the announcements at Next 25 about AI agents were perfectly nice, but not any kind of strategic change or differentiator for the company — at least not yet. To be sure, Google is building out its agent capabilities both internally and with APIs. Its Vertex AI and Agentspace offerings are designed to make it dead-simple for customers to pick models from a massive library, connect to just about any data source and choose from a gallery of agents or roll their own. On top of that, Google's new Agent2Agent open protocol promises to improve agent interoperability, even if the agents are on different frameworks. And as I said during the event, the team deserves credit for its simplicity in communicating about AI agents. So please don't get me wrong: all of this agentic stuff is good. My reservation is that I'm still not convinced that I see any clear differences among any of the horizontal agents offered by Google, AWS or Microsoft. And it's still very early days for agentic AI. I suspect we'll see a lot more changes in this area in the coming year or two. I just haven't seen anything yet that I would describe as an agentic watershed for any of the big CSPs — or as exciting for Google Cloud as the bigger strategic positioning in AI that I'm describing here. At the event, Kurian said that companies work with Google Cloud because it has an open, multi-cloud platform that is fully optimized to help them implement AI. I think that its path forward reflects those strengths. I really like the idea of combining Cloud WAN plus Cloud Interconnect — plus running Gemini on-prem (on high-performing Dell infrastructure) as a managed service. In fact, this may be the embodiment of the true hybrid multicloud vision that I've been talking about for the past 10 years. Why is this so important today? Well, stop me if you've heard me say this before, but something like 70% to 80% of all enterprise data lives on-prem, and the vast majority of it isn't moving to the cloud anytime soon. It doesn't matter if you think it should or if I think it should or if every SaaS vendor in the world thinks it should. What does matter is that for reasons of control, perceived security risks, costs and so on . . . it's just not moving. Yet enterprises still need to activate all that data to get value out of it, and some of the biggest levers available to do that are generative AI and, more and more each day, agentic AI. Google Cloud is in a position to deliver this specific solution — in all its many permutations — for enterprise customers across many industries. It has the hardware, the software and the know-how, and under the direction of Thomas Kurian and his team, it has a track record for smart execution. That's no guarantee of more success against AWS, Microsoft, Oracle and others, but I'll be fascinated to see how it plays out.

Is There Value In A Curated Enterprise AI Experience?
Is There Value In A Curated Enterprise AI Experience?

Forbes

time22-04-2025

  • Business
  • Forbes

Is There Value In A Curated Enterprise AI Experience?

Google Agentspace Ideation Agent Google Over the past couple of years, the frantic pace of AI innovation has had the big three cloud players vying to keep up with each other when it comes to AI capabilities. At this point, it's pretty easy to say that if one cloud vendor has a new AI capability, your preferred vendor either already has it or will have it within the next few weeks. One would think that AI presents new opportunities for a vendor to differentiate itself and take share from its competitors. But what's really happening is more a game of defense where the race for parity is about keeping existing customers. Heading into the Google Cloud Next 25 conference a couple of weeks ago, I was interested to see how Google would differentiate its AI offerings. I had some hope based upon its recent announcements about Agentspace (which I covered here) and the Customer Experience Suite (which I covered here). Both of those were notable in that the messaging was less about technology and more about creating business value and changing how people work. There were two key takeaways from the event. First, market execution is as important as technology in an ultracompetitive market like AI. Second, Google is using some proven but uncommon methods to differentiate its innovations. (Note: Google is an advisory client of my firm, Moor Insights & Strategy.) Google Cloud has held the number-3 revenue position in cloud services for a while now, but over the past year it's been taking market share from its competitors. It's even feasible that Google Cloud could eventually take the number-2 position away from Microsoft Azure. Certainly some of that has to do with the technology, but I think that hiring a new go-to-market leadership team and investing more in training and certifications — a trusted blueprint borrowed from many other established tech players — also has a lot to do with it. The timing for these initiatives is good, given the business-value-driven product messaging. Google has also made some smart decisions when it comes to building customer confidence, such as 30-day commits on spot pricing, along with making investments in customer education and services. Google is proving that how well you educate and take care of your customers is a major ingredient for tech sales and retention in times of disruption. This may not have always been the case for Google Cloud, but at Next 25, I became convinced that my perception of the company needed to be updated. In terms of technology, Google, like all other AI vendors, delivered a full plate of new innovations at its marquee event. And there were a lot of aha moments, including the reverse engineering and re-release of The Wizard of Oz. But it wasn't the technology itself that was amazing; it was how the technology is designed, how users are engaged and how solutions are deployed. That's not to say the technology isn't good. I would suggest that Google's core AI tech is competitive but only marginally differentiated. Rather, what I saw from Google was a deeper degree of business thinking and user-centric design than its AI competition. This is something I would call a curated experience. I believe this is deliberate, because a curated experience is a critical complement to Google's investments in improved market execution — not to mention how Google can gain new customers while retaining the ones it already has. To break this down a bit further, let's consider three big developer-related announcements. Agentspace represented a well-thought-out user-centric design. Google has had some success when it comes to AI user experience with other technologies such as NotebookLM. But Agentspace is a new type of work interface. For starters, it's personalized based upon the user's individual profile or other contextual inputs. For example, maybe you are running agents for a specific task. In that case, Agentspace will have the ability to present other relevant agents and downplay those that are irrelevant. Also, the UI looks more like a consumer product than something you'd typically see in an enterprise. The Agentspace product management team shared that this was a deliberate choice, and that they collaborated with Google's consumer UI teams to do it. The rationale was that AI adopters tend to get their initial introduction to AI from consumer-oriented projects. Therefore, give the user something they can learn more naturally — based upon experience rather than technical standards. To help drive further user adoption and engagement, Google also announced the Agent Developer Kit, which is an open source set of methods to foster collaboration between agents and other remote services. This makes a lot of sense because Google has leveraged open source to great effect in the past. The most notable example of this is Kubernetes, which is now the de facto standard in container management. Google's biggest contribution in the ADK was the Agent2Agent protocol, which is provided only with an AI platform like Google's own Vertex or Salesforce's Agentforce. The ADK also supports the emerging MCP standard. By open-sourcing ADK, Google will be able to attract developers to code collaborative agents without a lot of extra software and cost. (Why am I so sure? Look at the whole history of open source adoption in the enterprise.) It's a great way to get people to try out agents and, if they like it, to then consider Google's higher-end agent capabilities in Vertex and other products. Finally, in terms of solution deployment, Google's roadmap and packaging is quite clear. For example, in tooling there's Vertex and Firebase for professional developers and Agentspace for the no-code development environment. Another example is Model Garden, which, with 200 models available, is not too constrained but also avoids the chaos of more than one million models on Hugging Face. Google's simplicity here is quite refreshing, especially compared to the other cloud providers, which have more complex and entitlement-driven solutions. A curated experience is not a new idea. In fact, Red Hat Enterprise Linux was the poster child for the whole concept 20 years ago. By providing enterprises with gated access to the best of standards and open source and a world-class support team, Red Hat gave customers a feeling that they were tapping into both meaningful innovation and sensible risk mitigation. But we also have to be honest that RHEL was a lot cheaper than the Unix systems it was initially displacing. Eventually, as cloud came to the fore, open Linux also became more mainstream, and new competitors to Red Hat emerged (including cloud-specific variants from AWS and Google). Fortunately for Red Hat, it was able to transfer the RHEL thought process to other areas like virtualization and DevOps. (In this context, it's worth noting that Kubernetes is core to Red Hat's OpenShift cloud management platform.) Based upon this example, one can assume that curation can be a meaningful value proposition — but it may not be sustainable. Given the level of fragmentation, low standardization and user confusion about AI in the marketplace today, now is a good time for users to consider a curated experience. But how long it will last remains an open question. Additionally, Google's timeline may be different than Red Hat's, because Google is also tapping into something of a different positioning that may achieve a better result. Instead of providing 'leading-edge but not bleeding-edge' AI, the company is conveying a sense of 'Let's get going.' And for many mainstream companies and users, that may be the right type of encouragement to choose Google — particularly given how early we are in the age of AI. So, is Google's new TPU 10x faster than two years ago? Is Gemini better at a given benchmark this week? Yes to both — and those tech milestones certainly have their importance. But Google's real bet is that your experience in learning about and using AI is more important than those types of headlines.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store