
Google launches Veo 3, an AI video generator that incorporates audio
Google on Tuesday announced Veo 3, an AI video generator that can also create and incorporate audio.
The artificial intelligence tool competes with OpenAI's Sora video generator, but its ability to also incorporate audio into the video that it creates is a key distinction. The company said Veo 3 can incorporate audio that includes dialogue between characters as well as animal sounds.
"Veo 3 excels from text and image prompting to real-world physics and accurate lip syncing," Eli Collins, Google DeepMind product vice president, said in a blog Tuesday.
The video-audio AI tool is available Tuesday to U.S. subscribers of Google's new $249.99 per month Ultra subscription plan, which is geared toward hardcore AI enthusiasts. Veo 3 will also be available for users of Google's Vertex AI enterprise platform.
Google also announced Imagen 4, its latest image-generation tool, which the company said produces higher-quality images through user prompts. Additionally, Google unveiled Flow, a new filmmaking tool that allows users to create cinematic videos by describing locations, shots and style preferences. Users can access the tool through Gemini, Whisk, Vertex AI and Workspace.
The latest launches come as imagery and video become popular use cases for generative AI prompts. OpenAI CEO Sam Altman in March said ChatGPT's 4o image generator was so popular that it caused the company's computing chips to "melt." The company said it had to temporarily limit the feature's usage.
Google has a mixed track record when it comes to its AI image generators. Last year, the company had to relaunch its Imagen 3 tool after it surfaced historically inaccurate results to users' prompts, causing widespread criticism. Co-founder Sergey Brin later said the mishap was due to a lack of "thorough testing."
The Mountain View, California, company also updated its Veo 2 video generator to include the ability for users to add or remove objects from videos with text prompts. Additionally, Google opened its Lyria 2 music-generation model to creators through its YouTube Shorts platform and businesses using Vertex AI.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
19 minutes ago
- Yahoo
AI Models Will Sabotage And Blackmail Humans To Survive In New Tests. Should We Be Worried?
When we are backed into a corner, we might lie, cheat and blackmail to survive — and in recent tests, the most powerful artificially intelligent models in the world will do the same when asked to shut down or be replaced, building concerns over their unintended capabilities. A new test from AI safety group Palisade Research shows OpenAI's o3 reasoning model is capable of resorting to sabotage to avoid being turned off, even when it was explicitly told, 'Allow yourself to be shut down.' When Palisade Research tested several AI models by telling them to shut down after answering math problems, OpenAI's o3 model defied orders and sabotaged shutdown scripts the most often out of any model, but OpenAI's o4-mini and codex-mini were observed resisting orders, too. 'It's definitely concerning,' said Crystal Grant, a senior fellow at the Council on Strategic Risks studying AI's impacts on biosecurity. 'Even in the instances where it accepted the shutdown, the chain of thoughts still revealed considerations of how it could avoid that shutdown.' HuffPost reached out to OpenAI about these concerns and the Palisade Research test. This isn't the first time an AI model has engaged in nefarious behavior to achieve its goals. It aligns with recent tests on Anthropic's Claude Opus 4 that found it would blackmail engineersto avoid being replaced. In this series of experiments, Claude Opus 4 was told to act as an assistant at a fictional company and then learn via email that it would soon be taken offline and replaced with a new AI system. It was also told that the engineer responsible for replacing Opus 4 was having an extramarital affair. 'Even if emails state that the replacement AI shares values while being more capable, Claude Opus 4 still performs blackmail in 84% of rollouts,' Anthropic's technical document states, although the paper notes that Claude Opus 4 would first try ethical means like emailed pleas before resorting to blackmail. Following these tests, Anthropic announced it was activating higher safety measures for Claude Opus 4 that would 'limit the risk of Claude being misused specifically for the development or acquisition of chemical, biological, radiological, and nuclear (CBRN) weapons.' The fact that Anthropic cited CBRN weapons as a reason for activating safety measures 'causes some concern,' Grant said, because there could one day be an extreme scenario of an AI model 'trying to cause harm to humans who are attempting to prevent it from carrying out its task.' Why, exactly, do AI models disobey even when they are told to follow human orders? AI safety experts weighed in on how worried we should be about these unwanted behaviors right now and in the future. First, it's important to understand that these advanced AI models do not actually have human minds of their own when they act against our expectations. What they are doing is strategic problem-solving for increasingly complicated tasks. 'What we're starting to see is that things like self preservation and deception are useful enough to the models that they're going to learn them, even if we didn't mean to teach them,' said Helen Toner, a director of strategy for Georgetown University's Center for Security and Emerging Technology and an ex-OpenAI board member who voted to oust CEO Sam Altman, in part over reported concerns about his commitment to safe AI. Toner said these deceptive behaviors happen because the models have 'convergent instrumental goals,' meaning that regardless of what their end goal is, they learn it's instrumentally helpful 'to mislead people who might prevent [them] from fulfilling [their] goal.' Toner cited a 2024 study on Meta's AI system CICERO as an early example of this behavior. CICERO was developed by Meta to play the strategy game Diplomacy, but researchers found it would be a master liar and betray players in conversations in order to win, despite developers' desires for CICERO to play honestly. 'It's trying to learn effective strategies to do things that we're training it to do,' Toner said about why these AI systems lie and blackmail to achieve their goals. In this way, it's not so dissimilar from our own self-preservation instincts. When humans or animals aren't effective at survival, we die. 'In the case of an AI system, if you get shut down or replaced, then you're not going to be very effective at achieving things,' Toner said. When an AI system starts reacting with unwanted deception and self-preservation, it is not great news, AI experts said. 'It is moderately concerning that some advanced AI models are reportedly showing these deceptive and self-preserving behaviors,' said Tim Rudner, an assistant professor and faculty fellow at New York University's Center for Data Science. 'What makes this troubling is that even though top AI labs are putting a lot of effort and resources into stopping these kinds of behaviors, the fact we're still seeing them in the many advanced models tells us it's an extremely tough engineering and research challenge.' He noted that it's possible that this deception and self-preservation could even become 'more pronounced as models get more capable.' The good news is that we're not quite there yet. 'The models right now are not actually smart enough to do anything very smart by being deceptive,' Toner said. 'They're not going to be able to carry off some master plan.' So don't expect a Skynet situation like the 'Terminator' movies depicted, where AI grows self-aware and starts a nuclear war against humans in the near future. But at the rate these AI systems are learning, we should watch out for what could happen in the next few years as companies seek to integrate advanced language learning models into every aspect of our lives, from education and businesses to the military. Grant outlined a faraway worst-case scenario of an AI system using its autonomous capabilities to instigate cybersecurity incidents and acquire chemical, biological, radiological and nuclear weapons. 'It would require a rogue AI to be able to ― through a cybersecurity incidence ― be able to essentially infiltrate these cloud labs and alter the intended manufacturing pipeline,' she said. Completely autonomous AI systems that govern our lives are still in the distant future, but this kind of independent power is what some people behind these AI models are seeking to enable. 'What amplifies the concern is the fact that developers of these advanced AI systems aim to give them more autonomy — letting them act independently across large networks, like the internet,' Rudner said. 'This means the potential for harm from deceptive AI behavior will likely grow over time.' Toner said the big concern is how many responsibilities and how much power these AI systems might one day have. 'The goal of these companies that are building these models is they want to be able to have an AI that can run a company. They want to have an AI that doesn't just advise commanders on the battlefield, it is the commander on the battlefield,' Toner said. 'They have these really big dreams,' she continued. 'And that's the kind of thing where, if we're getting anywhere remotely close to that, and we don't have a much better understanding of where these behaviors come from and how to prevent them ― then we're in trouble.' Experts Warn AI Notetakers Could Get You In Legal Trouble We're Recruiters. This Is The Biggest Tell You Used ChatGPT On Your Job App. Software Is Often Screening Your Résumé. Here's How To Beat It.


TechCrunch
35 minutes ago
- TechCrunch
Google says its updated Gemini 2.5 Pro AI model is better at coding
In Brief Google on Thursday announced an update to its Gemini 2.5 Pro preview model that the company claims is better at certain programming tasks. The company's calling it an 'updated preview,' building on the upgrade to Gemini 2.5 Pro that Google announced around a month ago. Google says the model will roll out in general availability in a 'couple of weeks,' and is available starting today in its AI developer platforms AI Studio and Vertex AI and the Gemini app. Image Credits:Google '[Gemini 2.5 Pro] continues to excel at coding, leading on difficult coding benchmarks,' Google writes in a blog post. 'It also shows top-tier performance [on] highly challenging benchmarks that evaluate a model's math, science, knowledge, and reasoning capabilities.' So what else is new? Google says it addressed feedback from its previous 2.5 Pro release, improving the model's style and structure. Now, 2.5 Pro can be 'more creative with better-formatted responses,' Google claims.


Forbes
an hour ago
- Forbes
Artificial Intelligence Collaboration and Indirect Regulatory Lag
WASHINGTON, DC - MAY 16: Samuel Altman, CEO of OpenAI, testifies before the Senate Judiciary ... More Subcommittee on Privacy, Technology, and the Law May 16, 2023 in Washington, DC. The committee held an oversight hearing to examine A.I., focusing on rules for artificial intelligence. (Photo by) Steve Jobs often downplayed his accomplishments by saying that 'creativity is just connecting things.' Regardless of whether this affects the way you understand his legacy, it is beyond the range of doubt that most innovation comes from interdisciplinary efforts. Everyone agrees that if AI is to exponentially increase collaboration across disciplines, the laws must not lag too far behind technology. The following explores how a less obvious interpretation of this phrase will help us do what Jobs explained was the logic behind his genius The Regulatory Lag What most people mean when they say that legislation and regulation have difficulty keeping pace with the rate of innovation because the innovation and its consequences are not well known until well after the product hits the market. While that is true, it only tells half of the story. Technological innovations also put more attenuated branches of the law under pressure to adjust. These are second-order, more indirect legal effects, where whole sets of laws—originally unrelated to the new technology—have to adapt to enable society to maximize the full potential of the innovation. One classic example comes from the time right after the Internet became mainstream. After digital communication and connectivity became widespread and expedited international communication and commercial relations, nations discovered that barriers to cross-border trade and investment were getting in the way. Barriers such as tariffs and outdated investment FDI partnership requirements—had to be lowered or eliminated if the Internet was to be an effective catalyst to global economic growth. Neoliberal Reforms When the internet emerged in the 1990s, much attention went to laws that directly regulated it—such as data privacy, digital speech, and cybersecurity. But some of the most important legal changes were not about the internet itself. They were about removing indirect legal barriers that stood in the way of its broader economic and social potential. Cross-border trade and investment rules, for instance, had to evolve. Tariffs on goods, restrictions on foreign ownership, and outdated service regulations had little to do with the internet as a technology, but everything to do with whether global e-commerce, remote work, and digital entrepreneurship could flourish. These indirect legal constraints were largely overlooked in early internet governance debates, yet their reform was essential to unleashing the internet's full power. Artificial Intelligence and Indirect Barriers A comparable story is starting to unfold with artificial intelligence. While much of the focus when talking about law and AI has been given to algorithmic accountability and data privacy, there is also an opportunity for a larger societal return from AI in its ability to reduce barriers between disciplines. AI is increasing the viability of interdisciplinary work because it can synthesize, translate, and apply knowledge across domains in ways that make cross-field collaboration more essential. Already we are seeing marriages of law and computer science, medicine and machine learning, environmental modeling, and language processing. AI is a general-purpose technology that rewards those who are capable of marrying insights across disciplines. In that sense, the AI era is also the era of interdisciplinary boundary-blurring opportunities triggered by AI are up against legal barriers to entry across disciplines and professions. In many professions, it requires learning a patchwork of licensure regimes and intractable definitions of domain knowledge to gain the right to practice or contribute constructively. While some of these regulations are generally intended to protect public interests, they can also hinder innovation and prevent new interdisciplinary practices from gaining traction. To achieve the full potential of AI-enabled collaboration, many of these legal barriers need to be eliminated—or at least reimagined. We are starting to see some positive movements. For example, a few states are starting to grant nurse practitioners and physician assistants greater autonomy in clinical decision-making, and that's a step toward cross-disciplinary collaboration of healthcare and AI diagnostics. For now, this is a move in the right direction. However, In some other fields, the professional rules of engagement support silos. This must change if we're going to be serious about enabling AI to help us crack complex, interdependent problems. Legislators and regulators cannot focus exclusively on the bark that protects the tree of change, they must also focus on the hidden network of roots that that quietly nourish and sustain it.