
I made 25 videos using Google's Veo 3. Here's how it went.
Mishaal Rahman / Android Authority
When the AI-generated 'Will Smith eating spaghetti' video went viral a little over 2 years ago, I wasn't as skeptical as some about the future of AI video generation. I anticipated improvements, but I never imagined the technology would advance so quickly. Indeed, it was just last month that Google rolled out Veo 2, its second-generation AI video generator model, to the public, and the company is already back with a much more impressive model. After making over 25 videos with it, I'm convinced that Google's Veo 3 is a mind-blowing advancement in AI video generation, for better or worse.
What is Veo 3?
Veo 3 is Google's state-of-the-art text-to-video generation model. Like Veo 2, Veo 3 creates high-quality videos in a range of subjects and styles, even capturing nuanced object interactions and human expressions. Both models also block 'harmful requests and results' and mark their video outputs with an invisible watermark called SynthID.
The Veo 2 model could only produce silent videos, making it more like a high-quality GIF generator. The new Veo 3 model, however, supports native audio generation, putting it leaps and bounds ahead of its predecessor. The new model can not only generate sound effects and ambient noise but also create dialogue that's synced with the video.
Mishaal Rahman / Android Authority
Veo 3 vs Veo 2
While Veo 3 outputs are still limited to short, 8-second video clips, the addition of native audio generation has allowed people to create some truly mind-blowing AI videos that have taken the Internet by storm. I'm sure you've seen some of these videos already, but if not, we've put together a collection of over 25 videos made by Veo 3 that demonstrate the tool's prowess and its current limitations. While Veo 3 can be a pain to work with, its low barrier to entry makes it an incredible tool for anyone with enough time to create convincing, life-like videos, and I'm not convinced the world is ready for this.
Veo 3 almost makes it too easy to create realistic videos
If you've spent any amount of time on social media in the last few weeks, then you've probably seen people argue over whether 100 men could beat 1 gorilla in a fight. It's become somewhat of a meme, with laypeople and experts alike chiming in on the debate. Some amateur video makers have even created their own simulations of the hypothetical brawl. I wanted to see how easy it would be for me, someone with virtually no 3D animation experience, to make a video showing 100 men take on 1 gorilla.
It was as simple as asking the Gemini chatbot, 'Create a video showing 100 men fighting one silverback gorilla.'
Now, I'm sure if you pixel-peep, you can find some errors. Maybe you'll spot some men or weapons in the background appearing or disappearing randomly, or perhaps you'll notice that there clearly aren't 100 men in this 8-second clip. But if you were to simply watch this video on a small smartphone screen, you'd be hard-pressed to find any major issues at a glance.
This video definitely captures the chaotic, fast-paced action that would ensue when 100 men take on 1 gorilla. The sound that Veo 3 generated for the gorilla's punches had weight to it, making it feel believable. I knew it was an AI-generated video, of course, because I was the one who made it. But when I showed my mom — who was unaware of the memes on social media — this clip, she asked me what movie it was from!
Another video that demonstrates Veo 3's skills at simulating animal physics is this one:
I asked Gemini to make a video of 'a bull rampaging in a shop selling fine china,' and it created a video where, again, if you were to pixel-peep, you'd probably find issues. But at a glance, everything looks strikingly real: the bull's movement through the shop, the way dishes scatter, and the accompanying sound of them breaking. Most shops selling fine china would probably be more organized, but some out there might indeed look like this.
While I think the '100 men vs. 1 gorilla' video does a decent job showing how Veo 3 handles people, this next example better illustrates its ability to capture the nuances of human expressions. One of our readers asked us to create a video of 'a British Parliament debate between two men using the roadman accent,' and I was amazed at how it turned out.
Veo 3 generated some really realistic, subtle movements in the — need I remind you, AI-generated — hands and facial muscles of the man on the left as he says, 'Do you know what? Blud.' And the way the man on the right moves as he coughs in the other man's face felt incredibly real.
What I think Veo 3 does best is create realistic videos of utterly unreal situations. Sure, you probably won't ever see 100 men fight 1 gorilla, a bull rampaging through a fine china shop, or two British Parliamentarians arguing in a roadman accent in your lifetime, but they're all things that could conceivably happen.
You'll never see a real video depicting a 'hyper-magnified view of a bustling ant colony' where the ants are actually 'intricate clockwork robots' building mini skyscrapers from 'glowing sugar crystals,' only for this intricate scene to be disrupted by a human finger pressing down from above, culminating in a close-up of a spherical sugar crystal.
And you'll never see 'an asteroid crashing into an ocean of water balloons.'
Both of these videos demonstrate Veo 3's impressive skill at simulating physics interactions, even when dealing with such outlandish scenarios. To my amateur eyes, these look like they could have been meticulously created by an expert in 3D animation, but they weren't.
My favorite video to come from my testing is the one where the Google weather froggy mascot pops out of a Google Nest Hub to have a picnic on the kitchen counter. I just love how the frog pops out of the display, quickly lays out a picnic blanket, and waves at the camera. I also love how the camera pans out to show the kitchen, and how accurately Veo 3 rendered the frog's shadow and reflection on the kitchen counter.
The video isn't perfect, though, as it strangely doesn't have any audio. Also, the text on the Nest Hub says 'Nesst' instead of 'Nest,' showing that Veo 3 still has issues with rendering text accurately in videos. When I asked Veo 3 to redo the video but add sound, the result was much more unsettling: Instead of a cute frog popping out of the display, what looks like a person in a frog costume jumps over the display and starts a picnic on the counter before saying, 'Hello, everybody!'
This redo encapsulates some of the issues that Veo 3 still has with consistency and prompt adherence. In fact, this was just one of many examples of Veo 3 struggling to stick to my prompt, despite Google's stated improvements to prompt adherence. In one sense, I'm almost happy that Veo 3 isn't perfect; the model won't always give you what you're looking for, and that makes it harder for people with malicious intentions to abuse it.
Should we be happy that Veo 3 isn't perfect?
The videos I shared above, as well as the many great Veo 3-generated videos you may have already seen online, may have given you the impression that Veo 3 always produces excellent results. It can very easily do so, but it still struggles a lot with generating legible text and adhering to the prompt.
For example, when I asked it to create a video of 'a woman paragliding from the top of the Eiffel Tower,' it gave me a pretty realistic-looking video of a woman paragliding…next to the Eiffel Tower.
Or when someone asked me to make a video 'where someone types a prompt for Google Veo inside Google Veo,' the model churned out a video with unintelligible text on a laptop screen.
The most amusing issue I've encountered with Veo 3 is its inability to understand what a 'bugdroid' is. Bugdroid is the official nickname of the Android robot mascot, but Veo 3 consistently fails to accurately portray the robot in its generated videos, often creating generic robots with large eyes or bug-like antennas. It's not like the model refuses to generate videos featuring the Android mascot due to brand safety concerns; it can easily be made to do so if you tell it to make a 'green Android bot.'
Speaking of brand safety, it's nice to see that Veo 3 at least has some basic protections against generating videos featuring famous people. If you ask it to, for example, make a video featuring YouTube star Mr. Beast, it'll refuse to do so. If you try to work around this by describing the exact person instead of providing their name, Veo 3 will still refuse to generate the video.
With some really clever prompting, Veo 3 can definitely still be coaxed into creating videos featuring famous people, as we've already seen several Veo 3-made recreations of Will Smith eating spaghetti online. This is definitely problematic given how realistic Veo 3's videos can be. Even videos that don't feature celebrities can cause a stir, as demonstrated by a fake video of a woman being denied boarding due to her wanting to bring her 'emotional support kangaroo' onboard.
To combat this, Google has started to put a visible watermark on all Veo 3 videos generated through Gemini. However, there's an exception: this watermark isn't placed if you're a Google AI Ultra subscriber using Flow, Google's new AI-powered filmmaking tool, to generate videos. The Google AI Ultra plan costs an eye-watering $249.99 per month, which is pricey enough to deter some but not all persons with malicious intent.
With Flow, you can even guide Veo 3's output with your own images or AI-generated images. This allows for greater control over the output and enables generating videos that better align with your creative intent, capture your desired aesthetic, and match your characters' designs. It also opens up new avenues of video creation by allowing Veo 3 to generate expanded scenes, reimagine videos in different styles, remove unwanted objects from videos, and animate drawn characters.
Supplied by Google
Flow mitigates many of the inherent limitations of using Veo 3 in Gemini, and I can genuinely see it being useful for amateur and even professional filmmakers. But it also circumvents Google's new visible watermarking policy and only adds an invisible watermark that few platforms support. It also makes it easier to put together longer videos, which means it has a higher potential for abuse.
I don't think there's any turning back at this point; AI video generators are here to stay, and they're just going to keep getting better. Veo 3 is leaps and bounds better than the first text-to-video models, and it's only been a few years. With how much existing and new video data Google can pull from YouTube, the company will undoubtedly make major improvements in its upcoming Veo 4 model.
See all the videos we made with Veo 3
If you're curious, here's every single AI generated video we made with Veo 3 through Gemini: Create a video showing 100 men fighting one silverback gorilla. (Link) Can you repeat this, but make it like it was uploaded to Snapchat circa 2018, filmed on an iPhone? (Link)
Create a cinematic trailer for an imaginary sci-fi movie set on a distant planet with floating cities where the protagonist is secretly the son of the villain. (Link)
Create an animated video of garden gnomes constructing a futuristic AI supercomputer using CPUs that resemble carrots, potatoes, and broccoli. Show them working in a magical underground lab with glowing circuits and enchanted tools. (Link)
Make a video showing the Android bugdroid walking down a path by itself, looking at a smartphone it's holding. That bugdroid gets surprised by a couple of other bugdroids that are hanging out together and invite the first droid to join them. Each bugdroid should be wearing a hat that says 'Android Faithful'. (Link)
A man walking on water. (Link)
A green Android bot messing with a red Apple and finally eating it. (Link)
Make a video of 3 humanoid mechs fighting against an army of 20 red colored humanoid mechs via aeriel combat above the skies of Tokyo. All of the combatants are actively fighting and not taking turns or waiting around. The camera is slowly zooming outwards from the action throughout the entire scene. Behind a distant cloud, a silhouette of a flying lizard monster can be seen. (Link)
A bull rampaging in a shop selling fine china. (Link)
An influencer announces to the world, via a short form, vertical video, a 100 man vs 1 gorilla showdown while showing the contenders. (Link)
A woman paragliding from the top of the Eiffel Tower. (Link)
A surreal F1 race where iconic cars and drivers from different eras compete on a track that morphs through time and space. (Link)
A hyper-magnified view of a bustling ant colony, but instead of ants, tiny, intricate clockwork robots frantically build and dismantle miniature skyscrapers made of glowing sugar crystals. Suddenly, a colossal, slow-motion 'finger of doom' descends from above, casting a giant shadow, accompanied by a booming, distorted 'THUD' followed by the sound of crumbling glass and the squeaking, frantic whirring of the robots as everything collapses into sparkling dust. End on a close-up of a single, perfectly spherical sugar crystal landing silently. (Link)
Create a video of a basketball game between 5 bugdroids in yellow jerseys and 5 bugdroids in white jerseys where one of the yellow bugdroids dribbles the ball from his three point line to his free throw line then back to his three point line then turns around and shoots the ball to the basket. The ball bounces on the rim then hangs in the air then drops in. The bugdroids that shot the ball yells in delight while putting his hands around his throat and his other four teammates dressed in yellow run over to him to celebrate. Bugdroids, by the way, are the term for the Android OS robot mascot. (Link)
Create a video of a beach filled with people with everyone doing a unique beach activity while the camera pans to the left 180 degrees, then the camera pans right back to the original spot. In the foreground is a person with a surfboard walking left, and the person is eventually out of the frame, but will reappear once the camera pans over him again as it is returning to the first frame of the scene. (Link)
The froggy mascot popping out of a Google Nest Hub and having his picnic on a kitchen counter. He then greets the camera as it zooms out to reveal the kitchen. (Link)
Create a video of the Android Bot mascot holding a Google Pixel phone and texting his friend, a red Apple holding an iPhone. The perspective should be over the shoulder of the Android Bot, showing the Pixel phone's screen with the Google Messages app on screen. The message the Bot is sending should read, 'Thanks for getting the message!' As he hits send, the video should show the red Apple receiving the message on his iPhone, and the message should be shown in a green bubble in the Apple Messages app. (Link)
Make a video inspired by a Bollywood action movie scene that transitions to a musical dance sequence after a few seconds. Afterwards, the action sequence resumes from where it was suspended. All of the actors are of Indian descent and the lead actor is dressed in traditional Indian garbs. Alongside him is his sidekick who is dressed as a call center worker. (Link)
Create a scene where a black male dressed in Hawaiian clothing is running away from a T-Rex. The two are moving in the direction of the audience. This scene plays for 1.5 seconds and marks the end of segment A of the video. Segment B of the video now begins. The camera zooms in on the man's face as he looks to the left side of the screen. The scene pans to the left to reveal a close up shot of a white male dressed only in a beach towel and sunglasses. The new character lifts up his glasses slowly and shows a bewildered expression. This section marks segment B of the video and lasts for 4 seconds in total. Segment C of the video starts at this point and the camera now shifts back from the white male back to the black male running from the T-Rex. The camera zooms out and the T-Rex lets out a roar. In the background, a volcano can be seen erupting once the roaring starts. (Link)
A gamer who is live streaming themselves playing a tactical, turn-based, fantasy-themed Japanese RPG. The player is currently in a battle with a party of 4, one thief, one white mage, one paladin, and one black mage, fighting against a red dragon. The player can be seen in a small square overlay on the bottom right corner of the screen. (Link)
An asteroid crashing into an ocean of water balloons. (Link)
Show us a video of NASA astronauts finding life on mars from the perspective of the mission control room. (Link)
A video where someone types a prompt for Google Veo inside Google Veo. (Link)
Giants Causeway with the basalt columns rising and sinking into the sea in waves. (Link)
A war between unicellular and multicellular organisms. Video from the battlefield. (Link)
A British parliament debate between two men using the roadman accent. (Link)
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
29 minutes ago
- Yahoo
Analysts unveil bold forecast for Alphabet stock despite ChatGPT threat
Analysts unveil bold forecast for Alphabet stock despite ChatGPT threat originally appeared on TheStreet. You typed in a question and clicked a few links, and Google could get paid if you landed on an ad. For years, that simple cycle helped turn Google into a trillion-dollar titan. But now, that model is under threat. 💵💰💰💵 AI-powered chatbots like OpenAI's ChatGPT are rapidly changing how people find answers. Instead of browsing through links, users are getting direct summaries on AI. These 'zero-click' searches quietly erode the economics that built the modern internet. The number of users is growing fast. OpenAI CEO Sam Altman said in April that ChatGPT already has 'something like 10% of the world" in terms of users, pegging the number closer to 800 million, Forbes reported. Even Google seems to know it. It's giving AI answers, called AI Overviews, right at the top of the page. "What's changing is not that fewer people are searching the that more and more the answers to Google are being answered right on Google's page. That AI box at the top of Google is now absorbing that content that would have gone to the original content creators," Cloudflare CEO Matthew Prince said in a CNBC interview. Alphabet () , Google's parent company, isn't showing any cracks just yet. In April, the company posted first-quarter revenue of $90.23 billion, topping Wall Street expectations. Earnings per share came in at $2.81, far above the forecasted $ the backbone of Google's business, brought in $66.89 billion, accounting for nearly three-quarters of total revenue. Its 'Search and other' segment rose almost 10% year over year, hitting $50.7 billion. Meanwhile, Google's own AI tools are starting to show traction. AI Overviews now has 1.5 billion users per month, up from 1 billion in October, the company said. So far, the numbers suggest that AI isn't cannibalizing Google's business yet. Bank of America remains bullish on Alphabet stock. The firm reiterated a buy rating and a price target of $200, which implies a potential 15% upside from current levels, according to a recent research report. The firm said in May, Google's global average daily web visits held steady at 2.7 billion, unchanged from the previous month and down 2% from a year earlier. ChatGPT, meanwhile, saw a 3% month-over-month increase to 182 million, marking a 105% jump the U.S., Google traffic slipped 2% year-over-year to 524 million daily visits, while ChatGPT surged 112% over the same period to 26 million. Although Google has highlighted the growing reach of its AI Overviews, analysts are uncertain whether it's translating into more traffic. 'So far, we are not seeing a lift in Google traffic from AI Overviews expansion, though we think the search experience is much improved,' the analysts wrote. The competition is real. Google's global search share also edged down in May, falling 8 basis points month-over-month and 123 basis points year-over-year to 89.6%, according to Statcounter. Still, Bank of America analysts remain optimistic on Alphabet stock. "While ChatGPT's traffic continues to grow rapidly, we think Google remains well-positioned given its scale, multi-product reach, data assets, and robust monetization infrastructure," the analysts said. "AI can expand overall search monetization by better understanding the intent behind complex and long-tail queries that were previously hard to monetize," they added. Morningstar's Malik Ahmed Khan echoed that sentiment, saying Alphabet's diverse revenue streams and global exposure should cushion any hits, even as regulatory and AI risks mount, according to a May research report. Alphabet stock closed at $174.92 on June 6. The stock is down 8% unveil bold forecast for Alphabet stock despite ChatGPT threat first appeared on TheStreet on Jun 6, 2025 This story was originally reported by TheStreet on Jun 6, 2025, where it first appeared. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data


Forbes
an hour ago
- Forbes
Samsung Confirms Upgrade Choice—Galaxy Users Must Now Decide
This decision defines the future of your phone. Republished on June 7 with reports into Google's new decision for Android users. A timely warning from Samsung this week, which neatly sets out the biggest upgrade decision now facing Android users. As whispers start to spread suggesting a disconnect between Samsung and Google at the heart of Android, this is critical. We're talking AI and the new features and offerings now hitting phones and PCs at breakneck speed. This is where Galaxy has an advantage, Samsung says, 'in privacy-first, AI-powered experiences' which can 'protect you in the era of AI.' The question the Galaxy-maker asks in its latest post is the right one: 'This level of personalization' brought by AI 'can be incredibly helpful, but the more your phone knows, the more there is to protect. So, what's keeping all that personal data secure?' Samsung's answer is Knox. 'Every Galaxy device is protected from the chip up by a multi-layered approach, which includes on-device personalization, user-controlled cloud processing, and ecosystem-wide protection through Samsung Knox Matrix.' This is Samsung's secure ecosystem that is the closest replica to Apple's securely walled garden currently available on Android. 'At the core of this system is Samsung Knox Vault, Samsung's hardware-based solution for your most sensitive information.' Knox is not new and neither is the concept of hardware-enabled Galaxy data security. What is new is segmenting sensitive the latest AI-related data from the rest, and securing that alongside the more traditional PINs, passwords and credit card numbers. 'Location service metadata from your most personal photos,' Samsung says, 'could easily give away the exact location where the image was taken.' And there's not much data more sensitive than who did what, where and when. 'In the era of AI, personal information like your home address, face clustering ID, person ID, pet type, scene type and more need to be encrypted and stored in a safe location. These things aren't just files — they are deeply connected to your daily life.' It's unclear exactly what is being or will be segmented and how this plays into the various opt-ins that Samsung has added to distinguish between on-device and cloud AI, between what is only within your secure enclave and what is outside. But it's difficult not to read this push as a play against the latest announcements from Google and the cloud-based AI that will now run riot across sensitive data, including emails and even cloud data storage. Yes, there are always opt-outs, but it's all or nothing for users who want AI but are not yet worrying about privacy. 'As Galaxy AI becomes more useful,' Samsung says, 'it also becomes more personal — learning how you use your device and adapting to your needs… Knox Vault is more than a security feature, it's Galaxy's promise that no matter how advanced your devices become, or how much AI evolves, your privacy is secured.' Google, meanwhile, will not make this decision easy for Samsung user. No one is rolling out new smartphone AI innovations faster, and it will always overshadow what can be done if users take a privacy-centric, device-only approach. Per Android Police, the latest update is 'Google's Gemini replacing Google Assistant as the default AI assistant, taking on all digital assistance responsibilities as Assistant is phased out later this year. Gemini is gaining 'Scheduled Actions,' allowing users to automate recurring tasks and information delivery at specific times.' This is the stepping stone to so-called Agenctic AI on phones, where monitoring data and events and activities enables an agent to make decisions autonomously on a smartphone owner's behalf. This next step, with 'Scheduled Actions streamlining routines [and] offering personalized updates,' is just the start. As Mashable says, 'When combined with computer vision, which is what allows a model to 'see' a user's screen, we get the agentic AI everyone is so excited about… Agentic AI tools could order groceries online, browse and buy the best-reviewed espresso machine for you, or even research and book vacations. In fact, Google is already taking steps in this direction with its new AI shopping experience.' Allowing AI access to smartphones with all the data and insight they contain, pushed this to a level even beyond Windows's controversial Recall. It's decision time.


Tom's Guide
an hour ago
- Tom's Guide
5 features iOS 26 needs to steal from Google to catch up on AI
I've been enjoying Google's AI features on my Pixel phones for the last couple of years. Starting with the Pixel 8 Pro and proceeding with the Pixel 9 Pro, Google has proven to me that its AI features in its Pixel phones are unmatched — and Apple's in trouble if it doesn't catch up. With WWDC 2025 right around the corner, it's Apple's chance to redeem itself by introducing more Apple Intelligence features for what's presumably going to be the next iteration of its phone software: iOS 26. While there's been a handful of useful AI features, such as Visual Intelligence and Photo Clean Up to name a few, iPhones could still stand to get more. In fact, there are a number of Google AI features I think Apple needs to copy that could boost the iPhone experience. I'm not saying outright steal the same exact features, but at least come up with something similar — or if not, better one. If there's one AI feature that Apple desperately needs to copy from Pixel phones, it has to be none other than Call Screen. Not only is it one of the most underrated AI features I've tried in any phone, but it's also one of the most helpful. Call Screen allows Pixel phones to take incoming calls on your behalf, using Google Assistant to listen to callers and then provide you with contextual responses on your phone to choose. Think of it like an actual assistant who's fielding the call for you and relaying your response. I can't tell you how many times it's been such a lifesaver when I'm stuck in a work meeting. Although it technically debuted with the Galaxy S25 Ultra, the cross-app actions function migrated to Pixel phones and it shows the impressive abilities of AI. While Apple Intelligence can call on Siri to perform simple actions, it doesn't have the ability to connect with third-party apps — which is exactly what makes cross-app actions such a big game changer with Pixel phones. Get instant access to breaking news, the hottest reviews, great deals and helpful tips. Through simple voice commands, it can work with several apps to complete a certain request. For example, you can ask Gemini on a Pixel phone to summarize an email or find a nearby restaurant that's pet friendly and add a calendar appointment for it. Another feature that debuted with Samsung and eventually made its way to Pixel phones is Circle to Search. Apple currently doesn't have anything like it, although you could argue that Visual Intelligence can effectively function in almost the same way. With Circle to Search, it's a quick and convenient way to perform searches directly on-device, from whatever app you're using. When activated, you simply circle or select what you're looking at on your phone's screen to perform a search — which could result in answering a question, performing a general Google Search, identifying something, and even finding deals on a product. One AI feature I've come to appreciate as a photo editor is the Pixel's Reimagine tool, which allows me to select parts of a photo and transform it into something else through a text description. The closest Apple Intelligence feature to this would be Image Playground, but that generates images from scratch through a text description — it doesn't work with existing photos. Reimagine helps to make existing photos look better, whether it's to change up the scene entirely or make minor edits. I personally love being able to select the sky in my photos and change it up to something else, or using Reimagine to insert different elements with realism. Even though it could benefit from a few enhancements, Pixel Screenshots can be better at helping you recall information you might forget — or need to remember for later on. It's exclusively available on the Pixel 9, Pixel 9 Pro, Pixel 9 Pro XL, and Pixel 9 Pro Fold and lets you use the screenshot function and AI to recall details in them. For example, if you screenshot a pizza recipe you want to try for later, or the details about an upcoming party you're going to, Pixel Screenshots will allow you to perform a search to find the exact details about it. Apple doesn't have a comparable AI feature, but wouldn't it be neat if Apple Intelligence could recall the most obscure (or detailed) information that you go through on your iPhone.