SkyReels Open Sources the World's First Human-Centric Video Foundation Model for AI Short Drama Creation – SkyReels-V1, Reshaping the AI Short Drama Landscape
Singapore, Feb. 19, 2025 (GLOBE NEWSWIRE) -- On February 18, SkyReels open-sourced the world's first human-centric video foundation model for AI short drama creation, SkyReels-V1, and the world's first SOTA-level expressive portrait image animation based on video diffusion transformers, SkyReels-A1.
Open-Source Repositories:
SkyReels-V1https://github.com/SkyworkAI/SkyReels-V1
SkyReels-A1https://github.com/SkyworkAI/SkyReels-A1
Technical Report:https://arxiv.org/abs/2502.10841
Official Website:skyreels.ai
Addressing global pain points in AI video generation—such as closed-source models, limited accessibility, high costs, and usability issues—SkyReels is breaking new ground by open-sourcing two SOTA-level models and algorithms, SkyReels-V1 and SkyReels-A1. These cutting-edge technologies for AI short drama creation are now offered to the open-source community and AIGC users.
The production format for AI videos and short dramas has been market-validated. SkyReels helps address the challenges in traditional short drama production—such as complex offline processes including scriptwriting, casting, set design, storyboard creation, filming, and post-production, which require substantial manpower, incur high costs, and extend production cycles.
01SkyReels-V1: Human-Centric Video Foundation Model, the world's first open-source video generation model dedicated to AI short drama creation
AI short dramas require precise control over both cognitive and physical expressions—integrating lip-sync, facial expression, and body movement generation into a unified process. Currently, lip-sync generation is particularly well-developed, owing to its strong mapping with audio cues that enable high precision and superior user experience.
Yet, the true quality of AI short drama generation lies in the nuances of character performance. To dramatically enhance the controllability of facial expressions and body movements, SkyReels-V1 not only meticulously annotates performance details but also processes emotions, scene context, and acting intent, fine-tuning on tens of millions of high-quality, Hollywood-level data points.
Research team has implemented advanced technical upgrades to capture micro-expressions, performance subtleties, scene descriptions, lighting, and composition. As a result, characters generated by SkyReels now exhibit remarkably precise acting details—approaching an award-winning level.
SkyReels-V1 delivers cinematic-grade micro-expression performance, supporting 33 nuanced facial expressions and over 400 natural motion combinations that faithfully reproduce genuine human emotional expression. As demonstrated in the accompanying video, SkyReels-V1 can generate expressions ranging from hearty laughter, fierce roars, and astonishment to tears—showcasing rich, dynamic performance details.
Moreover, SkyReels-V1 brings cinematic-level lighting and aesthetics to AI video generation. Trained on Hollywood-level high-quality film data, every frame generated by SkyReels exhibits cinematic quality in composition, actor positioning, and camera angles.
Whether capturing solo performance details or multi-character scenes, the model now achieves precise expression control and high-quality visuals.
Importantly, SkyReels-V1 supports both text-to-video and image-to-video generation. It is the largest open-source video generation model supporting image-to-video tasks at equivalent resolution, achieving SOTA-level performance across multiple metrics.
Figure 1: Comparison of Text-to-Video Metrics for SkyReels-V1 (Source: SkyReels)
Such SOTA-level performance is made possible not only by SkyReels self-developed high-quality data cleaning and manual annotation pipeline—which has built a tens-of-millions–scale dataset from movies, TV shows, and documentaries—but also by 'Human-Centric' multimodal video understanding model, which significantly enhances the ability to interpret human-related elements in video, particularly through in-house character intelligence analysis system.
In summary, thanks to the robust data foundation and advanced character intelligence analysis system, SkyReels-V1 can achieve:
Cinematic Expression Recognition: 11 types of facial expression understanding for characters in film and drama, including expressions such as disdain, impatience, helplessness, and disgust, with emotional intensity levels categorized into strong, medium, and weak.
Character Spatial Awareness: Leveraging 3D reconstruction technology to comprehend spatial relationships among multiple characters, enabling cinematic positioning.
Behavioral Intent Understanding: Constructing over 400 behavioral semantic units for precise action interpretation.
Scene-Performance Correlation: Analyzing the interplay between characters, wardrobe, setting, and plot.
https://youtu.be/U2SXDzQGnbQSkyReels-V1 is not only among the very few open-source video foundation models worldwide, but it is also the most powerful in terms of performance for character-driven video generation.
With SkyReels self-developed inference optimization framework 'SkyReels-Infer,' the inference efficiency has been significantly improved —achieving 544p video generation on a single RTX 4090 in just 80 seconds. The framework supports distributed multi-GPU parallelism, Context Parallel, CFG Parallel, and VAE Parallel. Furthermore, by implementing FP8 quantization and parameter-level offload, it meets the requirements of low-memory GPUs; support for flash attention, SageAttention, and model compilation optimizations further reduces latency; and leveraging the open-source diffuser library enhances usability.Figure 2: Using equivalent RTX 4090 resources (4 GPUs), the SkyReels-Infer version reduces end-to-end latency by 58.3% compared to the HunyuanVideo official version (293.3s vs. 464.3s).
Figure 3: Under similar A800 resource conditions, the SkyReels-Infer version reduces end-to-end latency by 14.7%–28.2% compared to the HunyuanVideo official version, demonstrating a more robust multi-GPU deployment strategy.
02SkyReels-A1: The First SOTA-Level Expressive Portrait Image Animation Algorithm Based on Video Diffusion Transformers
To achieve even more precise and controllable character video generation, the SkyReels is open-sourcing SkyReels-A1, a SOTA-level algorithm based on video diffusion transformer for expression and action control. Comparable to Runway Act-One, SkyReels-A1 supports video-driven, film-grade expression capture, enabling high-fidelity micro-expression reproduction.
SkyReels-A1 can generate highly realistic and consistency videos for characters in any reference conditions—from portrait of half-body to full-body shots. It achieves a precise simulation of facial expressions, emotional nuances, skin textures, and body movements.
By inputting both a reference and a driving video, SkyReels-A1 'transplants' the facial expressions and actions details from the driving video onto the character in the reference image. The resulting video shows no distortion and faithfully reproduces the micro-expressions and body movements from the driving video, even surpassing the video quality generated by Runway Act-One in evaluation.
More encouragingly, SkyReels-A1 not only supports profile-based expression control but also enables highly realistic eyebrow and eye micro-expression alignment, along with more pronounced head movements and natural body motions.
For example, in the same dialogue scene, while the character generated by Runway Act-One shows noticeable distortion and deviates from the original appearance, SkyReels-A1 preserves the character's details, maintaining authentic nuance and seamlessly blending facial expressions as well as body movements.
Furthermore, SkyReels-A1 can drive more dramatic facial expressions scenes. Compared to Runway Act-One that fails to generate the desired effect, SkyReels-A1 support to transfer more complex expression dynamics, enabling the character's facial emotions to naturally synchronize with body movements and scene content for an exceptionally life–like performance.
03Empowering the Global AI Short Drama Ecosystem through Open-Sourcing
Video generation models are among the most challenging components of AI short drama creation. Although model generation capabilities have significantly improved over the past year, there remains a considerable gap—particularly given the high production costs.
By open-sourcing our SOTA-level models, SkyReels-V1 and SkyReels-A1, SkyReels becomes the first in the AI short drama industry to take such a step. This initiative not only represents a modest yet significant contribution to the industry but also marks a major leap toward fostering a flourishing ecosystem for AI short drama creation and video generation.
It's believed that with further advancements in inference optimization and the open-sourcing of controllable algorithms, these models will soon provide users with more cost-effective and highly controllable AIGC capabilities. SkyReels aims to empower users to create AI short dramas at minimal cost, overcome current issues of inconsistent video generation, and enable everyone to generate detailed, controllable character performances using their own computers.
This open-sourcing of our video generation models is not only a technological breakthrough that helps narrow the digital divide in the global content industry, but it is also a revolution in cultural production capacity. In the future, the convergence of short dramas, gaming, virtual reality, and other fields will accelerate industrial integration. AI short dramas have the potential to evolve from a 'tech experiment' into a mainstream creative medium and become a new vehicle for global cultural expression.
'Achieve artificial general intelligence and empower everyone to better shape and express themselves.' With this open-source initiative, SkyReels will continue to release more video generation models, algorithms, and universal models—advancing AGI equity and fostering the sustained growth and prosperity of the AI short drama ecosystem, while benefiting the open-source community, developer ecosystems, and the broader AI industry.
CONTACT: Jingnan Fu Skywork AI PTE.LTD. fujingnan(at)singularity-ai.comSign in to access your portfolio
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
15 minutes ago
- Yahoo
High Throughput Process Development Strategic Business Report 2025-2030: Increasing Focus on Speed and Efficiency in Drug Discovery Expands Market Opportunities
The global High Throughput Process Development (HTPD) market, pivotal for the biopharmaceutical industry, is projected to grow from $15.3 billion in 2024 to $24.6 billion by 2030, at a CAGR of 8.3%. Key drivers include technological innovations in automation and data analytics. The market is segmented by product types, including consumables and software, with major growth forecasted in regions such as China and the U.S. The report offers insights into market trends, regional analysis, and profiles leading companies like Agilent Technologies and Danaher Corporation. High Throughput Process Development Market Dublin, June 09, 2025 (GLOBE NEWSWIRE) -- The "High Throughput Process Development - Global Strategic Business Report" report has been added to global market for High Throughput Process Development was valued at US$15.3 Billion in 2024 and is projected to reach US$24.6 Billion by 2030, growing at a CAGR of 8.3% from 2024 to 2030. This comprehensive report provides an in-depth analysis of market trends, drivers, and forecasts, helping you make informed business decisions. The report includes the most recent global tariff developments and how they impact the High Throughput Process Development market. The growth in the high throughput process development market is driven by several factors, including the increasing complexity of biologic drugs, the rising demand for personalized medicines, and advancements in automation and data analytics. As the biopharmaceutical industry shifts toward more complex therapeutics, such as gene and cell therapies, the need for efficient process optimization is more critical than ever. HTPD allows companies to streamline the development of these therapies, reducing both time and costs. The growing demand for biosimilars, which require rapid process development to compete with branded biologics, is also contributing to market growth. Additionally, the adoption of automation and machine learning in process development is further enhancing the efficiency and accuracy of bioprocess optimization, driving the market's throughput process development (HTPD) is a critical tool in the biopharmaceutical industry, enabling faster, more efficient development of biologic drugs. HTPD involves the use of automated, miniaturized experiments that allow researchers to test multiple process conditions simultaneously, significantly accelerating the optimization of production processes for therapeutic proteins, vaccines, and monoclonal antibodies. The ability to screen numerous variables at once reduces the time and resources needed to scale up bioprocesses, from lab-scale to commercial manufacturing. As biopharmaceutical companies face increasing pressure to reduce costs and bring therapies to market faster, HTPD has become indispensable in ensuring both process efficiency and product Are Technological Innovations Shaping High Throughput Process Development?Technological innovations, particularly in automation, robotics, and data analytics, have revolutionized high throughput process development. Automation systems now allow for the simultaneous running of hundreds of experiments, enabling researchers to optimize cell culture conditions, purification processes, and formulation strategies more efficiently. The integration of advanced data analytics and machine learning into HTPD platforms has further enhanced process optimization, allowing for predictive modeling and real-time decision-making. Microfluidic technologies, which allow for the precise control of small volumes of liquids, are also playing a critical role in reducing reagent use and increasing the speed of bioprocess optimization. These technologies are enabling biopharmaceutical companies to streamline their development timelines and reduce production Do Market Segments Define the Growth of the High Throughput Process Development Market?Product types include consumables, software, and instruments, with consumables like reagents and microplates holding the largest market share due to their constant need in HTPD workflows. Services such as contract research and process development outsourcing are also expanding, as biopharmaceutical companies increasingly seek external expertise to accelerate drug development. The biopharmaceutical industry, particularly companies focused on biologics such as monoclonal antibodies, is the primary end-user of HTPD, as it enables rapid scale-up from R&D to manufacturing. Other key users include academic and research institutes involved in biotechnology Scope Report Features: Comprehensive Market Data: Independent analysis of annual sales and market forecasts in US$ Million from 2024 to 2030. In-Depth Regional Analysis: Detailed insights into key markets, including the U.S., China, Japan, Canada, Europe, Asia-Pacific, Latin America, Middle East, and Africa. Company Profiles: Coverage of players such as Agilent Technologies, Aurora Biomed, Axxam S.P.A., Bio-Rad Laboratories, Danaher Corporation and more. Complimentary Updates: Receive free report updates for one year to keep you informed of the latest market developments. Key Insights: Market Growth: Understand the significant growth trajectory of the Monoclonal Antibodies segment, which is expected to reach US$17.8 Billion by 2030 with a CAGR of a 9.1%. The Other Molecule Types segment is also set to grow at 6.5% CAGR over the analysis period. Regional Analysis: Gain insights into the U.S. market, valued at $3.9 Billion in 2024, and China, forecasted to grow at an impressive 12.6% CAGR to reach $6.4 Billion by 2030. Discover growth trends in other key regions, including Japan, Canada, Germany, and the Asia-Pacific. Segments Molecule Type (Monoclonal Antibodies, Other Molecule Types) Component (Tools & Systems, Consumables, Software, Other Components) Technology (Chromatography, Ultraviolet-Visible Spectroscopy, Other Technologies) End-Use (Biopharmaceutical Companies, Contract Research Organizations, Academic Research Institutes) Tariff Impact Analysis: Key Insights for 2025What's Included in This Edition: Tariff-adjusted market forecasts by region and segment Analysis of cost and supply chain implications by sourcing and trade exposure Strategic insights into geographic shifts Buyers receive a free July 2025 update with: Finalized tariff impacts and new trade agreement effects Updated projections reflecting global sourcing and cost shifts Expanded country-specific coverage across the industry Key Attributes: Report Attribute Details No. of Pages 89 Forecast Period 2024 - 2030 Estimated Market Value (USD) in 2024 $15.3 Billion Forecasted Market Value (USD) by 2030 $24.6 Billion Compound Annual Growth Rate 8.3% Regions Covered Global Key Topics Covered: MARKET OVERVIEW Influencer Market Insights Tariff Impact on Global Supply Chain Patterns Global Economic Update High Throughput Process Development - Global Key Competitors Percentage Market Share in 2024 (E) Competitive Market Presence - Strong/Active/Niche/Trivial for Players Worldwide in 2024 (E) MARKET TRENDS & DRIVERS Rising Demand for Biopharmaceuticals Drives Growth in High Throughput Process Development Technological Advancements in Automation and Robotics Propel Innovation in High Throughput Process Development Increasing Focus on Speed and Efficiency in Drug Discovery Expands Addressable Market for High Throughput Solutions Growing Adoption of High Throughput Technologies in Vaccine Development Spurs Market Growth Surge in Demand for Personalized Medicine Strengthens Business Case for High Throughput Process Development Increasing Use of High Throughput Screening in Protein and Antibody Discovery Expands Market Potential Technological Innovations in Microfluidics and Lab Automation Propel Market Innovation Growing Focus on Reducing Time-to-Market for Pharmaceuticals Drives Adoption of High Throughput Processes Rising Use of Artificial Intelligence (AI) in High Throughput Data Analysis Strengthens Competitive Advantage Increasing Focus on Process Optimization and Scalability in Biopharmaceutical Production Drives Innovation Rising Demand for High Throughput Analytical Techniques in Quality Control Expands Market Opportunities FOCUS ON SELECT PLAYERS:Some of the 33 companies featured in this report Agilent Technologies Aurora Biomed Axxam S.P.A. Bio-Rad Laboratories Danaher Corporation Eppendorf GE Healthcare Hamilton Company Merck Millipore Perkinelmer Sartorius Stedim Biotech Tecan Group Thermo Fisher Scientific, Inc. For more information about this report visit About is the world's leading source for international market research reports and market data. We provide you with the latest data on international and regional markets, key industries, the top companies, new products and the latest trends. Attachment High Throughput Process Development Market CONTACT: CONTACT: Laura Wood,Senior Press Manager press@ For E.S.T Office Hours Call 1-917-300-0470 For U.S./ CAN Toll Free Call 1-800-526-8630 For GMT Office Hours Call +353-1-416-8900Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data


Axios
16 minutes ago
- Axios
Behind the Curtain: The scariest AI reality
The wildest, scariest, indisputable truth about AI's large language models is that the companies building them don't know exactly why or how they work. Sit with that for a moment. The most powerful companies, racing to build the most powerful superhuman intelligence capabilities — ones they readily admit occasionally go rogue to make things up, or even threaten their users — don't know why their machines do what they do. Why it matters: With the companies pouring hundreds of billions of dollars into willing superhuman intelligence into a quick existence, and Washington doing nothing to slow or police them, it seems worth dissecting this Great Unknown. None of the AI companies dispute this. They marvel at the mystery — and muse about it publicly. They're working feverishly to better understand it. They argue you don't need to fully understand a technology to tame or trust it. Two years ago, Axios managing editor for tech Scott Rosenberg wrote a story, "AI's scariest mystery," saying it's common knowledge among AI developers that they can't always explain or predict their systems' behavior. And that's more true than ever. Yet there's no sign that the government or companies or general public will demand any deeper understanding — or scrutiny — of building a technology with capabilities beyond human understanding. They're convinced the race to beat China to the most advanced LLMs warrants the risk of the Great Unknown. The House, despite knowing so little about AI, tucked language into President Trump's "Big, Beautiful Bill" that would prohibit states and localities from any AI regulations for 10 years. The Senate is considering limitations on the provision. Neither the AI companies nor Congress understands the power of AI a year from now, much less a decade from now. The big picture: Our purpose with this column isn't to be alarmist or " doomers." It's to clinically explain why the inner workings of superhuman intelligence models are a black box, even to the technology's creators. We'll also show, in their own words, how CEOs and founders of the largest AI companies all agree it's a black box. Let's start with a basic overview of how LLMs work, to better explain the Great Unknown: LLMs — including Open AI's ChatGPT, Anthropic's Claude and Google's Gemini — aren't traditional software systems following clear, human-written instructions, like Microsoft Word. In the case of Word, it does precisely what it's engineered to do. Instead, LLMs are massive neural networks — like a brain — that ingest massive amounts of information (much of the internet) to learn to generate answers. The engineers know what they're setting in motion, and what data sources they draw on. But the LLM's size — the sheer inhuman number of variables in each choice of "best next word" it makes — means even the experts can't explain exactly why it chooses to say anything in particular. We asked ChatGPT to explain this (and a human at OpenAI confirmed its accuracy): "We can observe what an LLM outputs, but the process by which it decides on a response is largely opaque. As OpenAI's researchers bluntly put it, 'we have not yet developed human-understandable explanations for why the model generates particular outputs.'" "In fact," ChatGPT continued, "OpenAI admitted that when they tweaked their model architecture in GPT-4, 'more research is needed' to understand why certain versions started hallucinating more than earlier versions — a surprising, unintended behavior even its creators couldn't fully diagnose." Anthropic — which just released Claude 4, the latest model of its LLM, with great fanfare — admitted it was unsure why Claude, when given access to fictional emails during safety testing, threatened to blackmail an engineer over a supposed extramarital affair. This was part of responsible safety testing — but Anthropic can't fully explain the irresponsible action. Again, sit with that: The company doesn't know why its machine went rogue and malicious. And, in truth, the creators don't really know how smart or independent the LLMs could grow. Anthropic even said Claude 4 is powerful enough to pose a greater risk of being used to develop nuclear or chemical weapons. OpenAI's Sam Altman and others toss around the tame word of " interpretability" to describe the challenge. "We certainly have not solved interpretability," Altman told a summit in Geneva last year. What Altman and others mean is they can't interpret the why: Why are LLMs doing what they're doing? Anthropic CEO Dario Amodei, in an essay in April called "The Urgency of Interpretability," warned: "People outside the field are often surprised and alarmed to learn that we do not understand how our own AI creations work. They are right to be concerned: this lack of understanding is essentially unprecedented in the history of technology." Amodei called this a serious risk to humanity — yet his company keeps boasting of more powerful models nearing superhuman capabilities. Anthropic has been studying the interpretability issue for years, and Amodei has been vocal about warning it's important to solve. In a statement for this story, Anthropic said: "Understanding how AI works is an urgent issue to solve. It's core to deploying safe AI models and unlocking [AI's] full potential in accelerating scientific discovery and technological development. We have a dedicated research team focused on solving this issue, and they've made significant strides in moving the industry's understanding of the inner workings of AI forward. It's crucial we understand how AI works before it radically transforms our global economy and everyday lives." (Read a paper Anthropic published last year, "Mapping the Mind of a Large Language Model.") Elon Musk has warned for years that AI presents a civilizational risk. In other words, he literally thinks it could destroy humanity, and has said as much. Yet Musk is pouring billions into his own LLM called Grok. "I think AI is a significant existential threat," Musk said in Riyadh, Saudi Arabia, last fall. There's a 10%-20% chance "that it goes bad." Reality check: Apple published a paper last week, "The Illusion of Thinking," concluding that even the most advanced AI reasoning models don't really "think," and can fail when stress-tested. The study found that state-of-the-art models (OpenAI's o3-min, DeepSeek R1 and Anthropic's Claude-3.7-Sonnet) still fail to develop generalizable problem-solving capabilities, with accuracy ultimately collapsing to zero "beyond certain complexities." But a new report by AI researchers, including former OpenAI employees, called " AI 2027," explains how the Great Unknown could, in theory, turn catastrophic in less than two years. The report is long and often too technical for casual readers to fully grasp. It's wholly speculative, though built on current data about how fast the models are improving. It's being widely read inside the AI companies. It captures the belief — or fear — that LLMs could one day think for themselves and start to act on their own. Our purpose isn't to alarm or sound doomy. Rather, you should know what the people building these models talk about incessantly. You can dismiss it as hype or hysteria. But researchers at all these companies worry LLMs, because we don't fully understand them, could outsmart their human creators and go rogue. In the AI 2027 report, the authors warn that competition with China will push LLMs potentially beyond human control, because no one will want to slow progress even if they see signs of acute danger. The safe-landing theory: Google's Sundar Pichai — and really all of the big AI company CEOs — argue that humans will learn to better understand how these machines work and find clever, if yet unknown ways, to control them and " improve lives." The companies all have big research and safety teams, and a huge incentive to tame the technologies if they want to ever realize their full value.


Fast Company
20 minutes ago
- Fast Company
How to get the most out of Google's free AI Studio
Google's AI Studio and Labs let you experiment for free with new AI tools. I love the way these digital sandboxes—like the one from Hugging Face —let you try out creative new uses of AI. You can dabble around then download and share what you make, without having to master a complex new platform. Read on for a few Google AI experiments to try. All are free, fast, and easy to use. 1. Transform an image Upload a photo and use Gemini's AI Studio Image Generation to transform it with prompts. Iterate on your original image until you get a version you like. The model understands natural language, so you don't have to master prompt lingo. 2. Generate an AI voice conversation AI-generated voices are increasingly hard to distinguish from human ones. If you're surprised, try Generate Speech in the AI Studio or Google's NotebookLM. How to use Generate Speech in Google's AI Studio Paste in text, either for a narration or a conversation between two people Open the settings tab to pick from 30 AI voices. Each is labeled with a characteristic—e.g. upbeat, gravelly, or mature. Click run to generate the conversation. Optionally adjust the playback speed. Download the file if you want to keep it, or paste in different text to try again. Example: a silly 90-sec chat between two violinists I scripted with Gemini and rendered quickly with this Generate Speech tool. Use case: Make a narration track for an instructional video. ElevenLabs has a better professional model for this, but AI Studio's is free, easy and quick. Alternatives Google's Gemini AI app can also now generate audio overviews from files you upload, if you're on a paid plan. Google's free NotebookLM has a new mobile app, and now lets you generate an audio conversation in any of 50 languages. Unlike Generate Speech in AI Studio, NotebookLM audio overviews summarize your material, they don't perform words as written. Why NotebookLM is so useful. Google's Illuminate lets you generate, listen to, share, and download AI conversations about research papers and famous books. Here's an audio chat about David Copperfield, for example. A bit dry to listen to, but still useful. 3. Make a gif Alternative: You can also make a static image with Google's Imagen 3 or the new Imagen 4. Write a short prompt and select your preferred aspect ratio. So far I still prefer Ideogram (why I like it) and ChatGPT's new image engine. 4. Generate a short video Google's Veo 2 and Flow let you generate free short video clips almost instantly with a prompt. Create a clip to add vibrancy or humor to a presentation, or a visual metaphor to help you explain something. Here are 25 other quick ideas for how you might use little AI-generated video scenes. How to create a video clip with Veo 2 Pick a length (5 to 8 seconds) and select horizontal or vertical orientation Write a prompt & optionally upload a photo to suggest a visual direction Example: Take a look at a parakeet photo I started with and the 5-second video I generated from the photo with Veo 2. Tip: Convert short video clips into gifs for free with Ezgif or Giphy. Unlike video files, gifs are easy to share and auto-play in an email or presentation. What's next: Remarkably lifelike clips made with Google's newer Veo 3 model went viral this week. These AI-generated visuals—with sound—are only available on the $250/month(!) plan for now, so try Veo 2 for free. 5. Explain things with lots of tiny cats This playful mini app creates short, step-by-step visual guides using charming cat illustrations to explain any concept, from how a violin works to the concept behind the matrix.