logo
ElevenLabs Launches Eleven v3 (alpha) : New Expressive Text to Speech Model

ElevenLabs Launches Eleven v3 (alpha) : New Expressive Text to Speech Model

Geeky Gadgets06-06-2025
ElevenLabs has launched Eleven v3 (alpha), a new Text to Speech model designed to deliver highly expressive and realistic speech generation. This version introduces advanced features like multi-speaker dialogue, inline audio tags for emotional and tonal control, and support for over 70 languages. While it requires more prompt engineering than previous models, it offers significant improvements in expressiveness and naturalness, making it ideal for applications in media, audiobooks, and creative projects. A real-time version is under development, and API access will be available soon.
At the core of Eleven v3 is its ability to produce highly expressive and lifelike speech, offering users greater control over tone, emotion, and delivery. This is achieved through several innovative features: ElevenLabs Eleven v3 (alpha) Text to Speech AI Model Advanced emotional and tonal controls: Users can fine-tune voice delivery to convey specific emotions or tones, enhancing the natural flow of speech.
Users can fine-tune voice delivery to convey specific emotions or tones, enhancing the natural flow of speech. Inline audio tags: Tags such as '[whispers]' or '[laughs]' allow for the seamless integration of non-verbal cues like sighs, laughter, and whispers, making speech more dynamic and engaging.
Tags such as '[whispers]' or '[laughs]' allow for the seamless integration of non-verbal cues like sighs, laughter, and whispers, making speech more dynamic and engaging. Multi-speaker dialogue synthesis: The new Text-to-Dialogue API enables the creation of overlapping, realistic conversations between multiple speakers, complete with smooth transitions and nuanced emotional shifts.
These features make Eleven v3 particularly valuable for applications such as storytelling, audiobooks, media production, and interactive entertainment. By allowing more natural and expressive speech, the model enhances the overall user experience across a variety of platforms.
Watch this video on YouTube. Breaking Language Barriers
Eleven v3 addresses the growing demand for multilingual support by offering compatibility with over 70 languages. This capability ensures that speech output maintains natural stress, cadence, and contextual accuracy across diverse linguistic settings. Improved linguistic adaptability: The model demonstrates a deeper understanding of accents, dialects, and cultural nuances, making it suitable for a wide range of global audiences.
The model demonstrates a deeper understanding of accents, dialects, and cultural nuances, making it suitable for a wide range of global audiences. Applications in multilingual projects: Eleven v3 is well-suited for international audiobooks, educational content, and customer support systems, allowing creators to reach broader audiences.
By supporting diverse languages and accents, Eleven v3 fosters inclusive communication and helps bridge language gaps, making it a valuable tool for global accessibility. Real-Time Capabilities and Developer Integration
Although Eleven v3 currently requires more prompt engineering than its predecessors, a real-time version is under development. This future iteration is expected to cater to applications that demand instantaneous speech synthesis, such as live voiceovers and conversational AI systems.
The model also offers robust API integration, allowing developers to incorporate its features into existing workflows and platforms. This flexibility makes Eleven v3 a versatile tool for industries such as: Gaming: Creating lifelike character voices and immersive in-game dialogues.
Creating lifelike character voices and immersive in-game dialogues. Film and media: Enhancing voiceovers and character-driven narratives.
Enhancing voiceovers and character-driven narratives. Education: Generating engaging and accessible learning materials.
Generating engaging and accessible learning materials. Accessibility: Improving digital tools for individuals with disabilities.
The combination of real-time capabilities and developer-friendly integration ensures that Eleven v3 can meet the diverse needs of professionals across multiple sectors. Applications Across Industries
The enhanced expressiveness and realism of Eleven v3 open up a wide range of applications, particularly in creative and functional domains. Media and entertainment: Filmmakers and game developers can use the model to create lifelike character voices, while audiobook producers can deliver more emotionally resonant narratives.
Filmmakers and game developers can use the model to create lifelike character voices, while audiobook producers can deliver more emotionally resonant narratives. Accessibility tools: The model's ability to generate clear and expressive speech can improve digital experiences for individuals with visual impairments or other disabilities, making content more inclusive.
The model's ability to generate clear and expressive speech can improve digital experiences for individuals with visual impairments or other disabilities, making content more inclusive. Customer service: Multilingual and emotionally nuanced speech capabilities can enhance automated customer support systems, providing a more human-like interaction.
Multilingual and emotionally nuanced speech capabilities can enhance automated customer support systems, providing a more human-like interaction. Education: Eleven v3 can be used to create engaging educational content, including language learning tools and interactive lessons.
By offering a combination of emotional depth, linguistic versatility, and technical precision, Eleven v3 has the potential to transform how industries approach voice generation and communication. Availability and Future Developments
Eleven v3 is currently available on the ElevenLabs platform, with an 80% discount on the ElevenLabs app offered until the end of June. API access and Studio support are expected to roll out soon, with early access available through direct sales contact.
For applications requiring real-time speech synthesis, ElevenLabs recommends using v2.5 Turbo or Flash until the real-time version of v3 becomes available. Addressing Challenges and Advancing TTS Technology
Eleven v3 was designed to address the limitations of earlier models, particularly in terms of expressiveness and naturalness. By allowing lifelike and responsive speech, the model meets the needs of professionals in industries such as film, gaming, education, and accessibility.
As demand for realistic AI voice generation continues to grow, Eleven v3 represents a significant advancement in TTS technology. Its combination of emotional nuance, multilingual support, and developer-friendly integration positions it as a valuable tool for both creative and functional applications.
By focusing on realism, versatility, and accessibility, Eleven v3 demonstrates the potential of AI-driven speech synthesis to enhance communication and storytelling across a wide range of industries. Here are additional guides from our expansive article library that you may find useful on Text-to-Speech. Filed Under: AI, Top News
Latest Geeky Gadgets Deals
Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Trump pauses export controls to bolster China trade deal, FT says
Trump pauses export controls to bolster China trade deal, FT says

Reuters

time18 minutes ago

  • Reuters

Trump pauses export controls to bolster China trade deal, FT says

July 28 (Reuters) - The U.S. has paused curbs on tech exports to China to avoid disrupting trade talks with Beijing and support President Donald Trump's efforts to secure a meeting with President Xi Jinping this year, the Financial Times said on Monday. The industry and security bureau of the Commerce Department, which oversees export controls, has been told in recent months to avoid tough moves on China, the newspaper said, citing current and former officials. Reuters could not immediately verify the report. The White House and the department did not respond to Reuters' requests for comment outside business hours. Top U.S. and Chinese economic officials are set to resume talks in Stockholm on Monday to tackle longstanding economic disputes at the centre of a trade war between the world's top two economies. Tech giant Nvidia (NVDA.O), opens new tab said this month it would resume sales of its H20 graphics processing units (GPU) to China, reversing an export curb the Trump administration imposed in April to keep advanced AI chips out of Chinese hands over national security concerns. The planned resumption was part of U.S. negotiations on rare earths and magnets, Commerce Secretary Howard Lutnick has said. The paper said 20 security experts and former officials, including former deputy US national security adviser Matt Pottinger, will write on Monday to Lutnick to voice concern, however. "This move represents a strategic misstep that endangers the United States' economic and military edge in artificial intelligence," they write in the letter, it added.

Chinese AI firms form alliances to build domestic ecosystem amid US curbs
Chinese AI firms form alliances to build domestic ecosystem amid US curbs

Reuters

time18 minutes ago

  • Reuters

Chinese AI firms form alliances to build domestic ecosystem amid US curbs

SHANGHAI, July 28 (Reuters) - China's artificial intelligence companies have announced two new industry alliances, aiming to develop a domestic ecosystem to reduce dependence on foreign tech as they seek to cope with U.S. export restrictions on advanced Nvidia (NVDA.O), opens new tab chipsets. The announcements were timed to coincide with the three-day World Artificial Intelligence Conference in Shanghai ending on Monday. The conference also showcased a slew of new products, such as an AI computing system from Huawei ( that experts believe rivals Nvidia's most advanced offering, as well as consumer-friendly products such as several kinds of digital AI glasses. The "Model-Chip Ecosystem Innovation Alliance" brings together Chinese developers of large language models (LLMs) and AI chip manufacturers. "This is an innovative ecosystem that connects the complete technology chain from chips to models to infrastructure," said Zhao Lidong, CEO of Enflame, one of the participating chipmakers. Other manufacturers of graphics processing units (GPUs) in the alliance include Huawei, Biren, and Moore Threads, which have been hit by U.S. sanctions that block them from purchasing advanced tech made with U.S. know-how. The alliance was announced by StepFun, an LLM developer. A second alliance, the Shanghai General Chamber of Commerce AI Committee, aims to "promote the deep integration of AI technology and industrial transformation." Participants include SenseTime ( opens new tab, also sanctioned by the U.S. and which has pivoted from facial recognition technology to LLMs. Others are StepFun and another LLM developer, MiniMax, as well as chipmakers Metax and Iluvatar CoreX. One of the most talked about products at the conference was Huawei's CloudMatrix 384 which incorporates 384 of its latest 910C chips and outperforms Nvidia's GB200 NVL72 on some metrics, according to U.S. research firm SemiAnalysis. Huawei's system design capabilities have meant that it has been able to use more chips and system-level innovations to compensate for weaker individual chip performance, SemiAnalysis said. At least six other Chinese computing firms showcased similar "clustering" chip technology. Metax demonstrated an AI supernode featuring 128 C550 chips designed to support large-scale liquid-cooled data centre requirements. Other events included Tencent's ( opens new tab unveiling of its open-source Hunyuan3D World Model 1.0, which the company said enables users to generate interactive 3D environments through text or image prompts. Baidu ( opens new tab announced what it said was next-generation "digital human" technology that helps businesses to create virtual livestreamers. It features "cloning technology" that can replicate a human's voice, tone, and body language from just 10 minutes of sample footage. Alibaba ( opens new tab was among those announcing AI glasses. Its Quark AI Glasses are powered by its Qwen AI model and are due to be released in China by the end of 2025. They will allow users to access the tech giant's map service for easy navigating and to use Alipay by scanning QR codes with voice commands.

Council in Peterborough turns experienced worker into AI chatbot
Council in Peterborough turns experienced worker into AI chatbot

BBC News

time18 minutes ago

  • BBC News

Council in Peterborough turns experienced worker into AI chatbot

The knowledge of an experienced council worker has been turned into an AI chatbot to help fellow Jinks has worked for Peterborough City Council for 35 years as a therapy practitioner supporting adult social care, and her colleagues often ask her for council said she was often inundated with questions and, although she was happy to help, it did put pressure on her time. Her manager recognised this, and the authority started working on a technological solution, developing a tool called Hey Geraldine. Ms Jinks worked with the team to feed information to the AI tool so it could answer questions asked on a daily basis in her usual chatty and direct manner."It was so exciting to develop this chatbot," she said. "Staff can now ask me questions whenever they want and receive a reply straightaway."It's lovely that they've kept my name and some staff have told me that they actually thought they were chatting with me – I guess that's because I've written all the answers!"The council said Hey Geraldine could respond to "contextual and practice-based questions" such as, "I am working with a person who is forgetting to turn the oven off. What is available to help them?" It would then provide a range of answers and possible solutions and links to is hoped the technology will make it easier and quicker to discharge people from hospital. Shabina Qayyum is the council's cabinet member for adults and health and also works as a GP. "I know how incredibly frustrating it can be for patients who are delayed in returning to their home after a stay in hospital," she also have an impact on the finances of both the NHS and local authorities."The use of technology-assisted care like smart devices and home activity sensors can really help people to regain their independence quickly. "The Hey Geraldine chatbot helps staff to put these technologies in place with all the knowledge and care that Geraldine herself would use." Follow Peterborough news on BBC Sounds, Facebook, Instagram and X.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store