logo
Google made an AI model to talk to dolphins

Google made an AI model to talk to dolphins

Yahoo14-04-2025

A new large language model AI system may soon allow humans to converse with dolphins. Scheduled to debut in the coming months, researchers will test to see if DolphinGemma and its companion Cetacean Hearing Augmentation Telemetry (CHAT) system can translate and mimic some of the mammal's own complex vocalizations. If successful, the breakthrough may represent the culmination of over four decades' worth of work, documentation, and conservation efforts..
Dolphins are some of the Earth's smartest and most communicative animals. Their social interactions are so complex that researchers at the Wild Dolphin Project (WDP) have spent the last 40 years attempting to decipher them. In the process, WDP has amassed decades' worth of underwater audio and video documenting a single community of Atlantic spotted dolphins in the Bahamas. In the process, they have been able to correlate sounds with behavioral aspects like courtships, unique names, and dolphin squabbles.
Experts have long theorized it may be possible for humans to actually communicate with the cetaceans, but lacked technology advanced enough to parse and mimic the species' underwater whistles, clicks, and burst pulses. With the rise of large language models (LLMs), researchers recently wondered if the same principles underlying LLMs could be applied to dolphin interactions. To test this possibility, WDP recently partnered with Google and the Georgia Institute of Technology, supplying engineers with a massive, labeled dataset of dolphin whistles, clicks, and burst pulses for use in LLM training.
The result is DolphinGemma, an AI model built using the same technology that runs Google's Gemini systems. DolphinGemma is designed on roughly 400 million parameters to function in essentially the same way as predictive LLMs like ChatGPT—but for dolphins.
DolphinGemma first receives and interprets audio inputs, then predicts likely subsequent sounds for recreation. It is next partnered with the CHAT system installed on modified Google Pixel smartphones. CHAT isn't designed to fully translate a dolphin's natural language, but help humans convey and establish a more simplified, shared vocabulary. The plan is to ostensibly teach members of the WDP's Atlantic spotted dolphin community a series of synthetic whistles with their favorite objects such as seagrass, sargassum, and even researchers' scarves. Over time, experts hope that the dolphins will even learn to request desired items when they want to play.
There's still a lot of work to be done before humans and dolphins bridge the interspecies communication gap. But with this creative use of LLMs, those underwater conversations are another step closer.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Here's how Uber's product chief uses AI at work — and one tool he's going to use next
Here's how Uber's product chief uses AI at work — and one tool he's going to use next

Business Insider

timean hour ago

  • Business Insider

Here's how Uber's product chief uses AI at work — and one tool he's going to use next

Uber's chief product officer has one AI tool on his to-do list. In an episode of "Lenny's Podcast" released on Sunday, Uber's product chief, Sachin Kansal, shared two ways he is using AI for his everyday tasks at the ride-hailing giant and how he plans to add NotebookLM to his AI suite. Kansal joined Uber eight years ago as its director of product management after working at cybersecurity and taxi startups. He became Uber's product chief last year. Kansal said he uses OpenAI's ChatGPT and Google's Gemini to summarize long reports. "Some of these reports, they're 50 to 100 pages long," he said. "I will never have the time to read them." He said he uses the chatbots to acquaint himself with what's happening and how riders are feeling in Uber's various markets, such as South Africa, Brazil, and Korea. The CPO said his second use case is treating AI like a research assistant, because some large language models now offer a deep research feature. Kansal gave a recent example of when his team was thinking about a new driver feature. He asked ChatGPT's deep research mode about what drivers may think of the add-on. "It's an amazing research assistant and it's absolutely a starting point for a brainstorm with my team with some really, really good ideas," the CPO said. In April, Uber's CEO, Dara Khosrowshahi, said that not enough of his 30,000-odd employees are using AI. He said learning to work with AI agents to code is "going to be an absolute necessity at Uber within a year." Uber did not immediately respond to a request for comment from Business Insider. Kansal's next tool: NotebookLM On the podcast, Kansal also highlighted NotebookLM, Google Lab's research and note-taking tool, which is especially helpful for interacting with documents. He said he doesn't use the product yet, but wants to. "I know a lot of people who have started using it, and that is the next thing that I'm going to use," he said. "Just to be able to build an audio podcast based on a bunch of information that you can consume. I think that's awesome," he added. Kansal was referring to the "Audio Overview" feature, which summarizes uploaded content in the form of two AIs having a voice discussion. NotebookLM was launched in mid-2023 and has quickly become a must-have tool for researchers and AI enthusiasts. Andrej Karpathy, Tesla's former director of AI and OpenAI cofounder, is among those who have praised the tool and its podcast feature. "It's possible that NotebookLM podcast episode generation is touching on a whole new territory of highly compelling LLM product formats," he said in a September post on X. "Feels reminiscent of ChatGPT. Maybe I'm overreacting."

We're close to translating animal languages – what happens then?
We're close to translating animal languages – what happens then?

Yahoo

time2 hours ago

  • Yahoo

We're close to translating animal languages – what happens then?

Charles Darwin suggested that humans learned to speak by mimicking birdsong: our ancestors' first words may have been a kind of interspecies exchange. Perhaps it won't be long before we join the conversation once again. The race to translate what animals are saying is heating up, with riches as well as a place in history at stake. The Jeremy Coller Foundation has promised $10m to whichever researchers can crack the code. This is a race fuelled by generative AI; large language models can sort through millions of recorded animal vocalisations to find their hidden grammars. Most projects focus on cetaceans because, like us, they learn through vocal imitation and, also like us, they communicate via complex arrangements of sound that appear to have structure and hierarchy. Sperm whales communicate in codas – rapid sequences of clicks, each as brief as 1,000th of a second. Project Ceti (the Cetacean Translation Initiative) is using AI to analyse codas in order to reveal the mysteries of sperm whale speech. There is evidence the animals take turns, use specific clicks to refer to one another, and even have distinct dialects. Ceti has already isolated a click that may be a form of punctuation, and they hope to speak whaleish as soon as 2026. The linguistic barrier between species is already looking porous. Last month, Google released DolphinGemma, an AI program to translate dolphins, trained on 40 years of data. In 2013, scientists using an AI algorithm to sort dolphin communication identified a new click in the animals' interactions with one another, which they recognised as a sound they had previously trained the pod to associate with sargassum seaweed – the first recorded instance of a word passing from one species into another's native vocabulary. Humpback whale songs are incredible vocal performances, sometimes lasting up to 24 hours The prospect of speaking dolphin or whale is irresistible. And it seems that they are just as enthusiastic. In November last year, scientists in Alaska recorded an acoustic 'conversation' with a humpback whale called Twain, in which they exchanged a call-and-response form known as 'whup/throp' with the animal over a 20-minute period. In Florida, a dolphin named Zeus was found to have learned to mimic the vowel sounds, A, E, O, and U. But in the excitement we should not ignore the fact that other species are already bearing eloquent witness to our impact on the natural world. A living planet is a loud one. Healthy coral reefs pop and crackle with life. But soundscapes can decay just as ecosystems can. Degraded reefs are hushed deserts. Since the 1960s, shipping and mining have raised background noise in the oceans by about three decibels a decade. Humpback whale song occupies the same low-frequency bandwidth as deep-sea dredging and drilling for the rare earths that are vital for electronic devices. Ironically, mining the minerals we need to communicate cancels out whales' voices. Humpback whale songs are incredible vocal performances, sometimes lasting up to 24 hours. 'Song' is apt: they seem to include rhymed phrases, and their compositions travel the oceans with them, evolving as they go in a process called 'song revolutions', where a new cycle replaces the old. (Imagine if Nina Simone or the Beatles had erased their back catalogue with every new release.) They're crucial to migration and breeding seasons. But in today's louder soundscape, whale song is crowded out of its habitual bandwidth and even driven to silence – from up to 1.2 km away from commercial ships, humpback whales will cease singing rather than compete with the noise. In interspecies translation, sound only takes us so far. Animals communicate via an array of visual, chemical, thermal and mechanical cues, inhabiting worlds of perception very different to ours. Can we really understand what sound means to echolocating animals, for whom sound waves can be translated visually? The German ecologist Jakob von Uexküll called these impenetrable worlds umwelten. To truly translate animal language, we would need to step into that animal's umwelt – and then, what of us would be imprinted on her, or her on us? 'If a lion could talk,' writes Stephen Budiansky, revising Wittgenstein's famous aphorism in Philosophical Investigations, 'we probably could understand him. He just would not be a lion any more.' We should ask, then, how speaking with other beings might change us. Talking to another species might be very like talking to alien life. It's no coincidence that Ceti echoes Nasa's Seti – Search for Extraterrestrial Intelligence – Institute. In fact, a Seti team recorded the whup/throp exchange, on the basis that learning to speak with whales may help us if we ever meet intelligent extraterrestrials. In Denis Villeneuve's movie Arrival, whale-like aliens communicate via a script in which the distinction between past, present and future times collapses. For Louise, the linguist who translates the script, learning Heptapod lifts her mind out of linear time and into a reality in which her own past and future are equally available. The film mentions Edward Sapir and Benjamin Whorf's theory of linguistic determinism – the idea that our experience of reality is encoded in language – to explain this. The Sapir-Whorf hypothesis was dismissed in the mid-20th century, but linguists have since argued that there may be some truth to it. Pormpuraaw speakers in northern Australia refer to time moving from east to west, rather than forwards or backwards as in English, making time indivisible from the relationship between their body and the land. Whale songs are born from an experience of time that is radically different to ours. Humpbacks can project their voices over miles of open water; their songs span the widest oceans. Imagine the swell of oceanic feeling on which such sounds are borne. Speaking whale would expand our sense of space and time into a planetary song. I imagine we'd think very differently about polluting the ocean soundscape so carelessly. Where it counts, we are perfectly able to understand what nature has to say; the problem is, we choose not to. As incredible as it would be to have a conversation with another species, we ought to listen better to what they are already telling us. • David Farrier is the author of Nature's Genius: Evolution's Lessons for a Changing Planet (Canongate). Why Animals Talk by Arik Kershenbaum (Viking, £10.99) Philosophical Investigations by Ludwig Wittgenstein (Wiley-Blackwell, £24.95) An Immense World by Ed Yong (Vintage, £12.99)

Google Photos vs. iCloud: Which should you use for your photos?
Google Photos vs. iCloud: Which should you use for your photos?

Tom's Guide

time2 hours ago

  • Tom's Guide

Google Photos vs. iCloud: Which should you use for your photos?

Google Photos and iCloud are among the best cloud storage services that make it easy to sync your photos across devices and share them with friends and family. However, they differ in several key ways, and choosing the wrong platform for your needs could result in a headache. In this guide, we compare Google Photos vs. iCloud so you can decide which photo cloud storage app is right for you. While iCloud can be used for more than just photo storage, for the purposes of this comparison, we're going to highlight what you can do in the Apple Photos app vs. Google Photos. Let's dive into some key differences between Google Photos and iCloud, specifically Apple Photos. Everyone with a Google account receives 15GB of free storage space to use across Google Photos, Gmail, and Google Drive. You can buy additional data through Google One if you need to store more content. Paid plans offer 100GB for $1.99 per month, 1TB for $9.99 per month, and 2TB for $19.99 per month. iCloud offers only 5GB of free storage. iCloud+ plans provide an additional 50GB for $0.99 per month, 200GB for $2.99 per month, or 2TB for $9.99 per month. Plans for 6TB and 12TB of storage are also available. Importantly, Google Photos can store photos and videos only. You can store other types of content, such as documents and audio files, with Google Drive, but you have to switch between apps to access this content. Meanwhile, iCloud supports all types of files in a single location, making it an all-in-one cloud hub for your content. Another notable distinction is that Google Photos enables you to choose the resolution to upload when backing up photos and videos. On the other hand, iCloud automatically backs up original photos at full resolution and replaces the copies on your device with lower-resolution versions. Get instant access to breaking news, the hottest reviews, great deals and helpful tips. Google Photos offers apps for all devices, while iCloud provides apps for Apple devices only. You can install Google Photos on a Mac computer or iPhone, but you can't install iCloud on a Windows computer or Android phone. You can still use iCloud on Windows and Android devices, but you must go through your browser to access the iCloud web app. This significantly detracts from the point of using a cloud storage app. Your photos won't automatically upload to the cloud or sync across devices, and you'll need to visit the iCloud website to view them. Effectively, iCloud is only a good option if you have a Mac computer and an iPhone. If you use any non-Apple devices, you'll have a much better experience with Google Photos. Both Google Photos and Apple Photos enable you to organize images into albums. Likewise, both automatically curate albums based on photo data like location, date, and people. They use AI to tag elements of your photos, including people, objects, and landscapes (such as mountains or sunsets). This makes it easy to search your photos later using terms like 'family photos' or 'beach.' That said, Apple Photos has a slight edge over Google Photos in terms of image editing. You can easily open any photo in Apple's Photos app, which has relatively advanced touch-up options. Google Photos offers auto-correction and some basic color and light adjustments, but that's it. With Google Photos, you can invite anyone with a Google account to join a shared album. The album owner can customize permissions for others to control whether they can add photos to the album or comment. It's also easy to share photos to third-party apps like Facebook and X. iCloud enables you to create shared Apple Photos albums, but users must have an Apple ID to access them. This can be a barrier for people without an Apple device. In addition, you can't customize permissions in an iCloud album. Anyone invited can add or delete photos, and all content stored in the album counts against the album owner's storage limit. Both apps also enable you to share photos and albums with a link. The recipient doesn't need a Google or Apple account. Overall, Google Photos is a better choice than iCloud for anyone who isn't fully locked into the Apple ecosystem. It's also better if you have friends or family who aren't Apple users since iCloud makes it challenging to share photos with them. Interested in more photo storage options? Check out our guide to the best cloud storage solutions. Yes, Google Photos has an iOS app that enables iPhone users to easily back up and sync their photos. iCloud offers 5GB of free storage, but users can upgrade to a paid plan for additional space. Yes, you can manually download photos from iCloud and upload them to Google Photos. Apple also offers a tool to transfer images directly to Google Photos. iCloud is primarily designed for Apple devices, but you can access it via a web browser on Windows or Android (with limited functionality).

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store