logo
#

Latest news with #GusMartins

New Google AI tool translates sign language into text, currently in testing phase with launch by year-end
New Google AI tool translates sign language into text, currently in testing phase with launch by year-end

India Today

time3 days ago

  • Business
  • India Today

New Google AI tool translates sign language into text, currently in testing phase with launch by year-end

Sign language is essential for many people who have speech impairment. They use it to communicate with people around them but among the regular not many understand it. Now, AI is going to help here as well. Google is working on a AI model called SignGemma that will translate sign language into text. The company says this is its most capable artificial intelligence model to date, designed to translate sign language into spoken text. This new AI model is currently in its testing phase, and is slated for public launch by the end of the first unveiled SignGemma during the keynote at Google I/O, where Gemma Product Manager Gus Martins described it as the company's 'most capable sign language understanding model ever.' Martins noted that, unlike previous attempts at sign language translation, SignGemma stands out for its open model approach and its focus on delivering accurate, real-time translations to users. While the tool is trained to handle various sign languages, Google says the model currently performs best with American Sign Language (ASL) and English.'We're thrilled to announce SignGemma, our groundbreaking open model for sign language understanding, set for release later this year,' Martins said. 'It's the most capable sign language understanding model ever, and we can't wait for developers and Deaf and hard-of-hearing communities to take this foundation and build with it.' Google highlighted that with this tool, the company aims to bridge communication gaps for millions of Deaf and hard-of-hearing individuals to ensure the tool is both effective and respectful of its user base, Google is taking a collaborative approach to its development. The company has extended an open invitation to developers, researchers, and members of the global Deaf and Hard of Hearing communities to participate in early testing and provide feedback."We're thrilled to announce SignGemma, our groundbreaking open model for sign language understanding," reads the official post from DeepMind on X. "Your unique experiences, insights, and needs are crucial as we prepare for launch and beyond, to make SignGemma as useful and impactful as possible."The introduction of SignGemma comes at a time when Google is heavily focused on expanding its AI portfolio. At Google I/O 2025, accessibility took centre stage with the announcement of several new AI-powered features designed to make technology more inclusive for everyone. One of the highlights was the expansion of Gemini AI's integration with Android's TalkBack, which will now provide users AI-generated descriptions for images and allow them to ask follow-up questions about what's on their screen. Google has also introduced updates to Chrome, including automatic Optical Character Recognition (OCR) for scanned PDFs, enabling screen reader users to access, search, and interact with text in documents that were previously inaccessible. For students, on Chromebooks a new accessibility tool called Face Control allows users to control their device with facial gestures and head movements.

Google Unveils SignGemma: AI Tool to Translate Sign Language into Text by Year-End
Google Unveils SignGemma: AI Tool to Translate Sign Language into Text by Year-End

Hans India

time3 days ago

  • Business
  • Hans India

Google Unveils SignGemma: AI Tool to Translate Sign Language into Text by Year-End

At Google I/O 2025, the tech giant introduced SignGemma, a powerful AI model designed to translate sign language into spoken text. Currently in its testing phase, this tool is available to developers and selected users, with a broader rollout expected by the end of the year. For millions of Deaf and hard-of-hearing individuals around the world, sign language is a vital means of communication. However, it often presents barriers in daily interactions with those unfamiliar with it. Google's new AI initiative, SignGemma, aims to change that by offering real-time sign language-to-text translations, improving accessibility and inclusion on a global scale. Described as Google's 'most capable sign language understanding model ever,' SignGemma was unveiled by Gemma Product Manager Gus Martins during the keynote. According to Martins, the project stands apart from previous attempts thanks to its open model framework and ability to deliver real-time, accurate translations. 'We're thrilled to announce SignGemma, our groundbreaking open model for sign language understanding, set for release later this year,' Martins said. 'It's the most capable sign language understanding model ever, and we can't wait for developers and Deaf and hard-of-hearing communities to take this foundation and build with it.' At present, SignGemma is most accurate when translating American Sign Language (ASL) into English. However, Google has stated that the model is trained to support a range of sign languages and plans to expand its capabilities over time. The launch of SignGemma is part of a broader push by Google to prioritise accessibility in AI technology. At this year's I/O conference, the company announced several updates focused on inclusivity, including enhanced AI integration in Android's TalkBack feature. Users will now receive AI-generated descriptions of images and be able to ask follow-up questions about what's on their screen, making the Android experience more intuitive for visually impaired users. Additionally, Google has rolled out updates to Chrome, such as automatic Optical Character Recognition (OCR) for scanned PDFs. This makes previously inaccessible documents readable and searchable for screen reader users. On Chromebooks, a new feature called Face Control enables users to navigate their device using facial expressions and head gestures—another step forward in Google's mission to empower every user. To ensure SignGemma is both useful and respectful, Google is adopting a collaborative development approach. The company is actively inviting developers, researchers, and members of the global Deaf and hard-of-hearing communities to test the tool and share feedback. 'We're thrilled to announce SignGemma, our groundbreaking open model for sign language understanding,' read an official post from DeepMind on X. 'Your unique experiences, insights, and needs are crucial as we prepare for launch and beyond, to make SignGemma as useful and impactful as possible.' With SignGemma, Google is not just expanding its AI capabilities—it's building a bridge between the hearing and Deaf communities. As it nears public release, the tool stands to transform communication and redefine accessibility in the digital age.

The latest Google Gemma AI model can run on phones
The latest Google Gemma AI model can run on phones

Yahoo

time20-05-2025

  • Business
  • Yahoo

The latest Google Gemma AI model can run on phones

Google's family of "open" AI models, Gemma, is growing. During Google I/O 2025 on Tuesday, Google took the wraps off Gemma 3n, a model designed to run "smoothly" on phones, laptops, and tablets. Available in preview starting Tuesday, Gemma 3n can handle audio, text, images, and videos, according to Google. Models efficient enough to run offline and without the need for computing in the cloud have gained steam in the AI community in recent years. Not only are they cheaper to use than large models, but they preserve privacy by eliminating the need to transfer data to a remote data center. During a keynote at I/O, Gemma Product Manager Gus Martins said that Gemma 3n can run on devices with less than 2GB of RAM. "Gemma 3n shares the same architecture as Gemini Nano, and is and is engineered for incredible performance," he added. In addition to Gemma 3n, Google is releasing MedGemma through its Health AI Developer Foundations program. According to the company, MedGemma is its most capable open model for analyzing health-related text and images. "MedGemma [is] our [...] collection of open models for multimodal [health] text and image understanding," Martins said. "MedGemma works great across a range of image and text applications, so that developers [...] can adapt the models for their own health apps." Also on the horizon is SignGemma, an open model to translate sign language into spoken-language text. Google says that SignGemma will enable developers to create new apps and integrations for deaf and hard-of-hearing users. "SignGemma is a new family of models trained to translate sign language to spoken-language text, but it's best at American Sign Language and English," Martins said. "It's the most capable sign language understanding model ever, and we can't wait for you — developers and deaf and hard-of-hearing communities — to take this foundation and build with it." Worth noting is that Gemma has been criticized for its custom, non-standard licensing terms, which some developers say have made using the models commercially a risky proposition. That hasn't dissuaded developers from downloading Gemma models tens of millions of times collectively, however. Updated 2:40 p.m. Pacific: Added several quotes from Gemma Product Manager Gus Martins. This article originally appeared on TechCrunch at

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store