Latest news with #GoogleAIStudio


Time of India
23-05-2025
- Business
- Time of India
AI may turn legacy codebases into liabilities: Zoho founder Sridhar Vembu
The foundations of modern software development may be under threat as generative artificial intelligence evolves rapidly, according to Sridhar Vembu , founder of software-as-a-service company (SaaS) Zoho . In a series of posts on social media platform X, Vembu said that large, existing codebases, long considered prized assets, could soon become burdens. 'If AI makes us 100x productive, why not rewrite the whole thing with AI help?' Vembu asked, highlighting a potential shift in enterprise software strategy. With generative AI tools now able to write new code at high velocity, companies may find it more efficient to start from scratch rather than maintain legacy systems. — svembu (@svembu) This view counters a common concern in the tech industry that AI is currently poor at navigating complex, existing code structures. Vembu acknowledged the limitation but suggested it will not last. 'It is not there yet, but perhaps not for long,' he said. 'Viewed that way, large existing codebases may no longer be assets,' he said. 'They may be liabilities.' He pointed to AI-powered tools such as Zoho Creator and Google AI Studio, which are now capable of building standard CRUD (create, read, update, delete) applications with minimal human input. Having recently stepped back from his role as chief executive officer to focus on research, Vembu has been vocal about AI's growing role in software development. He has predicted that generative AI could eventually handle up to 90% of coding tasks, especially boilerplate code that consumes much of a developer's time. However, he noted that essential complexity, such as innovative and creative work, would still require human input. Discover the stories of your interest Blockchain 5 Stories Cyber-safety 7 Stories Fintech 9 Stories E-comm 9 Stories ML 8 Stories Edtech 6 Stories The Zoho founder warned developers not to count on high salaries or long-term job security. As AI reshapes the software development landscape, he urged the industry to stay vigilant or risk becoming obsolete.


Techday NZ
22-05-2025
- Business
- Techday NZ
Google unveils Gemini 2.5 upgrades for reasoning & security
Google has provided a series of updates to its Gemini 2.5 model series, with enhancements spanning advanced reasoning, developer capabilities and security safeguards. The company reported that Gemini 2.5 Pro is now the leading model on the WebDev Arena coding leaderboard, holding an ELO score of 1415. It also leads across all leaderboards in LMArena, a platform that measures human preferences in multiple dimensions. Additionally, Gemini 2.5 Pro's 1 million-token context window was highlighted as supporting strong long context and video understanding performance. Integration with LearnLM, a family of models developed with educational experts, resulted in Gemini 2.5 Pro apparently becoming the foremost model for learning. According to Google, in direct comparisons focusing on pedagogy and effectiveness, Gemini 2.5 Pro was favoured by educators and experts over other models in a wide range of scenarios. The model outperformed others based on the five principles of learning science used in AI system design for education. Gemini 2.5 Pro introduced an experimental capability called Deep Think, which is being tested to enable enhanced reasoning by allowing the model to consider multiple hypotheses before responding. The company said, "2.5 Pro Deep Think gets an impressive score on 2025 USAMO, currently one of the hardest math benchmarks. It also leads on LiveCodeBench, a difficult benchmark for competition-level coding, and scores 84.0% on MMMU, which tests multimodal reasoning." Safety and evaluation measures are being emphasised with Deep Think. "Because we're defining the frontier with 2.5 Pro DeepThink, we're taking extra time to conduct more frontier safety evaluations and get further input from safety experts. As part of that, we're going to make it available to trusted testers via the Gemini API to get their feedback before making it widely available," the company reported. Google announced improvements to 2.5 Flash, describing it as the most efficient in the series, tailored for speed and cost efficiency. This version now reportedly uses 20-30% fewer tokens in evaluations and delivers improved performance across benchmarks for reasoning, multimodality, code, and long-context tasks. The updated 2.5 Flash is now available for preview in Google AI Studio, Vertex AI, and the Gemini app. New features have also been added to the Gemini 2.5 series. The Live API now offers a preview version supporting audio-visual input and native audio output. This is designed to create more natural and expressive conversational experiences. According to Google, "It also allows the user to steer its tone, accent and style of speaking. For example, you can tell the model to use a dramatic voice when telling a story. And it supports tool use, to be able to search on your behalf." Early features in this update include Affective Dialogue, where the model can detect and respond to emotions in a user's voice; Proactive Audio, which enables the model to ignore background conversations and determine when to respond; and enhanced reasoning in live API use. Multi-speaker support has also been introduced for text-to-speech capabilities, allowing audio generation with two distinct voices and support for over 24 languages, including seamless transitions between them. Project Mariner's computer use capabilities are being integrated into the Gemini API and Vertex AI, with multiple enterprises testing the tool. Google stated, "Companies like Automation Anywhere, UiPath, Browserbase, Autotab, The Interaction Company and Cartwheel are exploring its potential, and we're excited to roll it out more broadly for developers to experiment with this summer." On the security front, Gemini 2.5 includes advanced safeguards against indirect prompt injections, which involve malicious instructions embedded into retrieved data. According to disclosures, "Our new security approach helped significantly increase Gemini's protection rate against indirect prompt injection attacks during tool use, making Gemini 2.5 our most secure model family to date." Google is introducing new developer tools with thought summaries in the Gemini API and Vertex AI. These summaries convert the model's raw processing into structured formats with headers and action notes. Google stated, "We hope that with a more structured, streamlined format on the model's thinking process, developers and users will find the interactions with Gemini models easier to understand and debug." Additional features include thinking budgets for 2.5 Pro, allowing developers to control the model's computation resources to balance quality and speed. This can also completely disable the model's advanced reasoning capability if desired. Model Context Protocol (MCP) support has been added for SDK integration, aiming to enable easier development of agentic applications using both open-source and hosted tools. Google affirmed its intention to sustain research and development efforts as the Gemini 2.5 series evolves, stating, "We're always innovating on new approaches to improve our models and our developer experience, including making them more efficient and performant, and continuing to respond to developer feedback, so please keep it coming! We also continue to double down on the breadth and depth of our fundamental research — pushing the frontiers of Gemini's capabilities. More to come soon."


Techday NZ
22-05-2025
- Business
- Techday NZ
Google announces major Gemini AI upgrades & new dev tools
Google has unveiled a range of updates to its developer products, aimed at improving the process of building artificial intelligence applications. Mat Velloso, Vice President, AI / ML Developer at Google, stated, "We believe developers are the architects of the future. That's why Google I/O is our most anticipated event of the year, and a perfect moment to bring developers together and share our efforts for all the amazing builders out there. In that spirit, we updated Gemini 2.5 Pro Preview with even better coding capabilities a few weeks ago. Today, we're unveiling a new wave of announcements across our developer products, designed to make building transformative AI applications even better." The company introduced an enhanced version of its Gemini 2.5 Flash Preview, described as delivering improved performance on coding and complex reasoning tasks while optimising for speed and efficiency. This model now includes "thought summaries" to increase transparency in its decision-making process, and its forthcoming "thinking budgets" feature is intended to help developers manage costs and exercise more control over model outputs. Both Gemini 2.5 Flash versions and 2.5 Pro are available in preview within Google AI Studio and Vertex AI, with general availability for Flash expected in early June, followed by Pro. Among the new models announced is Gemma 3n, designed to function efficiently on personal devices such as phones, laptops, and tablets. Gemma 3n can process audio, text, image, and video inputs and is available for preview on Google AI Studio and Google AI Edge. Also introduced is Gemini Diffusion, a text model that reportedly generates outputs at five times the speed of Google's previous fastest model while maintaining coding performance. Access to Gemini Diffusion is currently by waitlist. The Lyria RealTime model was also detailed. This experimental interactive music generation tool allows users to create, control, and perform music in real time. Lyria RealTime can be accessed via the Gemini API and trialled through a starter application in Google AI Studio. Several additional variants of the Gemma model family were announced, targeting specific use cases. MedGemma is described as the company's most capable multimodal medical model to date, intended to support developers creating healthcare applications such as medical image analysis. MedGemma is available now via the Health AI Developer Foundations programme. Another upcoming model, SignGemma, is designed to translate sign languages into spoken language text, currently optimised for American Sign Language to English. Google is soliciting feedback from the community to guide further development of SignGemma. Google outlined new features intended to facilitate the development of AI applications. A new, more agentic version of Colab will enable users to instruct the tool in plain language, with Colab subsequently taking actions such as fixing errors and transforming code automatically. Meanwhile, Gemini Code Assist, Google's free AI-coding assistant, and its associated code review agent for GitHub, are now generally available to all developers. These tools are now powered by Gemini 2.5 and will soon offer a two million token context window for standard and enterprise users on Vertex AI. Firebase Studio was presented as a new cloud-based workspace supporting rapid development of AI applications. Notably, Firebase Studio now integrates with Figma via a plugin, supporting the transition from design to app. It can also automatically detect and provision necessary back-end resources. Jules, another tool now generally available, is an asynchronous coding agent that can manage bug backlogs, handle multiple tasks, and develop new features, working directly with GitHub repositories and creating pull requests for project integration. A new offering called Stitch was also announced, designed to generate frontend code and user interface designs from natural language descriptions or image prompts, supporting iterative and conversational design adjustments with easy export to web or design platforms. For those developing with the Gemini API, updates to Google AI Studio were showcased, including native integration with Gemini 2.5 Pro and optimised use with the GenAI SDK for instant generation of web applications from input prompts spanning text, images, or videos. Developers will find new models for generative media alongside enhanced code editor support for prototyping. Additional technical features include proactive video and audio capabilities, affective dialogue responses, and advanced text-to-speech functions that enable control over voice style, accent, and pacing. The model updates also introduce asynchronous function calling to enable non-blocking operations and a Computer Use API that will allow applications to browse the web or utilise other software tools under user direction, initially available to trusted testers. The company is also rolling out URL context, an experimental tool for retrieving and analysing contextual information from web pages, and announcing support for the Model Context Protocol in the Gemini API and SDK, aiming to facilitate the use of a broader range of open-source developer tools.


Time of India
20-05-2025
- Business
- Time of India
Google I/O 2025: Gemini 2.5 Pro gets improved reasoning, audio features and multilingual support
At Google I/O 2025, the company announced new updates to its Gemini 2.5 model series adding more powerful reasoning, native audio output, security upgrades, and improved tools for developers. 'In March, we announced Gemini 2.5 Pro , our most intelligent model yet…Today, We're bringing new capabilities to 2.5 Pro and 2.5 Flash,' Google said, announcing the new updates. The upgraded Gemini 2.5 Pro model now tops performance charts, including WebDev Arena for coding and LMArena for human preference testing. It also features a 1 million-token context window, which allows it to handle longer inputs and video understanding tasks. Google said that thanks to LearnLM — a version of Gemini developed with educational experts — the model now leads in learning-related tasks as well. 'Educators and experts preferred Gemini 2.5 Pro over other models across a diverse range of scenarios,' the company said. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like Ótimas notícias para os cidadãos brasileiros! Leia mais Undo Native audio, emotional dialogue and multilingual support Google also introduced native audio output for a more natural AI experience. Gemini can now speak with different tones, accents, and styles — such as a dramatic voice when telling a story. It can also: Detect user emotions and respond accordingly (Affective Dialogue) Ignore background noise (Proactive Audio) Handle more complex voice tasks (Thinking in the Live API) The text-to-speech tool now supports multiple speakers and over 24 languages, and it can switch between languages mid-conversation. These features will be available later today through the Gemini API. New 'Deep Think' for complex tasks Google said that it is testing an enhanced reasoning mode called Deep Think, which helps Gemini consider multiple answers before responding. It's aimed at tough challenges like advanced math and programming. 'We're starting to test an enhanced reasoning mode called Deep Think,' the company said. 'We're taking extra time to conduct more frontier safety evaluations and get further input from safety experts.' Deep Think is already leading benchmarks like the 2025 USAMO (math), LiveCodeBench (coding), and MMMU (multimodal reasoning). Gemini 2.5 Flash gets faster and more efficient Gemini 2.5 Flash, the lightweight version of the model, now uses 20–30% fewer tokens while improving performance across reasoning, code, and multimodal tasks, the company announced. It is now available in the Gemini app, Google AI Studio , and Vertex AI. A general release of the updated model is expected in early June, with 2.5 Pro following soon after. AI Masterclass for Students. Upskill Young Ones Today!– Join Now


Geeky Gadgets
13-05-2025
- Geeky Gadgets
Unlock the Power of Google AI Studio : No Coding Required
Have you ever wondered how artificial intelligence could transform your daily tasks, even if you have zero coding experience? Imagine creating a professional-grade video from a simple text prompt or analyzing complex datasets with just a few clicks. Welcome to Google AI Studio, a platform that's redefining how we interact with AI. Designed for both beginners and seasoned developers, this tool offers everything from intuitive starter apps to advanced features like real-time collaboration and custom code execution. Whether you're a curious newcomer or a tech-savvy pro, Google AI Studio promises to make AI not just accessible, but downright exciting. In this beginner-friendly walkthrough by the AI Grid, you'll discover how to unlock the potential of Google AI Studio step by step. From exploring its versatile chat models to experimenting with tools like video generation and file analysis, this guide will show you how to harness AI for creativity, productivity, and problem-solving. But that's not all—along the way, you'll uncover tips for customizing AI behavior, using Google's ecosystem, and even using pre-built apps to simplify complex tasks. Ready to see how AI can fit seamlessly into your workflow? Let's explore the possibilities together. Overview of Google AI Studio Core Features of Google AI Studio Google AI Studio is built around a set of core functionalities that simplify AI adoption while offering advanced capabilities for technical users. These features are designed to cater to a wide range of tasks, making sure both accessibility and depth. 1. Chat Models The platform includes a variety of chat models tailored to specific tasks, such as reasoning, coding, and generating structured outputs. One of the standout options is the Gemini 2.5 Pro model, which excels at solving complex problems and programming tasks. Users can customize the model's behavior to suit their needs by: Defining a persona or tone: Use system instructions to set the model's tone and approach. Use system instructions to set the model's tone and approach. Adjusting response depth: Use 'thinking mode' and budget controls to refine the level of detail in responses. Use 'thinking mode' and budget controls to refine the level of detail in responses. Grounding responses: Use real-time Google Search to ensure outputs are accurate and up-to-date. Additionally, the platform supports structured outputs such as tables and checklists, making it easier to organize and present information. However, free-tier users should note that they are limited to 25 requests per day, which may influence how extensively the feature can be used. 2. File Analysis Google AI Studio enables users to analyze a wide range of file types, including images, videos, audio, and documents. This feature is particularly useful for extracting insights from large datasets or summarizing multimedia content. Key capabilities include: Video analysis: Generate transcripts and gain visual insights from uploaded videos. Generate transcripts and gain visual insights from uploaded videos. Sample media files: Explore the platform's capabilities without needing to upload personal data. The platform's long context window ensures detailed examination of files, making it a valuable tool for professionals working with complex or large-scale data. 3. Video Generation The video generation feature allows users to create videos from text prompts or uploaded images. This tool is highly customizable, allowing users to adjust key settings such as: Aspect ratio and duration: Tailor the video's dimensions and length to fit specific requirements. Tailor the video's dimensions and length to fit specific requirements. Negative prompts: Refine the output by specifying elements to exclude. Once the video is generated, it can be downloaded or saved directly to Google Drive, facilitating easy sharing and storage. This feature is particularly useful for content creators and educators looking to produce engaging visual materials. 4. Streaming with Gemini Live Gemini Live is the platform's real-time streaming feature, designed to enhance collaboration and support. It allows users to share their screens for tasks such as document reading or software navigation. This functionality is especially beneficial for team projects, remote assistance, or live demonstrations, making sure seamless communication and productivity. 5. Starter Apps For those new to AI, Google AI Studio offers pre-built apps that simplify common tasks. These starter apps are designed to provide an accessible entry point into the platform's capabilities. Examples include: Flashcard generation: Quickly create study materials tailored to specific topics. Quickly create study materials tailored to specific topics. Customizable templates: Use pre-designed templates as a foundation for building AI-powered applications. These tools make it easy for beginners to explore AI's potential without requiring extensive technical expertise, fostering creativity and experimentation. How To Use Google AI Studio : 2025 Beginners Guide Watch this video on YouTube. Here are more detailed guides and articles that you may find helpful on Google AI Studio. Advanced Tools for Developers Google AI Studio is not just for beginners; it also caters to developers and technical users with a range of advanced tools. These features are designed to enhance productivity and enable seamless integration with external systems, making the platform a powerful resource for professional use. Code Execution: Run scripts directly within the platform to generate outputs such as charts or data visualizations. Run scripts directly within the platform to generate outputs such as charts or data visualizations. Function Calling: Integrate AI with external tools or applications to extend its functionality and streamline workflows. Integrate AI with external tools or applications to extend its functionality and streamline workflows. Safety Settings: Customize content restrictions to align with specific use cases or organizational policies. Comparison Tools The platform includes tools for comparing different AI models side-by-side. This feature allows users to evaluate performance metrics and select the most suitable model for their specific tasks. By allowing informed decision-making, this tool ensures that users achieve optimal results in their projects. Prompt Gallery The prompt gallery is a curated collection of pre-designed prompts that showcase the platform's capabilities. These examples serve as inspiration for new use cases, helping users explore the potential of AI in innovative ways. The gallery is particularly useful for those seeking to expand their understanding of the platform's functionality. History and Accessibility Google AI Studio's history feature allows users to revisit past interactions. This capability is invaluable for referencing previous chats, building on earlier work, or refining ongoing projects. By providing easy access to past activities, the platform enhances workflow continuity and efficiency. Why Google AI Studio Stands Out Google AI Studio bridges the gap between AI beginners and experienced developers, offering a diverse range of tools that cater to both simple and complex applications. Its seamless integration into Google's ecosystem, combined with a focus on accessibility and customization, makes it a valuable resource for anyone looking to harness the power of artificial intelligence. Whether you are creating structured outputs, analyzing files, generating videos, or developing custom apps, Google AI Studio provides the tools and flexibility needed to succeed in the evolving AI landscape. Media Credit: TheAIGRID Filed Under: AI, Guides Latest Geeky Gadgets Deals Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.