logo
#

Latest news with #Claude4

Claude 4 vs ChatGPT: Which AI assistant is right for you?
Claude 4 vs ChatGPT: Which AI assistant is right for you?

Tom's Guide

time13 hours ago

  • Business
  • Tom's Guide

Claude 4 vs ChatGPT: Which AI assistant is right for you?

Not all chatbots are created equal. OpenAI's ChatGPT and Anthropic's Claude are two of the most capable AI assistants available today, but they're built with different strengths, personalities and design philosophies. Whether you're looking for a creative collaborator, a coding partner or an all-in-one productivity tool, knowing how they compare can help you choose the right one for your style and needs. I've been testing AI chatbots for years, so let me break down the ChatGPT vs Claude differences that you need to know. While many users are eagerly anticipating the release of ChatGPT-5, OpenAI's current flagship model is GPT-4o (short for 'Omni'), and the company's most advanced base model to date. It supports multimodal input and output, meaning it can understand and respond to text, image, and audio, and it enables real-time interactions, including natural voice conversations and image generation. Anthropic's Claude 4, released in May 2025, comes in two versions. Claude Sonnet 4 is the free, fast, general-purpose model built for everyday use, while Claude Opus 4 is a premium model designed for deep reasoning, advanced coding tasks and large-scale context handling. It's built to power more complex workflows and long-form problem-solving. ChatGPT is available on the web, iOS and Android, and it's tightly integrated into Microsoft's ecosystem through Copilot in Bing, Word, Excel and other Office apps. Get instant access to breaking news, the hottest reviews, great deals and helpful tips. Claude 4, meanwhile, is accessible via on the web, iOS and Android. Developers and enterprise users can also access Claude via the Anthropic API, as well as through platforms like GitHub Copilot, AWS Bedrock and Google Cloud's Vertex AI. ChatGPT, powered by GPT-4o, offers strong conversational memory and supports multimodal input, including text, images, audio, file uploads and code execution. Users can personalize interactions with memory settings and even build custom GPTs for specific tasks. Although Claude does not have a memory feature, Claude 4 supports up to 200,000 tokens of context, allowing it to handle and retain significantly more information in a single conversation; the equivalent of more than 500 pages of text. ChatGPT offers fast, responsive performance and is particularly strong when paired with tools like Microsoft Copilot. Its reasoning abilities are enhanced by plugins, file uploads, and its new Agent feature, which can browse the web, execute tasks and automate workflows across apps. Claude Sonnet is similarly fast and excels at maintaining depth and continuity in conversations, making it a strong writing partner or analytical assistant. Claude Opus takes things further with support for multi-hour workflows, legal reasoning, code explanation and complex research tasks. On industry benchmarks like SWE-bench, Claude Opus ranks among the top performers for software engineering use cases. When it comes to multimodal support, ChatGPT offers more with in-chat image generation, visual context and real-time voice conversations with memory. Claude supports image input and visual understanding, meaning users can upload an image and the chatbot will assist with comprehension, but it cannot generate images. Also, Claude does not offer real-time conversational chat. Claude 4, especially in its Sonnet form, stands out for its natural, thoughtful tone. It often feels more like a helpful assistant offering clear explanations, thoughtful suggestions and a calm, conversational flow. If you're writing, brainstorming or trying to make sense of something complex, Claude tends to feel more human in how it is professional, polished and highly adaptable, with the ability to shift tones based on a user's preference or task. It is arguably more chatty than Claude, which appeals to some users more than others. ChatGPT supports a broad set of integrations including custom GPTs, an extensive plugin ecosystem, and Microsoft Copilot. Its new ChatGPT Agent feature (currently in early access) allows it to take real action on your behalf: browsing the web, interacting with forms, generating presentations and more all within a secure sandboxed environment. Claude 4 doesn't yet offer a native agent feature through its user interface, but it does support tool use and extended workflows through its API. It also integrates with GitHub Copilot, AWS Bedrock and other developer platforms, making it a strong option for those building their own agentic systems. If you're looking for an all-in-one AI assistant, ChatGPT is the more versatile choice. It can speak, see, listen, browse, generate images, analyze documents and automate tasks. Its voice mode feels remarkably human and is useful for a hands-free experience. The chatbot's ability to create and understand images adds another layer of functionality that Claude doesn't natively support. ChatGPT also integrates seamlessly with Microsoft products and lets users build custom GPTs tailored to their specific needs. And for power users, the new ChatGPT Agent adds another level of autonomy and actionability that's unmatched in most other consumer-facing AI tools. Claude 4 shines in natural language understanding and thoughtful conversation. It's ideal for users who prioritize clarity, warmth and coherence in writing, editing and reasoning. Whether you're crafting long-form content or analyzing dense documents, Claude delivers consistently thoughtful and easy-to-follow responses. Its 200K-token context window makes it the better choice for handling long inputs like research papers or legal contracts. Claude Opus, in particular, is built for high-stakes reasoning tasks and delivers top-tier performance for technical, legal and academic work. Even better, Claude Sonnet 4 is free, making it an excellent entry point for users who want fast, capable AI without paying for a subscription. Both ChatGPT and Claude are chatbots that I use every day. But they serve different goals. ChatGPT is my go-to for versatility, real-time interaction and multimodal tasks while Claude is what I use most often for summarizing, reasoning, deep context and analytics. Knowing what each excels at can help you choose the right assistant, or mix and match based on what you're trying to accomplish. Which chatbot do you prefer and why? Let me know in the comments. Follow Tom's Guide on Google News to get our up-to-date news, how-tos, and reviews in your feeds. Make sure to click the Follow button.

Top AI Firms Fall Short on Safety, New Studies Find
Top AI Firms Fall Short on Safety, New Studies Find

Time​ Magazine

time7 days ago

  • Business
  • Time​ Magazine

Top AI Firms Fall Short on Safety, New Studies Find

The world's leading AI companies have 'unacceptable' levels of risk management, and a 'striking lack of commitment to many areas of safety,' according to two new studies published Thursday. The risks of even today's AI—by the admission of many top companies themselves—could include AI helping bad actors carry out cyberattacks or create bioweapons. Future AI models, top scientists worry, could escape human control altogether. The studies were carried out by the nonprofits SaferAI and the Future of Life Institute (FLI). Each was the second of its kind, in what the groups hope will be a running series that incentivizes top AI companies to improve their practices. 'We want to make it really easy for people to see who is not just talking the talk, but who is also walking the walk,' says Max Tegmark, president of the FLI. Read More: Some Top AI Labs Have 'Very Weak' Risk Management, Study Finds SaferAI assessed top AI companies' risk management protocols (also known as responsible scaling policies) to score each company on its approach to identifying and mitigating AI risks. No AI company scored better than 'weak' in SaferAI's assessment of their risk management maturity. The highest scorer was Anthropic (35%), followed by OpenAI (33%), Meta (22%), and Google DeepMind (20%). Elon Musk's xAI scored 18%. Two companies, Anthropic and Google DeepMind, received lower scores than the first time the study was carried out, in October 2024. The result means that OpenAI has overtaken Google as second place in SaferAI's ratings. Siméon Campos, founder of SaferAI, said Google scored comparatively low despite doing some good safety research, because the company makes few solid commitments in its policies. The company also released a frontier model earlier this year, Gemini 2.5, without sharing safety information—in what Campos called an 'egregious failure.' A spokesperson for Google DeepMind told TIME: 'We are committed to developing AI safely and securely to benefit society. AI safety measures encompass a wide spectrum of potential mitigations. These recent reports don't take into account all of Google DeepMind's AI safety efforts, nor all of the industry benchmarks. Our comprehensive approach to AI safety and security extends well beyond what's captured.' Anthropic's score also declined since SaferAI's last survey in October. This was due in part to changes the company made to its responsible scaling policy days before the release of Claude 4 models, which saw Anthropic remove its commitments to tackle insider threats by the time it released models of that caliber. 'That's very bad process,' Campos says. Anthropic did not immediately respond to a request for comment. The study's authors also said that its methodology had become more detailed since last October, which accounts for some of the differences in scoring. The companies that improved their scores the most were xAI, which scored 18% compared to 0% in October; and Meta, which scored 22% compared to its previous score of 14%. The FLI's study was broader—looking not only at risk management practices, but also companies' approaches to current harms, existential safety, governance, and information sharing. A panel of six independent experts scored each company based on a review of publicly available material such as policies, research papers, and news reports, together with additional nonpublic data that companies were given the opportunity to provide. The highest grade was scored by Anthropic (a C plus). OpenAI scored a C, and Google scored a C minus. (xAI and Meta both scored D.) However, in FLI's scores for each company's approach to 'existential safety,' every company scored D or below. 'They're all saying: we want to build superintelligent machines that can outsmart humans in every which way, and nonetheless, they don't have a plan for how they're going to control this stuff,' Tegmark says.

Amazon-backed Anthropic rolls out Claude AI for financial services
Amazon-backed Anthropic rolls out Claude AI for financial services

CNBC

time15-07-2025

  • Business
  • CNBC

Amazon-backed Anthropic rolls out Claude AI for financial services

Anthropic on Tuesday announced Claude tools for financial services as part of the artificial intelligence startup's latest attempt to court enterprise customers. The so-called Financial Analysis Solution can help financial professionals make investment decisions, analyze markets and conduct research, Anthropic said. It includes the company's Claude 4 models, Claude Code and Claude for Enterprise with expanded usage limits, implementation support and other features. "Claude provides the complete platform for financial AI—from immediate deployment to custom development," Anthropic said in a release. As part of its new Financial Analysis Solution, Claude will get real-time access to financial information through data providers like Box, PitchBook, Databricks, S&P Global and Snowflake. Anthropic said many of these integrations are available on Tuesday, with more to come. Anthropic's Financial Analysis Solution and Claude for Enterprise are available on AWS Marketplace. The company said Google Cloud Marketplace availability is coming soon. The Amazon-backed AI startup was founded by former OpenAI research executives, and its debut product was the Claude chatbot in March 2023. Anthropic announced its latest models, Claude Opus 4 and Claude Sonnet 4, in May, roughly nine months after it rolled out its business product called Claude for Enterprise. Anthropic's Claude models and AI assistant have exploded in popularity as more businesses work to incorporate generative AI across marketing, sales and other customer service functions. As of March, Anthropic closed its latest funding round at a $61.5 billion post-money valuation.

Claude vs Grok 4 App Build Test : Which AI Builds Your App Faster & More Efficiently
Claude vs Grok 4 App Build Test : Which AI Builds Your App Faster & More Efficiently

Geeky Gadgets

time13-07-2025

  • Business
  • Geeky Gadgets

Claude vs Grok 4 App Build Test : Which AI Builds Your App Faster & More Efficiently

What if the future of app development wasn't just about coding skills but choosing the right AI collaborator? In a rapidly evolving tech landscape, the battle between AI models like Claude 4 and Grok 4 is reshaping how developers approach their craft. Imagine building a complex to-do app in record time, with one AI seamlessly handling tasks while another stumbles over integration issues. This isn't just a hypothetical scenario—it's the reality revealed in a recent head-to-head evaluation of these two advanced models. While both promise to transform development workflows, their performance and reliability tell two very different stories. In this comparison, Rob Shocks uncovers how Claude 4 and Grok 4 stack up in critical areas like autonomous task execution, error handling, and integration stability. Whether you're a seasoned developer or exploring AI-powered tools for the first time, this perspective will illuminate the strengths and limitations of each model. From Grok 4's ambitious but inconsistent performance to Claude 4's polished and dependable efficiency, the findings offer valuable insights for anyone seeking to streamline app development. As you read on, consider this: is it better to bet on potential or proven reliability? Claude 4 vs Grok 4 Grok 4: A Promising Newcomer with Challenges Grok 4, developed by Elon Musk's xAI, represents a new entrant in the competitive AI landscape. It features a 128,000-token context window, allowing it to process extensive inputs, and is offered via subscription at $300 annually or $30 monthly. Integrated into Cursor, Grok 4 aims to simplify development tasks by using its advanced processing capabilities. Despite its potential, Grok 4 encountered several challenges during testing. Its integration stability proved inconsistent, often requiring manual intervention to complete tasks. This lack of reliability in managing complex workflows limits its effectiveness in demanding development scenarios. While Grok 4 shows promise, its current iteration struggles to meet the expectations of developers seeking seamless and autonomous solutions. Claude 4: A Mature and Reliable AI Model Claude 4, in contrast, demonstrated consistent and robust performance throughout the evaluation. Renowned for its ability to autonomously execute tasks, it required minimal prompting and excelled in error resolution. Its seamless integration with modern tech stacks and its efficiency in handling complex workflows make it a standout choice for app developers. Claude 4's maturity as an AI model was evident in its ability to deliver results with minimal disruptions, reinforcing its reputation as a reliable tool for development tasks. Claude 4 vs Grok 4 : Comprehensive AI App Development Comparison Watch this video on YouTube. Below are more guides on AI-powered app development from our extensive range of articles. Testing the Models: Building a To-Do App The evaluation involved a practical test: creating an AI-powered to-do app capable of generating task plans. The tech stack used for this project included React, Prisma, and Neon DB, all of which are widely used in modern app development. Both Claude 4 and Grok 4 relied on Cursor, a tool designed to assist in generating and executing development tasks. The results highlighted key differences in their performance: Claude 4: Completed the app with fewer prompts, showcasing superior error-handling capabilities. It required minimal manual adjustments, streamlining the development process. Completed the app with fewer prompts, showcasing superior error-handling capabilities. It required minimal manual adjustments, streamlining the development process. Grok 4: Required additional troubleshooting and manual intervention, which slowed progress and exposed its limitations in managing tasks autonomously. This test underscored Claude 4's ability to handle complex workflows efficiently, while Grok 4's performance revealed areas that need improvement. Cursor Integration: A Crucial Component Cursor played a pivotal role in the evaluation by structuring task execution and guiding workflows. Both AI models depended on Cursor to generate to-do lists and manage development processes. However, their ability to integrate with Cursor varied significantly: Claude 4: Used Cursor effectively, maintaining stability and efficiency throughout the development process. Used Cursor effectively, maintaining stability and efficiency throughout the development process. Grok 4: Struggled with integration stability, leading to interruptions and inefficiencies that hindered its overall performance. The disparity in their performance with Cursor further highlighted Claude 4's reliability and Grok 4's need for refinement. Key Performance Insights The comparison between Claude 4 and Grok 4 revealed several critical insights into their capabilities: Autonomous Task Execution: Claude 4 excelled at completing tasks with minimal input, reducing the need for manual intervention and improving efficiency. Claude 4 excelled at completing tasks with minimal input, reducing the need for manual intervention and improving efficiency. Error Handling: Claude 4 demonstrated superior error resolution, making sure a smoother and more reliable development process. Claude 4 demonstrated superior error resolution, making sure a smoother and more reliable development process. Integration Stability: Grok 4 faced challenges with tool integration, which negatively impacted its performance and usability. Grok 4 faced challenges with tool integration, which negatively impacted its performance and usability. Potential for Growth: While Grok 4 shows promise, it requires updates and refinements to match Claude 4's efficiency and reliability. These insights emphasize the importance of selecting an AI model that aligns with the specific needs of app development projects. Looking Ahead: The Future of AI Models in Development As AI technology continues to advance, both Claude 4 and Grok 4 have the potential to evolve further. Grok 4's upcoming code-optimized version may address its current shortcomings, particularly in integration stability and autonomous task execution. Such improvements could make it a more competitive option in the future. Additionally, tools like Cursor, which are designed to be AI-agnostic, provide developers with the flexibility to integrate multiple models into their workflows. This adaptability allows developers to use the strengths of different AI models while mitigating their weaknesses. As the field of AI-powered app development progresses, developers can expect even more sophisticated tools and models that enhance productivity and innovation. Media Credit: Rob Shocks Filed Under: AI, Guides Latest Geeky Gadgets Deals Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.

Why responsible AI is key for building better, brighter future for the world
Why responsible AI is key for building better, brighter future for the world

Hans India

time06-07-2025

  • Business
  • Hans India

Why responsible AI is key for building better, brighter future for the world

Artificial intelligence has become the defining technology of our era, with recent years marking remarkable milestones in AI development. As the debate around AI technology heats up, people are actively discussing its usage, impact, and transformative potential across industries. However, amidst all the excitement and concern, there's notably less conversation about the responsibility that comes with this powerful technology. This gap in discourse presents both a challenge and an opportunity to shape how we approach AI governance. Breakthrough models like OpenAI's O3 achieved near-human performance on complex reasoning tasks, while AI applications continue to expand across industries in ways that seemed impossible just a few years ago. Yet as we witness this extraordinary progress, we find ourselves at a crucial juncture where thoughtful governance can shape AI's trajectory for maximum human benefit. The question isn't whether AI will transform our world - it's how we'll guide that transformation. Historically, humanity has taken a reactive approach to governing powerful technologies. The atomic bombing and destruction of two cities during the World War II ultimately paved the way for the formation of the United Nations - a response born from tragedy rather than foresight. Rather than repeating this pattern of waiting for catastrophe to drive action, we have an unprecedented opportunity with AI to be proactive. This is the right time for us to debate and establish a world governing council for responsible AI, creating international governance frameworks while embracing AI's transformative potential. This proactive approach represents a fundamental shift from reactive crisis management to thoughtful stewardship of emerging technology. Frankly speaking, recent developments highlight both AI's transformative power and the importance of responsible deployment. Google's Gemini 2.0 and Anthropic's Claude 4 demonstrate unprecedented capabilities in autonomous planning and execution, opening new possibilities for scientific research and problem-solving. However, this rapid advancement brings important considerations that deserve our attention. Workforce transitions affecting 14 per cent of workers require thoughtful retraining programs and support systems. Meanwhile, AI's growing energy footprint presents ongoing opportunities for innovation in sustainable computing and renewable energy integration. Artificial intelligence and ethical considerations The responsible development of AI requires addressing important ethical considerations with the same rigor we apply to technology itself. Workforce transition support becomes crucial as industries evolve, creating opportunities for new skilled roles while ensuring affected workers receive retraining and meaningful career paths. Environmental stewardship is equally important. While major tech companies have seen significant emissions increase largely due to AI infrastructure expansion, this challenge drives innovation in green computing and renewable energy solutions. Additionally, securing personal data becomes paramount as AI systems become more sophisticated, with recent studies showing 77 per cent of businesses experiencing AI-related data challenges, highlighting the need for robust privacy frameworks that protect individuals while enabling beneficial AI applications. Current data protection frameworks like Europe's GDPR provide excellent foundations, but AI's unique characteristics require evolved approaches. Unlike traditional data processing, AI systems integrate information in ways that make conventional privacy protections challenging to implement. This creates opportunities to develop next-generation privacy frameworks that address AI's permanent data integration while maintaining cross-border collaboration. The goal is to create comprehensive global standards that protect individual privacy while enabling AI's beneficial applications across borders. Understanding how personal data interacts with AI systems empowers users to make informed choices about their digital interactions. When individuals interact with ChatGPT or similar systems, they contribute to model training, creating opportunities for both personalized assistance and privacy considerations. This presents an opportunity to develop AI systems that provide personalized benefits while maintaining strong privacy protections through advanced techniques like differential privacy and federated learning. AI's extraordinary potential However, the story doesn't end with challenges - it begins with extraordinary potential. AI's capacity for human advancement continues to inspire remarkable breakthroughs. NASA's Perseverance rover uses AI for autonomous Mars exploration, opening new frontiers for space discovery. AI-powered stroke detection systems prove twice as accurate as traditional methods, saving lives through early intervention. The AI healthcare market has grown to over $32 billion, with applications reducing emergency room wait times and accelerating drug discovery for previously incurable diseases. These beneficial applications - from climate monitoring to space exploration - demonstrate AI's extraordinary capacity to address humanity's greatest challenges. The 57 countries that signed the Paris Declaration on AI governance last year recognize this tremendous opportunity. Building on this foundation, we can establish a global governing body that fosters international cooperation while ensuring AI development serves human flourishing. Like international frameworks for nuclear technology that enable both peaceful energy production and prevent harmful applications, AI governance can maximize benefits while addressing potential risks through shared standards and collaborative oversight. No doubt, the world needs responsible AI that can enhance our quality of life in all spheres and spaces. And the opportunity before us is clear: proactive governance now can unlock AI's full potential for a better tomorrow. That's the bottom line. (Krishna Kumar is a technology explorer & strategist based in Austin, Texas in the US. Rakshitha Reddy is AI developer based in Atlanta, US)

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store