
China's new AI agent Manus calls its own shots
is a senior writer at Future Perfect, Vox's effective altruism-inspired section on the world's biggest challenges. She explores wide-ranging topics like climate change, artificial intelligence, vaccine development, and factory farms, and also writes the Future Perfect newsletter.
Modern large language models are really good at a lot of tasks, like coding, essay writing, translation, and research. But there are still a lot of basic tasks, especially in the 'personal assistant' realm, that the most highly trained AIs in the world remain hopeless at.
You can't ask ChatGPT or Claude 'order me a burrito from Chipotle' and get one, let alone 'book me a train from New York to Philadelphia.' OpenAI and Anthropic both offer AIs that can view your screen, move your cursor, and do some things on your computer as if they were a person (through their 'Operator' and 'Computer Use' functions, respectively).
This story was first featured in the Future Perfect newsletter.
Sign up here to explore the big, complicated problems the world faces and the most efficient ways to solve them. Sent twice a week.
That such 'AI agents' sometimes work, sort of, is about the strongest thing you can say for them right now. (Disclosure: Vox Media is one of several publishers that has signed partnership agreements with OpenAI. One of Anthropic's early investors is James McClave, whose BEMC Foundation helps fund Future Perfect. Our reporting remains editorially independent.)
This week, China launched a competitor: the AI agent Manus. It produced a blizzard of glowing posts and testimonials from highly selected influencers, along with some impressive website demos.
Manus is invite-only (and while I submitted a request for the tool, it hasn't been granted), so it's hard to tell from the outside how representative these highly selected examples are. After a few days of Manus fervor, though, the bubble popped a little and some more moderate reviews started coming out.
Manus, the growing consensus holds, is worse than OpenAI's DeepResearch at research tasks; but better than Operator or Computer Use at personal assistant tasks. It's a step forward toward something important — AIs that can take action beyond the chatbot window — but it's not a shocking out-of-nowhere advance.
Perhaps most importantly, Manus's usefulness for you will be sharply limited if you don't trust a Chinese company you've never heard of with your payment information so it can book things on your behalf. And you probably shouldn't.
The agents are arriving
When I first wrote about the risks of powerful AI systems displacing or destroying humanity, one very reasonable question was this: How could an AI act against humanity, when they really don't act at all?
This reasoning is right, as far as current technology goes. Claude or ChatGPT, which just respond to user prompts and don't act independently in the world, can't execute on a long-term plan; everything they do is in response to a prompt, and almost all that action takes place within the chat window.
But AI was never going to remain as a purely responsive tool simply because there is so much potential for profit in agents. People have been trying for years to create AIs that are built out of language models, but which make decisions independently, so that people can relate to them more like an employee or an assistant than like a chatbot.
Generally, this works by creating a small internal hierarchy of language models, like a little AI company. One of the models is carefully prompted and in some cases fine-tuned to do large-scale planning. It comes up with a long-term plan, which it delegates to other language models. Various sub-agents check their results and change approaches when one sub-agent fails or reports problems.
The concept is simple, and Manus is far from the first to try it. You may remember that last year we had Devin, which was marketed as a junior software engineering employee. It was an AI agent that you interacted with via Slack to give tasks, and which it would then work on achieving without further human input except, ideally, of the kind a human employee might occasionally need.
The economic incentives to build something like Manus or Devin are overwhelming. Tech companies pay junior software engineers as much as $100,000 a year or more. An AI that could actually provide that value would be stunningly profitable. Travel agents, curriculum developers, personal assistants — these are all fairly well-paid jobs, and an AI agent could in principle be able to do the work at a fraction of the cost, without needing breaks, benefits or vacations.
But Devin turned out to be overhyped, and didn't work well enough for the market it was aiming at. It's too soon to say whether Manus represents enough of an advance to have real commercial staying power, or whether, like Devin, its reach will exceed its grasp.
I'll say that it appears Manus works better than anything that has come before. But just working better isn't enough — to trust an AI to spend your money or plan your vacation, you'll need extremely high reliability. As long as Manus remains tightly limited in availability, it's hard to say if it will be able to offer that. My best guess is that AI agents that seamlessly work are still a year or two away — but only a year or two.
The China angle
Manus isn't just the latest and greatest attempt at an AI agent.
It is also the product of a Chinese company, and much of the coverage has dwelled on the Chinese angle. Manus is clearly proof that Chinese companies aren't just imitating what's being built here in America, as they've often been accused of doing, but improving on it.
That conclusion shouldn't be shocking to anyone who is aware of China's intense interest in AI. It also raises questions about whether we will be thoughtful about exporting all of our personal and financial data to Chinese companies that are not meaningfully accountable to US regulators or US law.
Installing Manus on your computer gives it a lot of access to your computer — it's hard for me to figure out the exact limits on its access or the security of its sandbox when I can't install it myself.
One thing we've learned in digital privacy debates is that a lot of people will do this without thinking about the implications if they feel Manus offers them enough convenience. And as the TikTok fight made clear, once millions of Americans love an app, the government will face a steep uphill battle in trying to restrict it or oblige it to follow data privacy rules.
But there are also clear reasons Manus came out of a Chinese company and not out of, say, Meta — and they're the very reasons we might prefer to use AI agents from Meta. Meta is subject to US liability law. If its agent makes a mistake and spends all your money on website hosting, or if it steals your Bitcoin or uploads your private photos, Meta will probably be liable. For all of these reasons, Meta (and its US competitors) are being cautious in this realm.
I think caution is appropriate, even as it may be insufficient. Building agents that act independently on the internet is a big deal, one that poses major safety questions, and I'd like us to have a robust legal framework about what they can do and who is ultimately accountable.
But the worst of all possible worlds is a state of uncertainty that punishes caution and encourages everyone to run agents that have no accountability at all. We have a year or two to figure out how to do better. Let's hope Manus prompts us to get to work on not just building those agents, but building the legal framework that will keep them safe.
A version of this story originally appeared in the Future Perfect newsletter. Sign up here!
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Fast Company
an hour ago
- Fast Company
This free AI supersite is like Gemini Deep Research on steroids
Everywhere you look these days, there it is—some manner of breathlessly hyped new 'AI' service that's, like, totally gonna change your life forever. (Like, totally. For realsies.) Or so they say. In reality, of course, most of this stuff is far more fallible, limited in utility, and inadvisable to use outside of super-specific scenarios than most tech companies (and self-declared 'AI gurus') would lead you to believe. But AI, in its current form, isn't entirely useless. Far from it, in fact: This type of tech can be quite helpful in the right sort of scenario and, critically, if you think about it in the right way—not as an end-all instant answer machine but as a starting point for certain types of specific tasks or info-seeking. And as we wade our way through a year that's absolutely overflowing with overwrought AI ballyhoo, I've got just the tool for you to sift through that sea and seek out some surprising shiny pearls amid all the overwhelming noise. Be the first to find all sorts of little-known tech treasures with my free Cool Tools newsletter from The Intelligence. One useful new discovery in your inbox every Wednesday! Deep research, done right So, you've probably heard all about ChatGPT, Gemini, Perplexity, and the likes, right? They're all generative AI chatbots, which means they use a snazzy-sounding word prediction engine to analyze language patterns and answer your questions, among other more ambitious tasks. 🔎 One of their biggest recent advancements is the ability to perform what everyone's calling 'deep research'—a fancy way of saying they'll dive deep into a topic for you and create a detailed report of info, almost like a custom-made dossier, based on knowledge from all over the web. Again, I can't emphasize enough: The info here isn't infallible. These systems can—and do—get stuff wrong and sometimes even flat-out make up nonsense out of thin air. 🧠 But, as a starting point—especially when they include links to their sources so you can confirm info on your own and use it as an entryway to research as opposed to the final product—it really can save you time and give you a great way to get into a complex topic. And the tool I want to show to you today makes that feature far more powerful, useful, and also affordable than it's ever been before. ⌚ It'll take you 20 seconds to try out for yourself. ➜ It's called, amusingly, Ithy. (Try saying that 10 times fast!) And all it does, in a nutshell, is bring together the 'deep research' tools from a slew of different AI engines—including ChatGPT and Google's Gemini along with Perplexity, Meta AI, and more—into a single streamlined prompt. That means you can use 'em all together to create a single super-report on any subject imaginable. ✅ It couldn't be much easier to make happen, either: First, open up Ithy in any browser, on any device you're using. Type your question or the subject you're thinking about into its box and tap or click the arrow icon within that same line to get going. Select either 'Fast,' if you don't feel like waiting, or 'Deep,' if you've got time and want this thing to go especially in-depth. (Even the 'Fast' path is pretty darn deep, if you ask me.) And, well, that's about it. Ithy will think for a bit, then serve up an impressively detailed dossier on whatever it is you requested—with info coming from a mix of all those AI engines, combined and seamlessly blended together. And I mean seriously detailed, too—with all sorts of sections, graphics, FAQs, and external links for original sources so you can do your own reading and see exactly where it got its info. 📌 Here's a link to the sample report shown here, if you want to look even more closely. ☝️ Now, for the especially cool part: Ithy lets you do all of this free of charge —up to a point. The site gives you five report-creating credits to start, even if you don't sign in. Once you create an account (for free), you'll get 10 credits per month and can optionally then bump up to an unlimited Pro plan—which includes access to the typically pricey pro levels of Gemini and OpenAI—for seven bucks a month, if you go for the annual setup. But even if you don't go that route, 10 in-depth reports per month from all the web's leading AI engines together is a pretty powerful perk to have at your fingertips, without so much as dropping a dime. Ithy is entirely web-based —no downloads or installations required. It's free for up to 5 reports total or 10 reports per month, if you create an account—and optionally available in $7-per-month (paid annually) or $20-per-month (paid monthly) plan for its fully featured, limit-free Pro version. Like most AI engines, Ithy does use questions submitted to its site as training to further improve its AI systems. The questions are also being shared with the associated third-party AI sites, of course. So you'll want to think carefully about what you ask and avoid sending anything especially sensitive or personal (but really, it's designed to answer questions and provide info, so hopefully you wouldn't be submitting your banking info and Social Security number, anyway!).
Yahoo
an hour ago
- Yahoo
OpenAI to appeal copyright ruling in NY Times case as Altman calls for 'AI privilege'
OpenAI CEO Sam Altman said Thursday that the company plans to appeal a court decision in a copyright case brought by The New York Times. "Recently the NYT asked a court to force us to not delete any user chats. We think this was an inappropriate request that sets a bad precedent," Altman wrote in a post on X, adding that the company is "appealing the decision." Ai-generated Attorney Outrages Judge Who Scolds Man Over Courtroom Fake: 'Not A Real Person' "We will fight any demand that compromises our users' privacy; this is a core principle." The OpenAI co-founder said the case has accelerated the need for a conversation about "AI privilege," in which "talking to an AI should be like talking to a lawyer or a doctor." "I hope society will figure this out soon," Altman wrote, suggesting that "spousal privilege" might be a better analogy. Read On The Fox Business App The copyright case, brought by The New York Times, requires OpenAI to preserve ChatGPT output data indefinitely at the outlet's request. Ai's Development Is Critically Important For America – And It All Hinges On These Freedoms U.S. District Judge Sidney Stein was asked to vacate the May data preservation order on June 3, according to a court filing obtained by Reuters. The New York Times sued OpenAI and Microsoft in late 2023, accusing them of using millions of its articles without permission to train the large language model powering ChatGPT. "Defendants' generative artificial intelligence ('GenAI') tools rely on large-language models ('LLMs') that were built by copying and using millions of The Times's copyrighted news articles, in-depth investigations, opinion pieces, reviews, how-to guides, and more," the lawsuit says. The lawsuit claims that while defendants "engaged in widescale copying from many sources," OpenAI "gave Times content particular emphasis when building their LLMs — revealing a preference that recognizes the value of those works." Chatgpt Will Now Combat Bias With New Measures Put Forth By Openai "Using the valuable intellectual property of others in these ways without paying for it has been extremely lucrative for Defendants," the lawsuit says. "Microsoft's deployment of Times-trained LLMs throughout its product line helped boost its market capitalization by a trillion dollars in the past year alone. And OpenAI's release of ChatGPT has driven its valuation to as high as $90 billion." Stein said in an April court opinion that The Times had made a case that OpenAI and Microsoft were responsible for inducing users to infringe its copyrights. The opinion explained an earlier order that rejected parts of an OpenAI and Microsoft motion to dismiss, saying that the Times' "numerous" and "widely publicized" examples of ChatGPT producing material from its articles justified allowing the claims to continue. OpenAI COO Brad Lightcap said The New York Times and other plaintiffs "have made a sweeping and unnecessary demand in their baseless lawsuit against us: retain consumer ChatGPT and API consumer data indefinitely." "This fundamentally conflicts with the privacy commitments we have made to our users. It abandons long-standing privacy norms and weakens privacy protections," Lightcap said in a press release. "We strongly believe this is an overreach by The New York Times. We're continuing to appeal this order so we can keep putting your trust and privacy first. Click Here To Get Fox Business On The Go FOX Business has reached out to The New York Times for comment on OpenAI's appeal. FOX Business' Danielle Wallace and Reuters contributed to this report. Original article source: OpenAI to appeal copyright ruling in NY Times case as Altman calls for 'AI privilege'

Business Insider
an hour ago
- Business Insider
AI search's user experience may be the best it'll ever get, says one founder
By day, Lily Clifford is the CEO and founder of Rime Labs. The startup creates the voice on the other end of the line when you call to order from restaurants like Domino's or Wingstop. Rime trains AI models to create voices with specific regional accents, tones, and other elements that make them easier to converse with. Clifford also uses AI in her daily life, especially in lieu of search engines, she told Business Insider. Instead of pulling up a search engine when she has a question, Clifford usually turns to generative AI chatbots like OpenAI's ChatGPT or Google's Gemini. She said the experience reminds her of using Google or other search engines in the late 1990s and early 2000s. That's when she thinks the user experience was at its prime. "My hot take here is these applications might be the best that they ever will be," she said. Search engines used to be simpler, Clifford said. There were far fewer ads and sponsored results. And optimizing webpages to get more clicks — a practice known as SEO — was in its infancy. Those developments spawned new businesses and became features of the modern internet. But Clifford said search results have also gotten worse for users. It's common to see multiple sponsored results above more relevant ones in a search, for instance. AI chatbots, meanwhile, haven't gone through the same evolution — yet. Companies and individuals are still experimenting with usinggenerative AI for lots of tasks, from writing emails to creating images for advertising campaigns. Many people, like Clifford, use AI as a replacement for search engines. Ask AI a question, and it will often give you an answer in just a few sentences. For some, that's more appealing than clicking through several results from a search engine until you find the information that you're looking for. AI search results can also give users contradictory or incorrect information, though, creating a potential downside to the quick-and-easy answers. Still, Clifford noticed the user experience gap between the chatbots and search engines during a recent trip to Milan, she said. While there, she used an AI chatbot to look for a local place to buy a silk blouse. The chatbot pointed her toward a local seamstress who sold custom blouses through Instagram. "It wasn't like 'Go to Forever 21,' which is probably what would've happened if I typed it into Google," she said. "It was totally wild and fun to use." But, Clifford thinks it's a matter of time before AI chatbots go the way of the search engines before them. Some companies with big investments in generative AI search tools are taking steps in that direction. Last month, Google said it would expand its use of ads in some of the AI Overviews that appear at the top of its search results, for example. And some marketing experts now offer help with " answer engine optimization," or AEO.