logo
China shuts down AI tools during nationwide college exams

China shuts down AI tools during nationwide college exams

The Verge2 hours ago

Chinese AI companies have temporarily paused some of their chatbot features to prevent students from using them to cheat during nationwide college exams, Bloomberg reports. Popular AI apps, including Alibaba's Qwen and ByteDance's Doubao, have stopped picture recognition features from responding to questions about test papers, while Tencent's Yuanbao, Moonshot's Kimi have suspended photo-recognition services entirely during exam hours.
The increasing availability of chatbots has made it easier than ever for students around the world to cheat their way through education. Schools in the US are trying to address the issue by reintroducing paper tests, with the Wall Street Journal reporting in May that sales of blue books have boomed in universities across the country over the last two years.
The rigorous multi-day 'gaokao' exams are sat by more than 13.3 million Chinese students between June 7-10th, each fighting to secure one of the limited spots at universities across the country. Students are already banned from using devices like phones and laptops during the hours-long tests, so the disabling of AI chatbots serves as an additional safety net to prevent cheating during exam season.
When asked to explain the suspension, Bloomberg reports the Yuanbao and Kimi chatbots responded that functions had been disabled 'to ensure the fairness of the college entrance examinations.' Similarly, the DeepSeek AI tool that went viral earlier this year is also blocking its service during specific hours 'to ensure fairness in the college entrance examination,' according to The Guardian.
We were unable to find any public announcements from the AI companies mentioned, with The Guardian reporting that news around the shutdowns is being driven by students on the Chinese social media platform Weibo. The gaokao entrance exam incites fierce competition as it's the only means to secure a college placement in China, driving concerns that students may try to improve their chances with AI tools.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

This AI Company Wants Washington To Keep Its Competitors Off the Market
This AI Company Wants Washington To Keep Its Competitors Off the Market

Yahoo

time25 minutes ago

  • Yahoo

This AI Company Wants Washington To Keep Its Competitors Off the Market

Dario Amodei, CEO of the artificial intelligence company Anthropic, published a guest essay in The New York Times Thursday arguing against a proposed 10-year moratorium on state AI regulation. Amodei argues that a patchwork of regulations would be better than no regulation whatsoever. Skepticism is warranted whenever the head of an incumbent firm calls for more regulation, and this case is no different. If Amodei gets his way, Anthropic would face less competition—to the detriment of AI innovation, AI security, and the consumer. Amodei's op-ed came in a response to a provision of the so-called One Big Beautiful Bill Act, which would prevent any states, cities, and counties from enforcing any regulation that specifically targets AI models, AI systems, or automated decision systems for 10 years. Senate Republicans have amended the clause from a simple requirement to a condition for receiving federal broadband funds, in order to comply with the Byrd Rule, which in Politico's words "blocks anything but budgetary issues from inclusion in reconciliation." Amodei begins by describing how, in a recent stress test conducted at his company, a chatbot threatened an experimenter to forward evidence of his adultery to his wife unless he withdrew plans to shut the AI down. The CEO also raises more tangible concerns, such as reports that a version of Google's Gemini model is "approaching a point where it could help people carry out cyberattacks." Matthew Mittelsteadt, a technology fellow at the Cato Institute, tells Reason that the stress test was "very contrived" and that "there are no AI systems where you must prompt it to turn it off." You can just turn it off. He also acknowledges that, while there is "a real cybersecurity danger [of] AI being used to spot and exploit cyber-vulnerabilities, it can also be used to spot and patch" them. Outside of cyberspace and in, well, actual space, Amodei sounds the alarm that AI could acquire the ability "to produce biological and other weapons." But there's nothing new about that: Knowledge and reasoning, organic or artificial—ultimately wielded by people in either case—can be used to cause problems as well as to solve them. An AI that can model three-dimensional protein structures to create cures for previously untreatable diseases can also create virulent, lethal pathogens. Amodei recognizes the double-edged nature of AI and says voluntary model evaluation and publication are insufficient to ensure that benefits outweigh costs. Instead of a 10-year moratorium, Amodei calls on the White House and Congress to work together on a transparency standard for AI companies. In lieu of federal testing standards, Amodei says state laws should pick up the slack without being "overly prescriptive or burdensome." But that caveat is exactly the kind of wishful thinking Amodei indicts proponents of the moratorium for: Not only would 50 state transparency laws be burdensome, says Mittelsteadt, but they could "actually make models less legible." Neil Chilson of the Abundance Institute also inveighed against Amodei's call for state-level regulation, which is much more onerous than Amodei suggests. "The leading state proposals…include audit requirements, algorithmic assessments, consumer disclosures, and some even have criminal penalties," Chilson tweeted, so "the real debate isn't 'transparency vs. nothing,' but 'transparency-only federal floor vs. intrusive state regimes with audits, liability, and even criminal sanctions.'" Mittelsteadt thinks national transparency regulation is "absolutely the way to go." But how the U.S. chooses to regulate AI might not have much bearing on Skynet-doomsday scenarios, because, while America leads the way in AI, it's not the only player in the game. "If bad actors abroad create Amodei's theoretical 'kill everyone bot,' no [American] law will matter," says Mittelsteadt. But such a law can "stand in the way of good actors using these tools for defense." Amodei is not the only CEO of a leading AI company to call for regulation. In 2023, Sam Altman, co-founder and then-CEO of Open AI, called on lawmakers to consider "intergovernmental oversight mechanisms and standard-setting" of AI. In both cases and in any others that come along, the public should beware of calls for AI regulation that will foreclose market entry, protect incumbent firms' profits from being bid away by competitors, and reduce the incentives to maintain market share the benign way: through innovation and product differentiation. The post This AI Company Wants Washington To Keep Its Competitors Off the Market appeared first on

Committee explores nuclear solutions to AI demand
Committee explores nuclear solutions to AI demand

E&E News

time26 minutes ago

  • E&E News

Committee explores nuclear solutions to AI demand

House Science, Space and Technology Committee lawmakers will meet this week to discuss how nuclear energy could help meet a projected surge in demand from artificial intelligence operations. The Energy Subcommittee hearing — to be led by Chair Randy Weber (R-Texas) — continues Republicans' early focus and significant concern regarding supply and demand in the 119th Congress. They believe baseload energy sources, such as nuclear and fossil fuels, need to be built at a rapid pace to offset a surge in intermittent, renewable energy generation that could put grid reliability at risk. Indeed, transmission providers are forecasting an 8.2 percent growth in electricity load over the next five years primarily due to AI data center proliferation. That's equivalent to hooking up nearly 50 million homes to the grid by 2029. Advertisement But whether nuclear energy can actually meet that demand remains a point of debate among energy and policy experts.

Apple Poised to Monetize AI at WWDC 2025
Apple Poised to Monetize AI at WWDC 2025

Yahoo

time26 minutes ago

  • Yahoo

Apple Poised to Monetize AI at WWDC 2025

Apple (NASDAQ:AAPL) looks ready to kick off its AI monetization era with system-wide updates at WWDC 2025, Wedbush's Daniel Ives says, setting the stage for paid AI features across the Apple ecosystem. The keynote at 1 p.m. ET today at Apple Park will unveil '26 upgrades for macOS, iOS and iPadOS powered by Apple Intelligence, countering Street skepticism about a slow AI rollout. In his Monday note, Ives argued that WWDC marks the start of Apple's AI cash flow, not a mere feature demo, as the company layers new AI-driven capabilities into core OS updates. He expects details on Siri's deeper integration with Google's Gemini and OpenAI's ChatGPT, demonstrating how Apple will embed AI across native apps to drive user engagementand, ultimately, device upgrades when iPhone 17 ships next year. With over 100 million iPhones in China due for an upgrade, Ives also eyes an announcement on Apple's partnership with Alibaba (NYSE:BABA) to deploy AI services locally, a move he calls critical for unlocking growth in the world's largest smartphone market. Wedbush reiterates its Outperform rating and raises its 12-month price target to $270, noting that Apple's edge isn't in building the most advanced large language model but in toll-collecting'' on its vast hardware base. Apple's unmatched ecosystem ensures that any third-party AI app must run through Cupertino, highlighting how AI unlocks new services revenue without upending its product strategy. Why It Matters: WWDC's AI announcements could shift Wall Street sentiment, validating Apple's long-term AI strategy and fueling expectations for recurring software revenue beyond hardware sales. Investors will watch for concrete details on AI feature pricing, Siri-Gemini integrations and the Alibaba tie-up during today's keynote and in the follow-up developer sessions. This article first appeared on GuruFocus. Error while retrieving data Sign in to access your portfolio Error while retrieving data Error while retrieving data Error while retrieving data Error while retrieving data

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store