Latest news with #DoubaoAI


News18
01-06-2025
- Science
- News18
Chinese Paraglider's Accidental 8,000m Feat Video AI-Generated? Check Details
Last Updated: The video clip contained no logo, but a portion of it uploaded to Facebook on May 25 had a Doubao AI watermark, indicating it was likely generated using Doubao AI's tool. A viral video claiming to show a Chinese paraglider's accidental ascent to over 8,000 meters has been found to be generated by artificial intelligence, at least partially, according to a report in NBC News. The paraglider, Peng Yujiang, 55, had reportedly claimed to have been lifted from 3,000 meters, but the unexpected conditions carried him nearly 5,000 metres higher. Peng Yujiang allegedly withstood temperatures of around -40°C and critically low oxygen levels, overcoming hypoxia and severe frostbite while remaining conscious, and safely landed back on the ground. However, the investigation reveals that the first five seconds of the video might be AI-generated. In the clip, Peng can be seen gliding at high altitude with his legs dangling, but the footage is seemingly cropped. The video clip contained no logo, but a portion of it uploaded to Facebook on May 25 had a Doubao AI watermark, indicating it was likely generated using ByteDance-owned Doubao AI's tool. advetisement 'It's unclear if the remaining footage of Peng gliding through the sky, which differs from the first five seconds, is authentic or not," the publication stated. AI-verification firm GetReal Labs analysed the footage and found evidence supporting the claim that AI was used to create it. 'We were able to extract a few frames and analyse them using our Inspect platform, and our models confirm that the images are synthetic," said GetReal Labs, adding that several other elements in the video differed from the rest of the footage. Meanwhile, news agency Reuters, which distributed the clip without the AI logo, has taken down the video. Other news outlets have also removed their versions of the video. 'We have reason to believe this is an AI-generated video and are currently working on killing this footage," Reuters said. Peng was testing used paragliding gear when the incident supposedly occurred. The video clip quickly went viral, with users praising Peng for remaining calm under extreme conditions. 'I had just bought a second-hand paragliding harness and wanted to test it, so I was conducting ground parachute shaking. After a while, the wind suddenly picked up and lifted me into the air. I tried to land as soon as possible, but I failed," Peng told the Chinese media. Peng and his friend, who posted the video, were banned from paragliding for six months for violating safety rules and flying without authorisation. Location : China First Published: June 01, 2025, 10:01 IST


Forbes
23-04-2025
- Business
- Forbes
China's ‘Smart' AI Assistants Echoes Microsoft Recall's Privacy Flaws
As TikTok owner ByteDance rolls out Doubao AI screen-reading AI assistants, convenience and ... More capability are accelerating — so are the stakes for privacy and transparency. If there's one idea the global artificial intelligence industry seems to agree on, it's this: intelligent agents are the next phase of AI. More than just chatbots, these tools are designed to plan, execute and assist across multiple domains. From OpenAI to China's Manus AI, developers are working toward smart AI assistants that can handle tasks autonomously — sometimes with minimal prompting. But while much of the conversation in the West centers on technical benchmarks and API integrations, China's largest tech platforms are pursuing a different path. There, the rise of 'screen-aware' agents — AI assistants that can view, interpret and act based on what's displayed on your screen by asking the user to turn on your PC's assistive functions — represents a shift in how AI is being embedded into everyday computing. One of the most notable examples is ByteDance's Doubao AI, a consumer-facing assistant that introduced a Screen Sharing Call feature last month. With the user's permission, the tool gains visibility over everything currently shown on the desktop during a voice interaction. It can read WeChat conversations, solve math problems on-screen, identify active applications and summarize video clips — all without switching context. This design echoes Microsoft's now-infamous Recall feature, which was similarly built to help users remember on-screen activities by capturing snapshots. Microsoft faced strong public backlash last year and delayed the rollout twice, ultimately limiting it to local data storage and making it opt-in only when it announced earlier this month that it was rereleasing it. Despite Recall's rocky start, this approach has signaled a broader trend in Chinese AI development — one focused on integrating agents directly into daily workflows by giving them wide-reaching visual context. While the technology promises to make assistants significantly more useful, it also raises serious and urgent questions about how AI developers are addressing privacy, consent and transparency. The key technical enabler behind Doubao AI's feature is access to a system-level setting known as accessibility services. Originally designed to help users with disabilities, this permission grants apps the ability to view and interact with nearly all visual elements on the screen. In practice, this means that once a user grants Doubao AI this permission, it can continuously observe what's happening on the desktop during a voice call. While the feature does request user consent, the language used is often technical, and the implications are not always clear. Unlike screen sharing in video calls — where users deliberately choose what to show — Doubao AI sees everything unless the user manually closes apps or disables the function. What's more, the information interpreted by Doubao AI is uploaded to ByteDance's servers for processing. This introduces a second layer of consideration: What happens to that data, how is it stored and what safeguards are in place to prevent misuse or unauthorized access? Pavan Davuluri, Microsoft corporate vice president, Windows and devices, speaks about Recall on May ... More 20, 2024. So far, Doubao hasn't seen the same level of scrutiny that Recall did — at least not publicly. That may reflect differences in user expectations or regulatory maturity, but it also highlights how quickly powerful features can become normalized without broad public discussion. Doubao isn't an outlier. Other Chinese developers are rolling out screen-aware capabilities as well. Zhipu AI's GLM-PC can carry out cross-application tasks on the desktop, while AutoGLM is focused on browser-based actions. Smartphone manufacturers like Xiaomi, OPPO, Honor and Vivo are embedding similar AI functions at the system level, often relying on accessibility permissions to parse visual content. In some cases, these integrations happen with minimal user notification. OPPO's ColorOS privacy policy, for instance, notes that its Xiaobu assistant can use accessibility services to scan screen elements — without explicitly prompting the user or offering an opt-out. This low-friction approach increases usability, but it also shifts more responsibility onto users to understand what they've enabled. Oppo store in Shanghai, China From a product standpoint, it's easy to see the appeal. AI assistants that can 'see' the user's screen are able to provide more relevant, contextual help. They can summarize a document that's open, highlight unread messages or offer quick actions tied to what's currently displayed. In environments where fast interaction matters — such as mobile operating systems or productivity tools — screen awareness can be a significant feature advantage. But it also raises structural questions about how digital consent is designed. If an AI assistant sees more than the user expects — or retains data longer than necessary — the balance between functionality and control can quickly shift. The differing responses to screen-aware AI in China and the West reflect deeper contrasts in how AI is being developed and governed. In the U.S. and Europe, AI assistants are typically sandboxed using formal APIs, explicit tool integrations and clear scopes of action. Developers emphasize consent mechanisms and often shape deployment timelines around public feedback — as Microsoft's Recall experience shows. In China, the reliance on accessibility services and GUI-based interpretation allows for quicker integration across apps and platforms, but often with less granular control. It's a practical workaround to API fragmentation, and it accelerates product development. But the trade-off is that users may not fully understand what permissions they've granted — or how their data is being handled. There are some signs of industry self-regulation. Xiaomi CEO Lei Jun has called for establishing standards for AI terminals, including a grading system and unified privacy frameworks. And the China Software Industry Association recently released guidelines on accessibility permissions, recommending limits on data access and greater transparency for users. Still, these measures are early-stage and largely advisory. As more assistants adopt screen-sharing or screen-reading features, the need for clearer norms and accountability will only grow. It's worth emphasizing: screen-aware assistants are not inherently problematic. Used carefully, they can enhance productivity, improve accessibility and simplify digital tasks. There are legitimate use cases for users who want AI support that adapts in real time to what's on screen. But like any powerful feature, the impact depends on how it's implemented — and how well users understand it. The key challenges include: Without clear answers to these questions, screen-sharing by default could become normalized before its implications are fully understood. As AI assistants evolve into intelligent agents capable of multistep reasoning and cross-platform actions, visual context will almost certainly become part of their toolkit. Whether through screenshot-based models, accessibility hooks or integrated APIs, screen awareness adds a layer of intelligence that static inputs can't replicate. The question is not whether this trend will continue — it will — but whether the industry builds the right infrastructure to ensure it's done responsibly. China's deployment of these tools offers a preview of what's technically possible and where frictionless design can lead. It also provides a cautionary tale: powerful AI features introduced without transparency risk undermining trust just as AI enters more personal spaces. Developers, regulators and users alike would benefit from treating screen-aware agents not as a niche product tweak but as a pivotal shift in how we interact with machines. If the smart AI assistant of the future is to be a true partner, it must be built on a foundation of clarity, respect and informed choice.