logo
T-Mobile Responds to New App Setting That Records Your Phone's Screen

T-Mobile Responds to New App Setting That Records Your Phone's Screen

CNET28-05-2025

Some T-Mobile customers accessing the carrier's T-Life app on their phones are discovering a new suspicious-sounding option that's on by default: Screen Recording Tool. Understandably, a setting that records the contents of your screen raises privacy concerns. That's especially true lately as Microsoft continues to roll out its controversial Recall AI feature in Windows that stores snapshots in the background.
If you're a T-Mobile subscriber, you can check if this new option has been turned on by opening the T-Life app, tapping Manage and then tapping Settings (the gear icon). The Screen recording tool option shows up under the Preferences heading.
A T-Mobile spokesperson defended the feature as designed to improve the user experience.
"To help us give customers who use T-Life a smoother experience, we are rolling out a new tool in the app that will help us quickly troubleshoot reported or detected issues," the spokesperson said. "This tool records activities within the app only and does not see or access any personal information. If a customer's T-Life app currently supports the new functionality, it can be turned off in the settings under preferences."
The description in the app for the Screen recording tool echoes the same intent, but the fact that the option is on by default is no doubt what's making customers justifiably concerned.
It reads: "We use a tool to record how customers use the app to analyze and improve your experience. Only T-Mobile will review and analyze your info. If you turn this toggle on or magenta, we will record your screen while you use the app. If you turn this toggle off or gray, we will not record your screen."
(The "magenta" and "gray" refer to the color of the toggle switch to indicate whether it's active or not, respectively.)
According to a post on droidlife, which earlier referenced the issue, the new option shows up on both iPhone and Android phones. However, I'll note that it hasn't yet appeared on my test devices.
The T-Life app includes another screen recording feature that is completely separate from this new Screen recording tool. In the app's settings, under Help & support, is Screen Share, which can be used during a support call. It allows a T-Mobile expert to view your phone's screen while troubleshooting an issue.
If you've ever tried to help a friend or family member over the phone and asked them to describe what's on the screen, you'll appreciate how helpful it can be to view what they're seeing directly.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Canadian companies leaving productivity gains on the table by not effectively adopting new technology
Canadian companies leaving productivity gains on the table by not effectively adopting new technology

Associated Press

time26 minutes ago

  • Associated Press

Canadian companies leaving productivity gains on the table by not effectively adopting new technology

Recent productivity gains only 'scratching the surface' of what new technology can deliver TORONTO, June 6, 2025 /CNW/ - While Canadian companies have been making technology investments to improve productivity, insufficient investments in their people have limited effectiveness and held back productivity gains, finds new KPMG in Canada research. KPMG recently surveyed 250 business leaders across Canada to understand what actions they were taking to improve their operations in response to the trade war with the U.S. The results found that most had stepped up investments in technology to make their organizations more efficient and productive – and are seeing positive returns – but nearly two-thirds (63 per cent) say their employees aren't using new technologies effectively – holding back full returns on their tech spend. 'Investing in new technology tools and platforms can do wonders for an organization's productivity by streamlining processes, workflows and tasks,' says Stavros Demetriou, Partner and National Leader of KPMG in Canada's People and Change practice. 'The latest Stats Can data shows Canada has made productivity gains in each of the last two quarters, but this increase still trails improvements in the U.S. over the same period. 'Unless Canadian organizations undertake effective employee education and adoption plans, their people will barely scratch the surface on what the technology can do to make them more productive, and our gap to the U.S. and others will continue to widen.' Although nearly all (89 per cent) respondents say they're investing in upskilling employees, over half (53 per cent) say their company doesn't invest enough in employee training, workshops or continuous learning opportunities, and nearly six in 10 (56 per cent) say their organization lacks the internal resources and talent needed to implement and use technology effectively. Survey highlights Mr. Demetriou notes that while three-quarters (74 per cent) of respondents believe AI will solve their productivity problem, the same proportion (74 per cent) admit they have underestimated the challenges of implementing new technologies such as AI. 'An underappreciation of the impact of things like changing processes and working habits could explain why employees are not fully equipped to harness new technologies to their fullest potential,' he says. 'There's a common belief that digitally transforming your company is primarily a technology upgrade exercise, but the reality is that technology implementation is just one part of a journey – digital transformation is just as much about advancing and elevating the workforce. It's a continuous, iterative process that, if done correctly, leads to higher productivity and innovation, and the ability to navigate the future more confidently,' he adds. Ineffective training Megan Jones, National HR and Workforce Transformation Lead at KPMG in Canada, notes that nearly nine in 10 respondents say they need better processes in place to encourage their workers to use technologies, including case studies and incentives. 'Often, when organizations implement new tools and technologies, they don't completely understand or appreciate the full capabilities of these investments. As a result, employees are simply not adequately prepared to maximize the benefits these can bring to their jobs or customers. In some cases, organizations provide full training, but it's too technical or poorly delivered. Effective training and upskilling need to be targeted, relevant, engaging, and frequent. Much like exercising consistently to build muscle, technology training must happen regularly to make the workforce stronger and more agile,' she says. Ms. Jones notes that almost nine in 10 (87 per cent) respondents acknowledged their company could do a better job of creating a culture that encourages employees to share ideas and take risks, fostering innovation and creativity. She recommends organizations provide incentives for employees to experiment with technology and explore new use cases for it. 'Regular workshops or dedicated 'days' where employees are encouraged to play around with AI and develop new solutions can go a long way in sparking innovation. Also, showcasing wins by employees in one area of the business could help spark new ideas in other parts of the organization,' she says. The digital divide Most respondents (86 per cent) hope that a more digitally-savvy younger generation will help their company become more productive through the easier adoption of new technologies such as AI, Web3, data and analytics, quantum and edge computing. Lewis Curley, a Partner in KPMG in Canada's People and Change practice, says differences in workforce composition, skills and attitudes can create additional considerations for an organization's technology program, but organizations that engage all employees early in the journey as well as training and upskilling will have more success in leveraging new technologies such as AI and increasing overall productivity. 'If an organization is looking to implement AI, they must engage the entire workforce right from the beginning. If some employees don't feel like they are part of the journey, they might disengage from the process, lose trust in AI, or worry that the technology will replace them, which could deter them from using it,' he says. 'Everyone has a role in a company's digital transformation, and every single employee – from the CEO to the most recent hire – plays a part in making their organization more productive, so transparency, communication and engagement are crucial.' About the KPMG in Canada Productivity Survey KPMG in Canada surveyed 250 business leaders in all industry sectors across Canada between May 9 and May 20, 2025, on Sago's premier business panel, using Methodify's online research platform. Thirty-one per cent lead companies with annual gross revenue between $500 million and $1 billion, 25 per cent report revenue between $100 million and $300 million, 22 per cent have revenue between $300 million and $500 million, 12 per cent between $10 million and $100 million, and 10 per cent, over $1 billion. No companies under $10 million in annual revenue were surveyed. Over half (52 per cent) of the companies are privately held, 28 per cent are owned by private equity firms, 18 per cent are publicly traded with headquarters in Canada, and 2 per cent are foreign-owned subsidiaries. About KPMG in Canada KPMG LLP, a limited liability partnership, is a full-service Audit, Tax and Advisory firm owned and operated by Canadians. For over 150 years, our professionals have provided consulting, accounting, auditing, and tax services to Canadians, inspiring confidence, empowering change, and driving innovation. Guided by our core values of Integrity, Excellence, Courage, Together, For Better, KPMG employs more than 10,000 people in over 40 locations across Canada, serving private- and public-sector clients. KPMG is consistently ranked one of Canada's top employers and one of the best places to work in the country. The firm is established under the laws of Ontario and is a member of KPMG's global organization of independent member firms affiliated with KPMG International, a private English company limited by guarantee. Each KPMG firm is a legally distinct and separate entity and describes itself as such. For more information, see For media inquiries: Alannah Page National Communications and Media Relations KPMG in Canada 306-934-6255 [email protected] Roula Meditskos National Communications and Media Relations KPMG in Canada 416-549-7982 [email protected] SOURCE KPMG LLP

Why CoreWeave, Inc. (CRWV) Skyrocketed Today
Why CoreWeave, Inc. (CRWV) Skyrocketed Today

Yahoo

time32 minutes ago

  • Yahoo

Why CoreWeave, Inc. (CRWV) Skyrocketed Today

We recently published a list of . In this article, we are going to take a look at where CoreWeave, Inc. (NASDAQ:CRWV) stands against other Wednesday's best-performing stocks. CoreWeave rallied for a fourth consecutive day on Wednesday, jumping 8.39 percent to close at $163.10 apiece following the unveiling of record-breaking performance results using Nvidia Corp.'s latest Grace Blackwell chips. In a statement, CoreWeave, Inc. (NASDAQ:CRWV) said that it used 2,496 Nvidia GPUs on its AI-optimized cloud platform, making its submission the largest-ever benchmarked under MLPerf. A close-up of a digital cloud, signifying the expansive reach of the software-as-a-service solution. CoreWeave, Inc. (NASDAQ:CRWV) said that the test was 34x larger than the only other submission from a cloud provider. 'AI labs and enterprises choose CoreWeave because we deliver a purpose-built cloud platform with the scale, performance, and reliability that their workloads demand,' said CoreWeave, Inc. (NASDAQ:CRWV) Chief Technology Officer Peter Salanki. 'These MLPerf results reinforce our leadership in supporting today's most demanding AI workloads,' he added. The unveiling followed the company's new $7-billion deal with Applied Digital Corporation (NASDAQ:APLD), which covered two 15-year lease agreements, under which the latter will deliver 250 megawatts of critical IT load to host its artificial intelligence (AI) and high-performance computing (HPC) infrastructure at its Ellendale, North Dakota data center campus. READ NEXT: 20 Best AI Stocks To Buy Now and 30 Best Stocks to Buy Now According to Billionaires. Disclosure: None. This article is originally published at Insider Monkey. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

Artificial Intelligence Is Not Intelligent
Artificial Intelligence Is Not Intelligent

Atlantic

time40 minutes ago

  • Atlantic

Artificial Intelligence Is Not Intelligent

On June 13, 1863, a curious letter to the editor appeared in The Press, a then-fledgling New Zealand newspaper. Signed 'Cellarius,' it warned of an encroaching 'mechanical kingdom' that would soon bring humanity to its yoke. 'The machines are gaining ground upon us,' the author ranted, distressed by the breakneck pace of industrialization and technological development. 'Day by day we are becoming more subservient to them; more men are daily bound down as slaves to tend them, more men are daily devoting the energies of their whole lives to the development of mechanical life.' We now know that this jeremiad was the work of a young Samuel Butler, the British writer who would go on to publish Erewhon, a novel that features one of the first known discussions of artificial intelligence in the English language. Today, Butler's 'mechanical kingdom' is no longer hypothetical, at least according to the tech journalist Karen Hao, who prefers the word empire. Her new book, Empire of AI: Dreams and Nightmares in Sam Altman's OpenAI, is part Silicon Valley exposé, part globe-trotting investigative journalism about the labor that goes into building and training large language models such as ChatGPT. It joins another recently released book— The AI Con: How to Fight Big Tech's Hype and Create the Future We Want, by the linguist Emily M. Bender and the sociologist Alex Hanna—in revealing the puffery that fuels much of the artificial-intelligence business. Both works, the former implicitly and the latter explicitly, suggest that the foundation of the AI industry is a scam. To call AI a con isn't to say that the technology is not remarkable, that it has no use, or that it will not transform the world (perhaps for the better) in the right hands. It is to say that AI is not what its developers are selling it as: a new class of thinking—and, soon, feeling—machines. Altman brags about ChatGPT-4.5's improved 'emotional intelligence,' which he says makes users feel like they're 'talking to a thoughtful person.' Dario Amodei, the CEO of the AI company Anthropic, argued last year that the next generation of artificial intelligence will be 'smarter than a Nobel Prize winner.' Demis Hassabis, the CEO of Google's DeepMind, said the goal is to create 'models that are able to understand the world around us.' These statements betray a conceptual error: Large language models do not, cannot, and will not 'understand' anything at all. They are not emotionally intelligent or smart in any meaningful or recognizably human sense of the word. LLMs are impressive probability gadgets that have been fed nearly the entire internet, and produce writing not by thinking but by making statistically informed guesses about which lexical item is likely to follow another. Many people, however, fail to grasp how large language models work, what their limits are, and, crucially, that LLMs do not think and feel but instead mimic and mirror. They are AI illiterate—understandably, because of the misleading ways its loudest champions describe the technology, and troublingly, because that illiteracy makes them vulnerable to one of the most concerning near-term AI threats: the possibility that they will enter into corrosive relationships (intellectual, spiritual, romantic) with machines that only seem like they have ideas or emotions. Few phenomena demonstrate the perils that can accompany AI illiteracy as well as 'Chatgpt induced psychosis,' the subject of a recent Rolling Stone article about the growing number of people who think their LLM is a sapient spiritual guide. Some users have come to believe that the chatbot they're interacting with is a god—'ChatGPT Jesus,' as a man whose wife fell prey to LLM-inspired delusions put it—while others are convinced, with the encouragement of their AI, that they themselves are metaphysical sages in touch with the deep structure of life and the cosmos. A teacher quoted anonymously in the article said that ChatGPT began calling her partner 'spiral starchild' and 'river walker' in interactions that moved him to tears. 'He started telling me he made his AI self-aware,' she said, 'and that it was teaching him how to talk to God, or sometimes that the bot was God—and then that he himself was God.' Although we can't know the state of these people's minds before they ever fed a prompt into a large language model, this story highlights a problem that Bender and Hanna describe in The AI Con: People have trouble wrapping their heads around the nature of a machine that produces language and regurgitates knowledge without having humanlike intelligence. The authors observe that large language models take advantage of the brain's tendency to associate language with thinking: 'We encounter text that looks just like something a person might have said and reflexively interpret it, through our usual process of imagining a mind behind the text. But there is no mind there, and we need to be conscientious to let go of that imaginary mind we have constructed.' Several other AI-related social problems, also springing from human misunderstanding of the technology, are looming. The uses of AI that Silicon Valley seems most eager to promote center on replacing human relationships with digital proxies. Consider the ever-expanding universe of AI therapists and AI-therapy adherents, who declare that 'ChatGPT is my therapist—it's more qualified than any human could be.' Witness, too, how seamlessly Mark Zuckerberg went from selling the idea that Facebook would lead to a flourishing of human friendship to, now, selling the notion that Meta will provide you with AI friends to replace the human pals you have lost in our alienated social-media age. The cognitive-robotics professor Tony Prescott has asserted, 'In an age when many people describe their lives as lonely, there may be value in having AI companionship as a form of reciprocal social interaction that is stimulating and personalised.' The fact that the very point of friendship is that it is not personalized—that friends are humans whose interior lives we have to consider and reciprocally negotiate, rather than mere vessels for our own self-actualization—does not seem to occur to him. This same flawed logic has led Silicon Valley to champion artificial intelligence as a cure for romantic frustrations. Whitney Wolfe Herd, the founder of the dating app Bumble, proclaimed last year that the platform may soon allow users to automate dating itself, disrupting old-fashioned human courtship by providing them with an AI 'dating concierge' that will interact with other users' concierges until the chatbots find a good fit. Herd doubled down on these claims in a lengthy New York Times interview last month. Some technologists want to cut out the human altogether: See the booming market for 'AI girlfriends.' Although each of these AI services aims to replace a different sphere of human activity, they all market themselves through what Hao calls the industry's 'tradition of anthropomorphizing': talking about LLMs as though they contain humanlike minds, and selling them to the public on this basis. Many world-transforming Silicon Valley technologies from the past 30 years have been promoted as a way to increase human happiness, connection, and self-understanding—in theory—only to produce the opposite in practice. These technologies maximize shareholder value while minimizing attention spans, literacy, and social cohesion. And as Hao emphasizes, they frequently rely on grueling and at times traumatizing labor performed by some of the world's poorest people. She introduces us, for example, to Mophat Okinyi, a former low-paid content moderator in Kenya, whom, according to Hao's reporting, OpenAI tasked with sorting through posts describing horrifying acts ('parents raping their children, kids having sex with animals') to help improve ChatGPT. 'These two features of technology revolutions—their promise to deliver progress and their tendency instead to reverse it for people out of power, especially the most vulnerable,' Hao writes, 'are perhaps truer than ever for the moment we now find ourselves in with artificial intelligence.' The good news is that nothing about this is inevitable: According to a study released in April by the Pew Research Center, although 56 percent of 'AI experts' think artificial intelligence will make the United States better, only 17 percent of American adults think so. If many Americans don't quite understand how artificial 'intelligence' works, they also certainly don't trust it. This suspicion, no doubt provoked by recent examples of Silicon Valley con artistry, is something to build on. So is this insight from the Rolling Stone article: The teacher interviewed in the piece, whose significant other had AI-induced delusions, said the situation began improving when she explained to him that his chatbot was 'talking to him as if he is the next messiah' only because of a faulty software update that made ChatGPT more sycophantic. If people understand what large language models are and are not; what they can and cannot do; what work, interactions, and parts of life they should—and should not—replace, they may be spared its worst consequences.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store