logo
AI tools collect and store data about you from all your devices, here's how to be aware of what you're revealing

AI tools collect and store data about you from all your devices, here's how to be aware of what you're revealing

Time of India13 hours ago

Tired of too many ads?
Remove Ads
Tired of too many ads?
Remove Ads
Tired of too many ads?
Remove Ads
Like it or not, artificial intelligence has become part of daily life. Many devices, including electric razors and toothbrushes, have become "AI-powered," using machine learning algorithms to track how a person uses the device, how the device is working in real time, and provide feedback. From asking questions to an AI assistant like ChatGPT or Microsoft Copilot to monitoring a daily fitness routine with a smartwatch, many people use an AI system or tool every day.While AI tools and technologies can make life easier, they also raise important questions about data privacy. These systems often collect large amounts of data, sometimes without people even realizing their data is being collected. The information can then be used to identify personal habits and preferences, and even predict future behaviours by drawing inferences from the aggregated data.As an assistant professor of cybersecurity at West Virginia University , I study how emerging technologies and various types of AI systems manage personal data and how we can build more secure, privacy-preserving systems for the future. Generative AI software uses large amounts of training data to create new content such as text or images. Predictive AI uses data to forecast outcomes based on past behaviour, such as how likely you are to hit your daily step goal, or what movies you may want to watch. Both types can be used to gather information about you.Generative AI assistants such as ChatGPT and Google Gemini collect all the information users type into a chat box. Every question, response and prompt that users enter is recorded, stored and analysed to improve the AI model.OpenAI's privacy policy informs users that "we may use content you provide us to improve our Services, for example to train the models that power ChatGPT." Even though OpenAI allows you to opt out of content use for model training, it still collects and retains your personal data. Although some companies promise that they anonymise this data, meaning they store it without naming the person who provided it, there is always a risk of data being reidentified.Beyond generative AI assistants, social media platforms like Facebook, Instagram and TikTok continuously gather data on their users to train predictive AI models. Every post, photo, video, like, share and comment, including the amount of time people spend looking at each of these, is collected as data points that are used to build digital data profiles for each person who uses the service.The profiles can be used to refine the social media platform's AI recommender systems. They can also be sold to data brokers, who sell a person's data to other companies to, for instance, help develop targeted advertisements that align with that person's interests.Many social media companies also track users across websites and applications by putting cookies and embedded tracking pixels on their computers. Cookies are small files that store information about who you are and what you clicked on while browsing a website.One of the most common uses of cookies is in digital shopping carts: When you place an item in your cart, leave the website and return later, the item will still be in your cart because the cookie stored that information. Tracking pixels are invisible images or snippets of code embedded in websites that notify companies of your activity when you visit their page. This helps them track your behaviour across the internet.This is why users often see or hear advertisements that are related to their browsing and shopping habits on many of the unrelated websites they browse, and even when they are using different devices, including computers, phones and smart speakers. One study found that some websites can store over 300 tracking cookies on your computer or mobile phone.Like generative AI platforms, social media platforms offer privacy settings and opt-outs, but these give people limited control over how their personal data is aggregated and monetized. As media theorist Douglas Rushkoff argued in 2011, if the service is free, you are the product.Many tools that include AI don't require a person to take any direct action for the tool to collect data about that person. Smart devices such as home speakers, fitness trackers and watches continually gather information through biometric sensors, voice recognition and location tracking. Smart home speakers continually listen for the command to activate or "wake up" the device. As the device is listening for this word, it picks up all the conversations happening around it, even though it does not seem to be active.Some companies claim that voice data is only stored when the wake word - what you say to wake up the device - is detected. However, people have raised concerns about accidental recordings, especially because these devices are often connected to cloud services, which allow voice data to be stored, synced and shared across multiple devices such as your phone, smart speaker and tablet.If the company allows, it's also possible for this data to be accessed by third parties, such as advertisers, data analytics firms or a law enforcement agency with a warrant.This potential for third-party access also applies to smartwatches and fitness trackers, which monitor health metrics and user activity patterns. Companies that produce wearable fitness devices are not considered "covered entities" and so are not bound by the Health Information Portability and Accountability Act. This means that they are legally allowed to sell health- and location-related data collected from their users.Concerns about HIPAA data arose in 2018, when Strava, a fitness company released a global heat map of user's exercise routes. In doing so, it accidentally revealed sensitive military locations across the globe through highlighting the exercise routes of military personnel.The Trump administration has tapped Palantir, a company that specializes in using AI for data analytics, to collate and analyse data about Americans. Meanwhile, Palantir has announced a partnership with a company that runs self-checkout systems.Such partnerships can expand corporate and government reach into everyday consumer behaviour. This one could be used to create detailed personal profiles on Americans by linking their consumer habits with other personal data. This raises concerns about increased surveillance and loss of anonymity. It could allow citizens to be tracked and analysed across multiple aspects of their lives without their knowledge or consent.Some smart device companies are also rolling back privacy protections instead of strengthening them. Amazon recently announced that starting on March 28, 2025, all voice recordings from Amazon Echo devices would be sent to Amazon's cloud by default, and users will no longer have the option to turn this function off. This is different from previous settings, which allowed users to limit private data collection.Changes like these raise concerns about how much control consumers have over their own data when using smart devices. Many privacy experts consider cloud storage of voice recordings a form of data collection, especially when used to improve algorithms or build user profiles, which has implications for data privacy laws designed to protect online privacy.All of this brings up serious privacy concerns for people and governments on how AI tools collect, store, use and transmit data. The biggest concern is transparency. People don't know what data is being collected, how the data is being used, and who has access to that data.Companies tend to use complicated privacy policies filled with technical jargon to make it difficult for people to understand the terms of a service that they agree to. People also tend not to read terms of service documents. One study found that people averaged 73 seconds reading a terms of service document that had an average read time of 29-32 minutes.Data collected by AI tools may initially reside with a company that you trust, but can easily be sold and given to a company that you don't trust.AI tools, the companies in charge of them and the companies that have access to the data they collect can also be subject to cyberattacks and data breaches that can reveal sensitive personal information. These attacks can by carried out by cybercriminals who are in it for the money, or by so-called advanced persistent threats, which are typically nation/state- sponsored attackers who gain access to networks and systems and remain there undetected, collecting information and personal data to eventually cause disruption or harm.While laws and regulations such as the General Data Protection Regulation in the European Union and the California Consumer Privacy Act aim to safeguard user data, AI development and use have often outpaced the legislative process. The laws are still catching up on AI and data privacy. For now, you should assume any AI-powered device or platform is collecting data on your inputs, behaviours and patterns.Although AI tools collect people's data, and the way this accumulation of data affects people's data privacy is concerning, the tools can also be useful. AI-powered applications can streamline workflows, automate repetitive tasks and provide valuable insights.But it's crucial to approach these tools with awareness and caution.When using a generative AI platform that gives you answers to questions you type in a prompt, don't include any personally identifiable information, including names, birth dates, Social Security numbers or home addresses. At the workplace, don't include trade secrets or classified information. In general, don't put anything into a prompt that you wouldn't feel comfortable revealing to the public or seeing on a billboard. Remember, once you hit enter on the prompt, you've lost control of that information.Remember that devices which are turned on are always listening - even if they're asleep. If you use smart home or embedded devices, turn them off when you need to have a private conversation. A device that's asleep looks inactive, but it is still powered on and listening for a wake word or signal. Unplugging a device or removing its batteries is a good way of making sure the device is truly off.Finally, be aware of the terms of service and data collection policies of the devices and platforms that you are using. You might be surprised by what you've already agreed to.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Meta's data broker
Meta's data broker

The Hindu

time6 hours ago

  • The Hindu

Meta's data broker

Alexandr Wang, a 28-year-old entrepreneur, has emerged as a key player in the rapidly evolving landscape of artificial intelligence. His company, Scale AI, has strategically positioned itself at the forefront of this transformation. Founded in 2016 as a data labelling company for autonomous vehicles, Scale AI quickly gained prominence. It secured partnerships with major tech giants such as OpenAI, Microsoft, Nvidia, and even the U.S. government. In a significant development, Meta Platforms has invested $14.3 billion to acquire a 49% stake in Scale AI. This acquisition marks a pivotal moment for the company, which now specialises in data curation and intelligent dataset management. As part of this deal, Mr. Alexandr is expected to assume a leadership role within Meta, leading the company's new 'superintelligence' research lab. This move is a much-needed talent refresh and reputation boost for Meta, which has been losing AI researchers and struggling with its Llama 4 launch. The company is bleeding AI talent as researchers are leaving for rival French startup Mistral. Meta's investment in AI is a significant move, and it aligns with the strategies employed by other tech giants. For instance, Microsoft has backed OpenAI, while Amazon supports Anthropic. These partnerships highlight the crucial role of specialised AI firms in the broader ecosystem. However, the Meta-Scale AI deal has sparked discussions about its implications. Some believe it represents an 'exit' for Scale AI, potentially impacting its ability to serve Meta's competitors like Google and OpenAI. Others question whether it's a strategic partnership to enhance Scale AI's capabilities while bolstering Meta's AI standing, or if it's a talent acquisition move by Meta. The deal's structure, with Mr. Alexandr reportedly retaining voting control over Scale despite his move to Meta, indicates a complex arrangement aimed at balancing these interests. Regardless, this move underscores Meta's commitment to leading the AI race. Mr. Alexandr, a vocal advocate for the strategic importance of AI for the U.S. as a matter of national security, was shaped by his upbringing in Los Alamos, New Mexico, the birthplace of the atomic bomb. In a TED Talk, he drew parallels between the transformative impact of the atomic bomb on past warfare and the current potential of AI to redefine global power. AI arms race Mr. Alexandr asserts that, as seen in the Second World War with the atomic bomb, the country that rapidly and effectively integrates new technology into warfare emerges victorious. He warns that the U.S. risks falling behind in the AI arms race, particularly against China, which he claims is ahead in accumulating the vast datasets necessary to develop powerful AI models. Mr. Alexandr has urged the U.S. government to establish a national data reserve, treating data as a strategic national resource akin to petroleum. Scale AI has actively pursued defence contracts, collaborating with the U.S. Department of Defense, Air Force, and the Army to test and evaluate large language models for military purposes. Mr. Alexandr views this work as a 'moral imperative', especially after witnessing China's advances in AI for surveillance and military applications. Mr. Alexandr predicts that AI will dominate warfare within the next decade. He emphasises that 'data will become a new kind of ammunition in the era of AI warfare'. This underscores his call for robust investment in data infrastructure, not only for military superiority but also as a deterrent, akin to nuclear capabilities. For Mr. Alexandr and Scale AI, the mission is clear: to provide the foundational data layer that accelerates AI development across all sectors, including national security. While Meta's significant investment marks a new chapter for Scale AI, potentially altering its client dynamics, Mr. Alexandr's core message about the paramount importance of data and the strategic implications of AI remains unchanged.

Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...
Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...

Time of India

time10 hours ago

  • Time of India

Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...

Nvidia CEO Jensen Huang strongly refuted Anthropic CEO Dario Amodei's predictions regarding AI-driven job automation during a press briefing. Huang dismissed Amodei's claim that AI could eliminate 50% of entry-level white-collar jobs within five years, advocating for transparent AI development. While acknowledging job transformation, Huang emphasized AI's potential to create new opportunities and boost productivity. Nvidia CEO Jensen Huang expressed sharp disagreement with Anthropic CEO Dario Amodei 's predictions about AI-driven job automation during a press briefing at VivaTech in Paris. Huang specifically challenged Amodei's recent claim that AI could eliminate up to 50% of entry-level white-collar jobs within five years, dismissing the notion as overly alarmist. 'I pretty much disagree with almost everything [Amodei] says,' Huang stated, critiquing three key points he attributed to Amodei: that AI is so dangerous only a select few should develop it, that it's too costly for widespread development, and that its power will lead to massive job losses. 'If you want things to be done safely and responsibly, you do it in the open … Don't do it in a dark room and tell me it's safe,' Huang added, advocating for transparent and collaborative AI development. What Anthropic CEO said on AI Amodei, in a recent Axios interview, had warned that AI could disrupt half of entry-level office roles and urged policymakers to prepare for economic impacts. Huang countered that while AI will transform jobs, it will also create new opportunities, citing increased productivity as a driver for hiring. 'Some jobs will be obsolete, but many jobs are going to be created,' he said. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like 선납금·위약금 없이 벤츠를 원하는 기간만큼 이용하세요. 계약 만료시 반납, 인수 모두 가능합니다. 써치오토모빌 더 알아보기 Undo Anthropic on Nvidia CEO Jensen Huang's statement In response, an Anthropic spokesperson clarified to Fortune that Amodei has never claimed only Anthropic should build AI. 'Dario has advocated for a national transparency standard for AI developers … so the public and policymakers are aware of the models' capabilities and risks,' the statement read. The spokesperson reaffirmed Amodei's concerns about AI's economic impact and his commitment to addressing them. Anthropic, founded in 2021 by Amodei and other ex-OpenAI researchers, prioritizes safety in AI development, a focus rooted in the founders' reported concerns over OpenAI's direction. The public clash between Huang and Amodei highlights ongoing debates in the AI industry about its societal and economic implications. AI Masterclass for Students. Upskill Young Ones Today!– Join Now

Google, Scale AI's largest customer, plans split after Meta deal
Google, Scale AI's largest customer, plans split after Meta deal

Time of India

time11 hours ago

  • Time of India

Google, Scale AI's largest customer, plans split after Meta deal

HighlightsAlphabet's Google plans to sever ties with Scale AI after rival Meta Platforms Inc. acquires a 49% stake in the AI data-labeling startup, potentially impacting Scale's revenue significantly. The acquisition of Scale AI by Meta Platforms Inc. raises concerns among competing AI companies, as they fear that sharing proprietary data with Scale could expose their research strategies and technical blueprints. Following the news of Meta's investment, competitors of Scale AI, such as Turing and Labelbox, anticipate a surge in business as companies seek alternative data-labeling services. Alphabet's Google , the largest customer of Scale AI , plans to cut ties with Scale after news broke that rival Meta is taking a 49% stake in the AI data-labeling startup, five sources familiar with the matter told Reuters. Google had planned to pay Scale AI about $200 million this year for the human-labeled training data that is crucial for developing technology, including the sophisticated AI models that power Gemini, its ChatGPT competitor, one of the sources said. The search giant already held conversations with several of Scale AI's rivals this week as it seeks to shift away much of that workload, sources added. Scale's loss of significant business comes as Meta takes a big stake in the company, valuing it at $29 billion. Scale was worth $14 billion before the deal. Scale AI intends to keep its business running while its CEO, Alexandr Wang, along with a few employees, move over to Meta. Since its core business is concentrated around a few customers, it could suffer greatly if it loses key customers like Google. In a statement, a Scale AI spokesperson said its business, which spans work with major companies and governments, remains strong, as it is committed to protecting customer data. The company declined to comment on specifics with Google. Scale AI raked in $870 million in revenue in 2024, and Google spent some $150 million on Scale AI's services last year, sources said. Other major tech companies that are customers of Scale's, including Microsoft, are also backing away. Elon Musk's xAI is also looking to exit, one of the sources said. OpenAI decided to pull back from Scale several months ago, according to sources familiar with the matter, though it spends far less money than Google. OpenAI's CFO that the company will continue to work with Scale AI, as one of its many data vendors. Companies that compete with Meta in developing cutting-edge AI models are concerned that doing business with Scale could expose their research priorities and road map to a rival, five sources said. By contracting with Scale AI, customers often share proprietary data as well as prototype products for which Scale's workers are providing data-labeling services. With Meta now taking a 49% stake, AI companies are concerned that one of their chief rivals could gain knowledge about their business strategy and technical blueprints. Google, Microsoft and OpenAI declined to comment. xAI did not respond to a request for comment. RIVALS SEE OPENINGS The bulk of Scale AI's revenue comes from charging generative AI model makers for providing access to a network of human trainers with specialized knowledge - from historians to scientists, some with doctorate degrees. The humans annotate complex datasets that are used to "post-train" AI models, and as AI models have become smarter, the demand for the sophisticated human-provided examples has surged, and one annotation could cost as much as $100. Scale also does data-labeling for enterprises like self-driving car companies and the U.S. government, which are likely to stay, according to the sources. But its biggest money-maker is in partnering with generative AI model makers, the sources said. Google had already sought to diversify its data service providers for more than a year, three of the sources said. But Meta's moves this week have led Google to seek to move off Scale AI on all its key contracts, the sources added. Because of the way data-labeling contracts are structured, that process could happen quickly, two sources said. This will provide an opening for Scale AI's rivals to jump in. "The Meta-Scale deal marks a turning point," said Jonathan Siddharth, CEO of Turing, a Scale AI competitor. "Leading AI labs are realizing neutrality is no longer optional, it's essential." Labelbox, another competitor, will "probably generate hundreds of millions of new revenue" by the end of the year from customers fleeing Scale, its CEO, Manu Sharma, told Reuters. Handshake, a competitor focusing on building a network of PhDs and experts, saw a surge of workload from top AI labs that compete with Meta. "Our demand has tripled overnight after the news," said Garrett Lord, CEO at Handshake. Many AI labs now want to hire in-house data-labelers, which allows their data to remain secure, said Brendan Foody, CEO of Mercor, a startup that in addition to competing directly with Scale AI also builds technology around being able to recruit and vet candidates in an automated way, enabling AI labs to scale up their data labeling operations quickly. Founded in 2016, Scale AI provides vast amounts of labeled data or curated training data, which is crucial for developing sophisticated tools such as OpenAI's ChatGPT. The Meta deal will be a boon for Scale AI's investors including Accel and Index Ventures, as well as its current and former employees. As part of the deal, Scale AI's CEO, Wang, will take a top position leading Meta's AI efforts. Meta is fighting the perception that it may have fallen behind in the AI race after its initial set of Llama 4 large language models released in April fell short of performance expectations.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store