logo
Scale AI has a data security problem

Scale AI has a data security problem

Business Insider13 hours ago

As Scale AI seeks to reassure customers that their data is secure following Meta's $14.3 billion investment, leaked files and the startup's own contractors indicate it has some serious security holes.
Scale AI routinely uses public Google Docs to track work for high-profile customers like Google, Meta, and xAI, leaving multiple AI training documents labeled "confidential" accessible to anyone with the link, Business Insider found.
Contractors told BI the company relies on public Google Docs to share internal files, a method that's efficient for its vast army of at least 240,000 contractors and presents clear cybersecurity and confidentiality risks.
Scale AI also left public Google Docs with sensitive details about thousands of its contractors, including their private email addresses and whether they were suspected of "cheating." Some of those documents can be viewed and also edited by anyone with the right URL.
There's no indication that Scale AI has suffered a breach because of this. Two cybersecurity experts told BI that such practices could leave the company and its clients vulnerable to various kinds of hacks, such as hackers impersonating contractors or uploading malware into accessible files.
Scale AI told Business Insider it takes data security seriously and is looking into the matter.
"We are conducting a thorough investigation and have disabled any user's ability to publicly share documents from Scale-managed systems," a Scale AI spokesperson said. "We remain committed to robust technical and policy safeguards to protect confidential information and are always working to strengthen our practices."
Meta declined to comment. Google and xAI didn't respond to requests for comment.
In the wake of Meta's blockbuster investment, clients like Google, OpenAI, and xAI paused work with Scale. In a blog post last week, Scale reassured Big Tech clients that it remains a neutral and independent partner with strict security standards.
The company said that "ensuring customer trust has been and will always be a top priority," and that it has "robust technical and policy safeguards to protect customers' confidential information."
BI's findings raise questions about whether it did enough to ensure security and whether Meta was aware of the issue before writing the check.
Confidential AI projects were accessible
BI was able to view thousands of pages of project documents across 85 individual Google Docs tied to Scale AI's work with Big Tech clients. The documents include sensitive details, such as how Google used ChatGPT to improve its own struggling chatbot, then called Bard.
Scale also left public at least seven instruction manuals marked "confidential" by Google, which were accessible to anyone with the link. Those documents spell out what Google thought was wrong with Bard — that it had difficulties answering complex questions — and how Scale contractors should fix it.
For Elon Musk's xAI, for which Scale ran at least 10 generative AI projects as of April, public Google documents and spreadsheets show details of "Project Xylophone," BI reported earlier this month. Training documents and a list of 700 conversation prompts revealed how the project focused on improving the AI's conversation skills about a wide array of topics, from zombie apocalypses to plumbing.
Meta training documents, marked confidential at the top, were also left public to anyone with the link. These included links to accessible audio files with examples of "good" and "bad" speech prompts, suggesting the standards Meta set for expressiveness in its AI products.
Some of those projects focused on training Meta's chatbots to be more conversational and emotionally engaging while ensuring they handled sensitive topics safely, BI previously reported. As of April, Meta had at least 21 generative AI projects with Scale.
Several Scale AI contractors interviewed by BI said it was easy to figure out which client they worked for, even though they were codenamed, often just from the nature of the task or the way the instructions were phrased. Sometimes it was even easier: One presentation seen by BI had Google's logo.
Even when projects were meant to be anonymized, contractors across different projects described instantly recognizing clients or products. In some cases, simply prompting the model or asking it directly which chatbot it was would reveal the underlying client, contractors said.
Scale AI left contractor information public
Other Google Docs exposed sensitive personal information about Scale's contractors. BI reviewed spreadsheets that were not locked down and that listed the names and private Gmail addresses of thousands of workers. Several contacted by BI said they were surprised to learn their details were accessible to anyone with the URL of the document.
Many documents include details about their work performance.
One spreadsheet titled "Good and Bad Folks" categorizes dozens of workers as either "high quality" or suspected of "cheating." Another list of hundreds of personal email addresses is titled "move all cheating taskers," which also flagged workers for "suspicious behavior."
Another sheet names nearly 1,000 contractors who were "mistakenly banned" from Scale AI's platforms.
Other documents show how much individual contractors were paid, along with detailed notes on pay disputes and discrepancies.
The system seemed 'incredibly janky'
Five current and former Scale AI contractors who worked on separate projects told BI that the use of public Google Docs was widespread across the company.
Contractors said that using them streamlined operations for Scale, which relies mostly on freelance contributors. Managing individual access permissions for each contractor would have slowed down the process.
Scale AI's internal platform requires workers to verify themselves, sometimes using their camera, contractors told BI.
At the same time, many documents containing information on training AI models can be accessed through public links or links in other documents without verification.
"The whole Google Docs system always seemed incredibly janky," one worker said.
Two other workers said they retained access to old projects they no longer worked on, which were sometimes updated with requests from the client company regarding how the models should be trained.
'Of course it's dangerous'
Organizing internal work through public Google Docs can create serious cybersecurity risks, Joseph Steinberg, a Columbia University cybersecurity lecturer, told BI.
"Of course it's dangerous. In the best-case scenario, it's just enabling social engineering," he said.
Social engineering refers to attacks where hackers trick employees or contractors into giving up access, often by impersonating someone within the company.
Leaving details about thousands of contractors easily accessible creates many opportunities for that kind of breach, Steinberg said.
At the same time, investing more in security can slow down growth-oriented startups.
"The companies that actually spend time doing security right very often lose out because other companies move faster to market," Steinberg said.
The fact that some of the Google Docs were editable by anyone creates risks, such as bad actors inserting malicious links into the documents for others to click, Stephanie Kurtz, a regional director at cyber firm Trace3, told BI.
Kurtz added that companies should start with managing access via invites.
"Putting it out there and hoping somebody doesn't share a link, that's not a great strategy there," she said.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

‘Big Balls' No Longer Works for the US Government
‘Big Balls' No Longer Works for the US Government

WIRED

time28 minutes ago

  • WIRED

‘Big Balls' No Longer Works for the US Government

Photograph:Edward 'Big Balls' Coristine, one of the first technologists hired as part of Elon Musk's Department of Government Efficiency (DOGE), is no longer working for the federal government, according to multiple sources. 'Edward Coristine resigned yesterday,' a White House official tells WIRED. Coristine received full-time employment status at the GSA late last month, as reported by WIRED. As of Tuesday afternoon, his Google Suite account with the General Services Administration (GSA) was no longer active, according to a source with direct knowledge. His name also no longer appears on a White House contact list of current DOGE employees on the federal payroll maintained by a senior administration official, the official says. In May, Coristine appeared to be attending regular meetings with departments and agencies, including a May 5 meeting with the Commerce Department to discuss Trump's golden Visa, a May 15 meeting to discuss implementing the DOGE agenda at the military, and a May 22 meeting with the Treasury Department. He was listed on a report regarding the GSA workforce on June 10. That report comes out once a month. Previously, Coristine was one of a handful of DOGE-affiliated staffers who were occasionally seen at GSA headquarters in Washington DC. He worked from the sixth floor of the office, which was cordoned off for employees affiliated with DOGE and the agency's leadership team. The GSA and Coristine did not immediately respond to requests for comment from WIRED. When Musk exited government in an official capacity last month, DOGE's future was uncertain. Musk's top lieutenant, Steve Davis, president of the Boring Company, another Musk venture, also announced he would leave DOGE with Musk. Sahil Lavingia, a former DOGE member who WIRED first identified at the Department of Veterans' Affairs says, 'I have heard since Elon and Steve have supposedly departed, they've terminated a lot of those that got hired. They're all still probationary, right? There's that two-year probationary period. So there's a good chance that a lot of those people end up getting fired anyway.' Coristine, according to Lavingia, was one of a small group of technologists who were highly trusted within DOGE and deployed across multiple federal agencies, and given multiple federal laptops. 'The core group of pre-inauguration engineers joked about how many laptops they had. It was almost like a competition in the sense to have seven, eight different laptops that they would run around with,' says Lavingia. During his time in government—a period in which DOGE sought to dismantle large parts of the federal bureaucracy, obtained direct access to sensitive federal payment systems, and worked to join siloed systems together to enable surveillance of people in the US—Coristine appeared at the GSA, Office of Personnel Management, the US Agency for International Development, the Department of Education, and the Small Business Administration. Coristine, who has gone by the handle 'Big Balls' online, is a 19-year-old high school graduate who worked at Neuralink for several months and founded a company called LLC in 2021. The company owns several domains including two registered in Russia, one of which is an AI Discord bot. He also briefly worked for a company founded by reformed black hat hackers called the Path Network, and a Telegram handle associated with Coristine appears to have solicited a distributed denial of service attack.

Snowflake Inc. (SNOW): A Bull Case Theory
Snowflake Inc. (SNOW): A Bull Case Theory

Yahoo

time29 minutes ago

  • Yahoo

Snowflake Inc. (SNOW): A Bull Case Theory

We came across a bullish thesis on Snowflake Inc. (SNOW) on Sergey's Substack. In this article, we will summarize the bulls' thesis on SNOW. Snowflake Inc. (SNOW)'s share was trading at $209.12 as of 16th June. SNOW's forward P/E was 185.19 respectively according to Yahoo Finance. A software company's engineer staring at a computer monitor with intense concentration. Snowflake is rapidly evolving from a data warehouse into an AI-native data cloud platform, with significant momentum under CEO Sridhar Ramaswamy. In Q1 FY2026, Snowflake accelerated innovation with over 125 new product launches—a 100% YoY increase—including Cortex Analyst and Agent, which enable natural language queries and AI workflows. The company is also repositioning its Marketplace as a core distribution hub for AI-native apps, now supporting agentic workflows through Cortex Knowledge Extensions and seamless data ingestion via OpenFlow. While new listings slowed, this reflects a shift toward higher-quality offerings as Snowflake updates its revenue-sharing and governance models. Cortex now supports top-tier LLMs like Llama 4 and OpenAI on Azure, powering weekly usage by over 5,200 enterprises, including BlackRock and WHOOP. Gen2 compute and AISQL offer major performance and cost improvements, while Data Science Agent and Semantic Model sharing extend ML accessibility and consistency. Strategic acquisitions like Crunchy Data strengthen transactional and public sector capabilities, where Snowflake recently secured DoD-level authorization. Customer momentum remains strong: 419 net adds (+18% YoY), 26 new $ 1 M+ ARR clients, and two $ $100 M+ deals. Industry wins include AstraZeneca, Dentsu, Siemens, Samsung Ads, and Kraft Heinz. Retention remains best-in-class at 124%, with net new ARR up 4% YoY and RPO growing 34.3%. Despite lower FCF margins (17.6%) and high SBC (39% of revenue), Snowflake's ongoing shift to back-end weighted consumption and active buybacks reflects disciplined capital management. With high switching costs, a growing network effect (39% of usage now data sharing), and increasing AI adoption, Snowflake's long-term competitive moat appears to be deepening. Previously, we highlighted a bullish thesis on Snowflake Inc. (SNOW) from Compounding Your Wealth, which framed the company as an enterprise data leader accelerating growth through AI integration, margin expansion, and robust platform adoption across sectors like financial services and public institutions. Sergey's thesis builds on this foundation, emphasizing Snowflake's transformation into an AI-native data cloud platform under new CEO Sridhar Ramaswamy. It underscores how Snowflake's 125+ product launches, such as Cortex Agent, Gen2 compute, and OpenFlow, are redefining its value proposition and extending its reach into high-performance analytics and AI-native app distribution. Snowflake Inc. (SNOW) is not on our list of the 30 Most Popular Stocks Among Hedge Funds. As per our database, 94 hedge fund portfolios held SNOW at the end of the first quarter which was 85 in the previous quarter. While we acknowledge the risk and potential of SNOW as an investment, our conviction lies in the belief that some AI stocks hold greater promise for delivering higher returns and have limited downside risk. If you are looking for an extremely cheap AI stock that is also a major beneficiary of Trump tariffs and onshoring, see our free report on the best short-term AI stock. READ NEXT: 8 Best Wide Moat Stocks to Buy Now and 30 Most Important AI Stocks According to BlackRock.

‘Techno King' Elon Musk Doesn't Own a Computer, His Lawyers Tell Court
‘Techno King' Elon Musk Doesn't Own a Computer, His Lawyers Tell Court

Gizmodo

time30 minutes ago

  • Gizmodo

‘Techno King' Elon Musk Doesn't Own a Computer, His Lawyers Tell Court

He fancies himself the innovator of our time. 'Disrupt' could be his middle name. Technology, it seems, is in his DNA. And yet, Elon Musk apparently—supposedly—has a secret Luddite streak: he doesn't use a computer. The revelation didn't come from a biography or a tell-all interview. It came from a legal filing in the high-stakes, mud-slinging lawsuit between Musk and OpenAI, where the future of artificial intelligence is on trial in a Northern California courtroom. Amid thorny questions of corporate betrayal and billion-dollar secrets, this strange detail stole the show. Yes, you heard that correctly. It's not a joke from a late-night show. That's not me saying it. It's coming from Musk's own lawyers. In a legal letter filed on June 22, Musk's legal team pushed back against accusations from OpenAI that they were failing to turn over relevant documents. When OpenAI claimed Musk's team was refusing to collect certain materials, his lawyers called the accusation 'incorrect' and, in the process, dropped the bombshell. 'Mr. Musk does not use a computer,' his lawyers at Toberoff & Associates wrote on the first page of the three-page document. There's just one problem with that claim: public evidence, including from Musk himself, suggests otherwise. While employees at X told WIRED Musk primarily works from his phone, they also note he has been seen using a laptop on occasion. More pointedly, Musk has referenced owning a computer in his own social media posts. In a December 2024 post on X, he shared an image with the text, 'This is a pic of my laptop,' explaining that he was using it to test Starlink's streaming capabilities in-flight. More recently, in May 2025, when asked about his gaming setup, Musk replied on X that he is 'still using my ancient PC laptop with the @DOGE sticker made long ago by a fan.' This is a pic of my laptop. It's about 3 years old. A guy in Germany gave me this cool sticker, so I don't want to upgrade it and lose the sticker. — Elon Musk (@elonmusk) December 31, 2024This contradiction emerges from the messy 'discovery' phase of his lawsuit against OpenAI, where he accuses CEO Sam Altman of betraying their founding mission. As both sides fight over internal documents, the battle has intensified. This context makes the 'no computer' claim seem less like a personal quirk and more like a potential legal tactic to limit the scope of discoverable documents. After all, if there are no computers, there are no computer files to hand over. The Musk v. Altman case is a proxy war over the governance and ownership of AI. Musk, a co-founder of OpenAI, now portrays himself as its most prominent critic, arguing the company sold its soul to Microsoft. OpenAI, in turn, depicts Musk as a bitter ex-partner trying to interfere with a company he chose to leave. But for now, the legal drama is being overshadowed by a bizarre claim that is seemingly contradicted by Musk's own public statements: the man suing over the future of artificial intelligence may be trying to persuade a judge that he has personally abandoned one of the most fundamental tools of the digital world.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store