
AI-based mammography is here, and it has a trust problem
Enthusiasm is growing for the technology, as prospective trials conducted in Europe suggest that some AI tools can detect more cancers than radiologists alone. But beyond radiology networks that are implementing their own algorithms, many imaging centers and the radiologists who work for them are still skeptical.
Advertisement
'It'll take time for us to really gain a lot of trust and confidence in it,' said Manisha Bahl, a breast radiologist at Massachusetts General Hospital, which is currently testing several AI-based tools for mammography in a head-to-head study.
One large
Advertisement
Radiologists were much less likely to call patients for follow-up when their mammogram was flagged by the AI than when it was flagged by a human radiologist — even though the AI-flagged cases were more likely to be cancerous. If both members of a human-AI pair flagged a case for possible malignancy, only 39 percent of patients were called back. If it was two radiologists that flagged a case, callbacks shot up to 57 percent. The researchers used technology from South Korean developer Lunit, which also partly funded the study.
'Providing a superior performing algorithm doesn't actually necessarily improve human performance,' said Adam Rodman, a clinical reasoning researcher and practicing internist who co-directs the iMED initiative at Beth Israel Deaconess Medical Center. 'This study does a really good job at explaining why that is, and it's trust.'
It's a common refrain for breast radiologists, who have been burned by technology before. Computer-aided detection tools for mammograms became standard starting in the 1980s, but in the long run, they never improved cancer detection or recall rates. 'We have to be very careful what we do with AI once it's out in the wild,' said Etta Pisano, chief research officer at the American College of Radiology, at the meeting of the Radiological Society of North America in December.
Advertisement
Performance for breast cancer AI — which marks or annotates suspicious lesions on a mammogram and provides a score or some indication of the likelihood of malignancy — is typically reported based on how well it identifies cancers in databases of old mammograms that have been previously screened by human radiologists. But in practice, that performance can vary significantly from center to center and radiologist to radiologist.
The radiologist's level of trust, the way the software is integrated into their
With all those variables, how can a radiologist make sure that they're using AI the most effective way? 'The answer is, nobody knows,' said Rodman.
Normally, Bahl, the radiologist at Mass General, opens up a mammogram and interprets it entirely on her own before she activates any AI-based image processing. That's to cut down on the risk of automation bias — where she learns to rely too much on the machine. That approach makes sense for radiologists who trust their professional judgement more than a new tool, and it aligns with surveys of patients that suggest they typically want a doctor making the final call on image analysis, said Sanjay Aneja, a radiation oncologist at Yale Cancer Center who studies mammography AI.
But that approach doesn't fulfill AI's other main promise: the potential for efficiency. 'If anything, it might be a little bit less efficient,' said Aneja. With skilled radiologists in shortage, AI will ideally help them work smarter and faster — which means looking at the algorithm's output upfront.
Advertisement
At radiology practices that are more aggressively leaning into mammogram AI, 'radiologists gain confidence and start to look earlier in the process,' said Chris McKinney, director of North American sales for Lunit.
That kind of comfort doesn't emerge unless radiologists are getting regular practice with a new system — and adoption of the tools is inconsistent, since they aren't reimbursed by insurance. Instead, most practices deploying the tools ask patients to pay a cash fee of $40 to $90 for an AI add-on. At the top-adopting practice within Rezolut, a network of more than 40 US radiology practices that use Lunit's breast AI system, about 40 percent of patients opt to pay.
'Every radiologist is going to have to work with the AI product and get more comfortable with it and then utilize it individually,' said Stamatia Destounis, chair of the American College of Radiology's breast imaging commission. 'I don't think everyone's going to use it the same way.'
But 'practice makes perfect' isn't enough for many AI researchers and physicians, who want more explicit guidance when it comes to appropriate deployment of these tools as they become more widely used. 'The problem is that there's actually very little guardrails with a lot of these devices as they get put out there,' said Aneja.
Imaging quality can vary significantly between radiology practices, for example. 'There's very few algorithms that say, 'This image is of poor quality, can't evaluate it,'' said Aneja. 'We want an algorithm to be able to say what they don't know.'
Advertisement
Chiara Corti, an oncologist from Italy and clinical fellow at Dana-Farber, calls for more disclosure of the race and ethnicity of patients whose mammograms were used to train AI tools to ensure accuracy across groups. (ACR's AI
The performance — and relative value — of an AI tool also depends on the radiologists who are using it. Not every radiologist is an expert in reading mammograms, and 'one of the biggest benefits of AI is to disseminate that level of knowledge,' said Aneja.
To ensure radiology practices use this new generation of algorithms in a way that drives better cancer outcomes, they need more real-world, prospective research in the US 'We need to see those human interactions,' said Destounis. 'It's impossible to deduce from a study going back in time how every radiologist is going to interact with the AI system in practice, daily.'
A 2024
Advertisement
That kind of real-world evidence is one of the goals of a program out of ARPA-H called ACTR, for Advancing Clinical Trial Readiness. Pisano, the program's director, said at the December RSNA meeting that ACTR planned to test AI products in real-world, pragmatic trials out of imaging centers across the country, most likely starting with trials of breast cancer screening algorithms that could include hundreds of thousands of mammograms.
The trials, she said at the time, were planned to start this spring. They have yet to be announced.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
24 minutes ago
- Yahoo
Cathie Wood sells $22.8 million of hot stock near all-time highs
Cathie Wood sells $22.8 million of hot stock near all-time highs originally appeared on TheStreet. Cathie Wood has long been aggressive in hunting tech stocks that she believes will have a 'disruptive' impact on the future world. However, she sometimes sells a stock when it is high to secure gains. In the past week, the head of Ark Investment Management sold a popular AI stock that has surged nearly 70% year-to-date. 💵💰Don't miss the move: Subscribe to TheStreet's free daily newsletter 💰💵 Cathie Wood's investments have had a volatile ride this year, swinging from strong gains to sharp losses, and now back to outperforming the broader market. In January and February, the Ark funds rallied as investors bet on the Trump administration's potential deregulation that could benefit Wood's tech bets. But the funds stumbled in the following weeks, underperforming sharply as several of its top holdings —especially Tesla, its largest position — declined amid macroeconomic and trade policy uncertainties. Now, the fund is regaining momentum. As of June 6, the flagship Ark Innovation ETF () is up 6.11% year-to-date, outpacing the S&P 500's 2.02% gain. Wood gained a remarkable 153% in 2020, which helped build her reputation and attract loyal investors. Still, her long-term performance has made many others skeptical of her aggressive style. As of June 6, Ark Innovation ETF, with $5 billion under management, has delivered a five-year annualized return of negative 0.5%. In comparison, the S&P 500 has an annualized return of 15.18% over the same period. Wood's investment strategy is straightforward: Her Ark ETFs typically buy shares in emerging high-tech companies in fields such as artificial intelligence, blockchain, biomedical technology and robotics. Wood says these companies have the potential to reshape industries, but their volatility leads to major fluctuations in Ark funds' Ark Innovation ETF wiped out $7 billion in investor wealth over the 10 years ending in 2024, according to an analysis by Morningstar's analyst Amy Arnott. That made it the third-biggest wealth destroyer among mutual funds and ETFs in Arnott's ranking. Wood said the U.S. is coming out of a three-year 'rolling recession' and heading into a productivity-led recovery that could trigger a broader bull market. In a letter to investors published on April 30, she dismissed predictions of a recession dragging into 2026, as she expects "more clarity on tariffs, taxes, regulations, and interest rates over the next three to six months." "If the current tariff turmoil results in freer trade, as tariffs and non-tariff barriers come down in tandem with declines in other taxes, regulations, and interest rates, then real GDP growth and productivity should surprise on the high side of expectations at some point during the second half of this year," she wrote. She also struck an optimistic tone for tech stocks. "During the current turbulent transition in the US, we think consumers and businesses are likely to accelerate the shift to technologically enabled innovation platforms including artificial intelligence, robotics, energy storage, blockchain technology, and multiomics sequencing," she said. But not everyone shares Wood's bullish outlook. Her flagship Ark Innovation ETF has seen $2.23 billion in net outflows over the past year through June 5, including nearly $154 million in the last month alone, according to ETF research firm VettaFi. From June 2 to June 5, Wood's Ark funds sold 179,846 shares of Palantir Technologies () , which was valued at roughly $22.8 million. Palantir is known for providing AI-driven data analytics software to the U.S. government, military, and commercial clients worldwide, including JPMorgan Chase, Airbus, and Merck. The company reported stronger-than-expected first-quarter revenue in early May and raised its full-year outlook as demand for AI tools increased. 'We are delivering the operating system for the modern enterprise in the era of AI,' CEO Alex Karp said. While many tech stocks have struggled this year, Palantir has stood out. Its shares are up roughly 69% in 2025 and just hit a record close of $133.17 on June of the recent momentum comes from its government work. Back in May 2024, Palantir won a $480 million, five-year U.S. Army contract to build its Maven Smart System, which is a battlefield AI prototype. Last month, the Defense Department modified the contract, increasing the licensing ceiling from $480 million to $1.275 billion. Palantir's Foundry platform has been adopted by at least four federal agencies, including the Department of Homeland Security and the Department of Health and Human Services, according to a New York Times report published May 30. Fannie Mae also announced a partnership with Palantir in May to work on AI-based fraud detection. However, the New York Times article also raised concerns about the company's relationship with the Trump administration, alleging that the U.S. president could use Palantir's technology to target immigrants and political opponents. The article also claimed that some Palantir employees felt uncomfortable with the company's decision to work with the Trump administration and that it "risks becoming the face of Mr. Trump's political agenda." Palantir responded in a June 3 post on X, denying the accusations. More Palantir Palantir gets great news from the Pentagon Wall Street veteran doubles down on Palantir Palantir bull sends message after CEO joins Trump for Saudi visit 'The recently published article by The New York Times is blatantly untrue,' the company wrote. 'Palantir never collects data to unlawfully surveil Americans.' Palantir remains a core position for Wood even after recent trims. The stock is now the 9th largest holding in the ARK Innovation ETF, accounting for 4.54%. Wood's latest trades in the past week include buying shares of Advanced Micro Devices () , () , Guardant Health () and Veracyte () . At the same time, she trimmed positions in Tesla () , Roblox () , Robinhood () , and Meta Platforms () .Cathie Wood sells $22.8 million of hot stock near all-time highs first appeared on TheStreet on Jun 8, 2025 This story was originally reported by TheStreet on Jun 8, 2025, where it first appeared.


Forbes
31 minutes ago
- Forbes
Why We Need Global Prosocial AI Governance — Now
Abstract image of a person's profile symbolically composed of dials of different sizes. Concept man, ... More time, space The artificial intelligence revolution isn't coming — it's here. But unlike previous technological waves, AI's transformative power is being concentrated in the hands of remarkably few players, creating global imbalances that threaten to entrench existing inequalities for generations. As AI systems increasingly shape our economies, societies, and daily lives, we face a critical choice: Will we allow narrow market forces and geopolitical power dynamics to dictate AI's development, or will we proactively steer this technology toward benefiting humanity as a whole? It is late to set the stage for global prosocial AI governance, but it is not too late – yet. Before examining governance frameworks, we must confront an uncomfortable truth: the AI revolution is built on a foundation of extreme market concentration that makes Big Tech's dominance look almost quaint by comparison. Nvidia controls approximately 80 percent of revenues and shipments for datacenter GPU computing, the essential infrastructure powering modern AI systems. This isn't just market leadership — it's approaching technological hegemony. The implications extend far beyond corporate balance sheets. Collectively, the global south is home to just over 1 percent of the world's top computers, and Africa just 0.04 percent. Meanwhile, the U.S. government further restricts AI chip and technology exports, dividing up the world to keep advanced computing power in the United States and among its allies. This creates what development economists call a digital colonialism scenario — entire regions become structurally dependent on technology controlled by a handful of corporations and governments. The concentration isn't limited to hardware. Three cloud providers — Amazon, Microsoft, and Google — control over 65% of global cloud infrastructure, creating additional bottlenecks for AI access. When you need specialized chips from one company, hosted on infrastructure controlled by three others, and governed by regulations written primarily in wealthy nations, the barriers to entry become virtually insurmountable for most of the world's population. This hardware concentration translates into stark global inequalities that dwarf previous technological divides. The economic and social benefits of AI remain geographically concentrated, primarily in the Global North. But unlike the gradual rollout of previous technologies like the internet or mobile phones, AI's infrastructure requirements create immediate exclusion rather than delayed adoption. Consider the practical reality: training a state-of-the-art AI model requires computational resources that cost millions of dollars and consume as much electricity as entire cities. The rise of AI could exacerbate both within-country and between-country inequality, placing upward pressure on global inequality as high-income individuals and regions benefit disproportionately while resource-poor regions risk being left behind. This creates a vicious cycle. Countries and regions without access to AI infrastructure become less competitive economically, reducing their ability to invest in the very infrastructure they need to participate in the AI economy. Meanwhile, AI-enabled automation threatens to disrupt traditional export industries that many developing economies rely on, from manufacturing to service outsourcing. The result is what economists call premature deindustrialization — developing countries losing industrial competitiveness before achieving full industrialization. But now it's happening at digital speed, compressed from decades into years. Yet maybe the fundamental challenge with AI isn't the technology itself — it's the intention behind its development and deployment, now amplified by a sharpened concentration of control. Today's AI systems are predominantly designed to maximize engagement, extract value, or optimize narrow business metrics determined by a small number of actors. Social media algorithms amplify divisive content because controversy drives clicks. Hiring algorithms perpetuate bias because they're trained on historical data that reflects past discrimination. Financial AI systems may optimize for short-term profits while creating systemic risks. This is where prosocial AI governance becomes essential. Unlike traditional regulatory approaches that focus on constraining harmful outcomes, prosocial AI governance aims to actively incentivize beneficial behaviors from the outset. ProSocial AI can enhance access to essential services, improve efficiency in resource use, and promote sustainable practices across all levels of society — but only if we design governance systems that prioritize broad-based benefits over narrow optimization. The global AI regulation landscape is fragmented and rapidly evolving. Earlier optimism that global policymakers would enhance cooperation and interoperability within the regulatory landscape now seems distant. The European Union has pioneered comprehensive AI regulation through its AI Act, while other jurisdictions take vastly different approaches — from the United States' innovation-first philosophy to China's state-directed development model. This fragmentation creates several problems. First, it allows AI developers to engage in regulatory arbitrage, developing systems in jurisdictions with the most permissive rules. Second, it prevents the emergence of global standards that could ensure AI systems operate prosocially across borders. Third, it creates competitive disadvantages for companies that voluntarily adopt higher ethical standards. Given the borderless nature of this issue, an internationally coordinated response is necessary. AI systems don't respect national boundaries — a biased hiring algorithm developed in one country can perpetuate discrimination globally, while misinformation generated by AI can destabilize societies worldwide. Traditional regulatory approaches tend to prove inadequate for rapidly evolving technologies. By the time regulators identify and respond to harms, the damage has already been done. Prosocial AI governance offers a different approach: building beneficial outcomes into the DNA of AI systems from the beginning. This means designing AI systems that actively promote human flourishing rather than merely avoiding harm. Instead of social media algorithms that maximize engagement at all costs, we need systems that promote constructive dialogue and community building. Rather than AI systems that automate away human jobs without consideration for displaced workers, we need technologies that augment human capabilities and create new opportunities for meaningful work. Companies with strong environmental, social, and governance frameworks, enhanced by AI, outperform competitors financially and foster greater brand loyalty. This suggests that prosocial AI isn't just morally imperative — it's also economically advantageous for businesses that adopt it early. Forward-thinking business leaders are beginning to recognize that prosocial AI governance isn't a constraint on innovation—it's a competitive advantage. Organizations that proactively embed prosocial values into their AI systems build stronger relationships with customers, employees, and communities. They reduce regulatory risk, attract top talent who want to work on meaningful problems, and position themselves as leaders in an increasingly values-driven marketplace. Moreover, prosocial AI often leads to better technical outcomes. Systems designed with diverse stakeholders in mind tend to be more robust, adaptable, and effective across different contexts. AI systems built with fairness and transparency as core requirements often discover innovative solutions that benefit everyone. The economic argument becomes even stronger when considering systemic risks. AI systems that prioritize narrow optimization over broader social welfare can create negative externalities that ultimately harm the very markets they operate in. Financial AI that ignores systemic risk can contribute to market crashes. Recommendation systems that polarize societies can undermine the social cohesion that stable markets depend on. Establishing global prosocial AI governance requires coordinated action across multiple levels. International bodies need to develop frameworks that incentivize prosocial AI development while allowing for innovation and adaptation to local contexts. These frameworks should focus on outcomes rather than specific technologies, creating space for diverse approaches while ensuring consistent prosocial objectives. At the organizational level, companies need to move beyond compliance-based approaches to AI ethics. This means embedding prosocial considerations into product development processes, establishing clear accountability mechanisms, and investing in the technical infrastructure needed to build genuinely beneficial AI systems. Technical standards organizations should develop metrics and evaluation frameworks that measure prosocial outcomes, not just traditional performance metrics. We need ways to assess whether AI systems actually promote human flourishing, environmental sustainability, and social cohesion. The urgency cannot be overstated. As AI systems become more powerful and pervasive, the window for establishing prosocial governance frameworks is rapidly closing. Once entrenched systems and business models become established, changing them becomes exponentially more difficult and expensive. We're at a pivotal moment where the next generation of AI systems will be designed and deployed. The decisions we make now about how to govern these systems will shape society for decades to come. We can either allow narrow economic interests to drive AI development, or we can proactively steer this technology toward broadly beneficial outcomes. The challenge of prosocial AI governance isn't someone else's problem — it's a defining challenge of our time that requires leadership from every sector of society. Business leaders, policymakers, technologists, civil society organizations and ultimately each of us have roles in the AI-infused play that society has become. Prosocial AI governance isn't a constraint on innovation — it's the foundation for sustainable technological progress that benefits everyone. The time to act is now, before today's AI solutions become tomorrow's entrenched problems.
Yahoo
32 minutes ago
- Yahoo
Scale AI's 30-year-old billionaire founder still shops at Shein and pulls up to work in a Honda Civic: ‘Act broke, stay rich,' Lucy Guo says
Billionaire college dropout Lucy Guo is reportedly the youngest self-made woman on the planet—knocking Taylor Swift off the top spot. But even with a $1.3 billion reported net worth, the 30-year-old cofounder of Scale AI tells Fortune she still pinches the pennies and shops at Shein. Despite founding and retaining a $1.3 billion stake in an AI unicorn, you won't catch Lucy Guo wasting her billions on a lavish lifestyle to match her new status. 'I don't like wasting money,' the frugal 30-year-old tells Fortune. Of course, sometimes Guo will splurge—like if she's got a 16 hour flight to endure, she says she'll opt for business class. And there's the odd designer dress hanging in her closet for when she needs it. 'But in terms of like daily life, my assistant just drives me in a pretty old Honda Civic. I don't care,' she says. 'Everything I wear is free or from Shein… Some of them aren't going to be that great quality, but there's always like two pieces or so that really work out, and I just wear them every day,' the billionaire founder laughs. 'I still literally buy buy-one-get-one-free on Uber Eats.' Guo, who is currently the founder and CEO of the creator community platform Passes, adds that a quote she stumbled on on the morning of our interview perfectly summarises her approach: 'It's like, act broke, stay rich.' Guo hit the jackpot after the AI startup she cofounded, Scale AI, was reportedly valued at $25 billion in April as part of a share sale. Although she left the company in 2018 (two years after founding it), the 5% stake she held onto is now worth an estimated $1.2 billion—making the millennial one of just 5 female billionaires under 40 according to Forbes' latest ranking, including Rihanna and Anthropic's cofounder Daniela Amodei. It's why Guo no longer feels the need to prove her wealth with a Patek Philippe everyday watch, or a Hermès Birkin to carry her laptop. That, she says, is the behaviour of millionaires. 'Who you see typically wasting money on, designer clothes, a nice car, et cetera, they're technically in the millionaire range,' Guo explains. 'All their friends are multimillionaires, or billionaires and they feel a little bit insecure, so they feel the need to be flashy to show other people, 'look, I'm successful.'' 'I'm not showing off to anyone, right?' Indeed, for our interview, she's makeup-free, dressed down, and could pass for any other millennial. But earlier in her career, Guo admits she, too, may have been dripping in designer gear. 'I do think that this is actually something that I personally went through, and I think a lot of people go through when you're in that middle ground of you're successful, but not as successful as you want to be.' 'And I think the reason most billionaires dress in a t-shirt, jeans, hoodies, is that they can. They don't need to be in the suit 24/7 because they're done proving themselves to the rest of the world. The rest of the world is just sucking up to them,' she adds. 'And I think that's kind of how I like feel, where I'm past that hump. I don't really have to prove myself to anyone.' 'No one's going to look at me and point at me like, 'Haha, she's so broke' when I'm pulling up in a Honda Civic because whatever, it doesn't matter.' Guo's not the only ultra-wealthy to admit she's 'pretty frugal.' The world's most powerful have been boasting about their quiet luxury lifestyle for some time now. They've been donning logo-less angora wool jumpers and linen trousers that could be from anywhere to the unassuming eye. Experts say their wealthy peers can tell who is wearing Zara from who is in Loro Piana, but the point is to resemble people in lower tax brackets. Others, like KeKe Palmer and Warren Buffet, have been less subtle about how they lead very normal lives, despite their huge net worths—with the world's most famous investor going as far to call himself 'cheap'. But in Guo's eyes, she's one of the few who actually are as cheap as they say they are. 'I think that people want to fit into society. Specifically in America, I do think there is a 'we hate billionaires situation.' So because of that, people want to show, 'look, I'm not your typical billionaire. I'm frugal,'' she explains. 'I'm not saying it to be like, 'let me show you the world that I'm not like other billionaires,' Guo adds. 'I fully admit it, I have gone through that spending spree when I was more insecure, and I felt like I needed to show something.' And those who really aren't spending their millions? They aren't doing it to be relatable, she says it's because like her they had their flashy era—then reached the inevitable realisation: 'Why am I wasting my money on something that doesn't matter?' This story was originally featured on Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data