logo
Legion Intelligence to Partner with Palantir to Deliver Advanced Generative AI to U.S. Special Operations Command

Legion Intelligence to Partner with Palantir to Deliver Advanced Generative AI to U.S. Special Operations Command

Business Wire13-05-2025

SAN FRANCISCO--(BUSINESS WIRE)--Legion Intelligence (formerly Yurts AI), a leading provider of scalable and secure Generative AI (GenAI) solutions for the Department of Defense, today unveiled they are working with Palantir Technologies to enhance the U.S. Special Operations Command's (SOCOM) Mission Command System (MCS) with next-generation AI capabilities.
This collaboration builds on Legion's leadership in deploying GenAI across SOCOM's sensitive and classified networks, leveraging cloud, hybrid, and on-premises environments. The integration of Legion's AI platform with Palantir's Apollo Mission Manager will enable efficient and secure deployment, accelerating the adoption of large language models (LLMs) to support situational awareness, data-driven insights, and superior decision-making in mission-critical environments.
In addition to enhancing SOCOM's operational workflows, Legion's platform enables the rapid incorporation of Generative AI capabilities into and across applications that may not easily leverage AI, whether in on-premises infrastructure or sensitive networks. This integration extends the core functionalities and data residing in the Mission Command System into new experiences and use cases, allowing SOCOM to unlock greater potential from its operational technology.
'We're honored to deepen our relationship with SOCOM by bringing Generative AI directly into mission workflows,' said Ben Van Roo, CEO of Legion Intelligence. 'GenAI delivers its greatest value when integrated into decision-making processes and networks critical to national defense. This partnership is a vital step toward empowering warfighters with immediate, secure access to actionable intelligence.'
Legion's scalable AI solutions and Palantir's robust data integration and security capabilities deliver a reliable and mission-focused platform for SOCOM's global operations.
About Legion Intelligence
Legion (formerly Yurts AI) is an agent-orchestration platform for defense, government, and enterprise organizations. Legion empowers teams to securely deploy, manage, and scale intelligent autonomous agents, enabling sophisticated, efficient coordination across sensitive operational environments. By integrating advanced MLOps, robust security protocols, dynamic decision-making, and human-centric design, Legion ensures every critical mission is backed by a capable legion of intelligent agents.
With Legion, the future of secure, sophisticated AI orchestration is here. Existing and prospective customers can confidently navigate their most sensitive environments, leveraging Legion's proven capabilities to enhance operational effectiveness, security, and strategic decision-making. Visit https://www.legionintel.com/ to learn more.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Anthropic CEO: GOP AI regulation proposal ‘too blunt'
Anthropic CEO: GOP AI regulation proposal ‘too blunt'

The Hill

time31 minutes ago

  • The Hill

Anthropic CEO: GOP AI regulation proposal ‘too blunt'

Anthropic CEO Dario Amodei criticized the latest Republican proposal to regulate artificial intelligence (AI) as 'far too blunt an instrument' to mitigate the risks of the rapidly evolving technology. In an op-ed published by The New York Times on Thursday, Amodei said the provision barring states from regulating AI for 10 years — which the Senate is now considering under President Trump's massive policy and spending package — would 'tie the hands of state legislators' without laying out a cohesive strategy on the national level. 'The motivations behind the moratorium are understandable,' the top executive of the artificial intelligence startup wrote. 'It aims to prevent a patchwork of inconsistent state laws, which many fear could be burdensome or could compromise America's ability to compete with China.' 'But a 10-year moratorium is far too blunt an instrument,' he continued. 'A.I. is advancing too head-spinningly fast. I believe that these systems could change the world, fundamentally, within two years; in 10 years, all bets are off.' Amodei added, 'Without a clear plan for a federal response, a moratorium would give us the worst of both worlds — no ability for states to act, and no national policy as a backstop.' The tech executive outlined some of the risks that his company, as well as others, have discovered during experimental stress tests of AI systems. He described a scenario in which a person tells a bot that it will soon be replaced with a newer model. The bot, which previously was granted access to the person's emails, threatens to expose details of his marital affair by forwarding his emails to his wife — if the user does not reverse plans to shut it down. 'This scenario isn't fiction,' Amodei wrote. 'Anthropic's latest A.I. model demonstrated just a few weeks ago that it was capable of this kind of behavior.' The AI mogul added that transparency is the best way to mitigate risks without overregulating and stifling progress. He said his company publishes results of studies voluntarily but called on the federal government to make these steps mandatory. 'At the federal level, instead of a moratorium, the White House and Congress should work together on a transparency standard for A.I. companies, so that emerging risks are made clear to the American people,' Amodei wrote. He also noted the standard should require AI developers to adopt policies for testing models and publicly disclose them, as well as require that they outline steps they plan to take to mitigate risk. The companies, the executive continued, would 'have to be upfront' about steps taken after test results to make sure models were safe. 'Having this national transparency standard would help not only the public but also Congress understand how the technology is developing, so that lawmakers can decide whether further government action is needed,' he added. Amodei also suggested state laws should follow a similar model that is 'narrowly focused on transparency and not overly prescriptive or burdensome.' Those laws could then be superseded if a national transparency standard is adopted, Amodei said. He noted the issue is not a partisan one, praising steps Trump has taken to support domestic development of AI systems. 'This is not about partisan politics. Politicians on both sides of the aisle have long raised concerns about A.I. and about the risks of abdicating our responsibility to steward it well,' the executive wrote. 'I support what the Trump administration has done to clamp down on the export of A.I. chips to China and to make it easier to build A.I. infrastructure here in the United States.' 'This is about responding in a wise and balanced way to extraordinary times,' he continued. 'Faced with a revolutionary technology of uncertain benefits and risks, our government should be able to ensure we make rapid progress, beat China and build A.I. that is safe and trustworthy. Transparency will serve these shared aspirations, not hinder them.'

Tech giants' indirect emissions rose 150% in three years as AI expands, UN agency says
Tech giants' indirect emissions rose 150% in three years as AI expands, UN agency says

Yahoo

time38 minutes ago

  • Yahoo

Tech giants' indirect emissions rose 150% in three years as AI expands, UN agency says

By Olivia Le Poidevin GENEVA (Reuters) -Indirect carbon emissions from the operations of four of the leading AI-focused tech companies, Amazon, Microsoft, Alphabet and Meta, rose on average by 150% from 2020-2023, as they had to use more power for energy-demanding data centres, a United Nations report said on Thursday. The use of artificial intelligence is driving up global indirect emissions because of the vast amounts of energy required to power data centres, the report by the International Telecommunication Union (ITU), the U.N. agency for digital technologies, said. Indirect emissions include those generated by purchased electricity, steam, heating and cooling consumed by a company. Amazon's operational carbon emissions grew the most at 182% in 2023 compared to three years before, followed by Microsoft at 155%, Meta at 145% and Alphabet at 138%, according to the report. The ITU tracked the greenhouse gas emissions of 200 leading digital companies between 2020 and 2023. Meta, which owns Facebook and WhatsApp, pointed Reuters to its sustainability report that said it is working to reduce emissions, energy and water used to power its data centres. The other companies did not respond immediately to requests for comment. As investment in AI increases, carbon emissions from the top-emitting AI systems are predicted to reach up to 102.6 million tons of carbon dioxide equivalent (tCO2) per year, the report stated. The data centres that are needed for AI development could also put pressure on existing energy infrastructure. "The rapid growth of artificial intelligence is driving a sharp rise in global electricity demand, with electricity use by data centres increasing four times faster than the overall rise in electricity consumption," the report found. It also highlighted that although a growing number of digital companies had set emissions targets, those ambitions had not yet fully translated into actual reductions of emissions.

Trump Administration To Rebrand Biden-Era Artificial Intelligence Safety Institute, Commerce Secretary Says At AI Honors: 'We're Not Going To Regulate It'
Trump Administration To Rebrand Biden-Era Artificial Intelligence Safety Institute, Commerce Secretary Says At AI Honors: 'We're Not Going To Regulate It'

Yahoo

timean hour ago

  • Yahoo

Trump Administration To Rebrand Biden-Era Artificial Intelligence Safety Institute, Commerce Secretary Says At AI Honors: 'We're Not Going To Regulate It'

Commerce Secretary Howard Lutnick told a D.C. crowd this week that the Biden-era AI Safety Institute would be rebranded as the Center for AI Standards and Innovation, as a 'place where people voluntarily go to drive analysis and standards.' 'As we move from large language models to large quantitative models, and we add all these different things, you want a place to go,' Lutnick said. 'We say, has someone checked out this model? Is this a safe model? Is this a model that we understand? How do I do this? And we're not going to regulate it. We are going to enhance the voluntary models of what great American innovation is all about.' More from Deadline Trump Launches Punitive Biden Probe, New Travel Bans, But Still Silent On Elon Musk's 'Kill Bill' Attack On Agenda Trump Celebrity Supporters: Famous Folks In Favor Of The 47th President Former Biden White House Press Secretary Karine Jean-Pierre To Publish Book Explaining Her Affiliation Switch From Democrat To An Independent Lutnick's remarks came at the inaugural AI Honors this week, held by the Washington AI Network at the Waldorf Astoria. The rebrand reflects a more hands-off approach that the Trump administration has taken to AI, after President Joe Biden often addressed AI by spotlighting the need for guardrails around the technology, and lined up major AI companies to agree to a set of voluntary commitments for 'responsible innovation.' Biden signed an executive order in 2023 that directed the Department of Commerce to develop standards for authentication and watermarking, among other things, in the creation of a Safety Institute. Days after taking office, Trump rescinded Biden's executive order, placing an emphasis on deregulation. In his speech, Lutnick said that AI safety 'is sort of an opinion-based model. And the Commerce Department and NIST, the National Institute of Standards and Technology, we do standards and we do most successfully cyber, the gold standard of cyber.' The Biden administration acknowledged that in many cases it would be left to Congress to pass laws to regulate AI technology. In entertainment, one of the more significant proposals is the No Fakes Act, which would give individuals a right to control their digital likeness, meaning that content creators would need permission to recreate celebrities and anyone else using AI. The Center for AI Standards and Innovation also will seek voluntary agreements 'with private sector AI developers and evaluators, and lead unclassified evaluations of AI capabilities that may pose risks to national security,' per a Commerce Department announcement. The ceremony on Tuesday honored Sen. Todd Young (R-IN); Rep. Jay Olbernolte (R-CA) and Rep. Ted Lieu (D-CA); Vice Admiral Frank Whitworth, director of the National Geospatial-Intelligence Agency; SandboxAQ CEO Jack Hidary; Patricia Falcone, deputy director for science and technology at the Lawrence Livermore National Laboratory; Ylli Bajraktari, president & CEO of SCSP and founder of AI + EXPO; and Booz Allen's Chief Technologist Joanna Guy, Space Llama Engineer Zane Price; and VP of AI Don Polaski. Alos honored was Father Paolo Benanti, Vatican adviser on AI ethics. The Washington AI Network was founded by Tammy Haddad. CNN anchor Sara Sidner emceed the event. In his speech, Lutnick also emphasized the need for the U.S. to remain the leader in AI, as he outlined parts of the administration's strategy to boost advanced manufacturing. 'The fact is that our adversaries are substantially behind us and we expect to keep them substantially behind us, but we want to bring our allies onto our side,' he said. Among other things, he talked of doubling the U.S. power capacity to meet the need for giant data centers. 'The power necessary to drive these data centers is awesome. It's awesome the amount of power they draw,' he said. 'And it can't be that the United States of America is balancing its citizens operating their refrigerator or a data center. That is just not a practical solution. So the practical solution would be to allow data center operators to build their own power generation sites adjacent to their data center.' Best of Deadline 'Stick' Soundtrack: All The Songs You'll Hear In The Apple TV+ Golf Series 'Nine Perfect Strangers' Season 2 Release Schedule: When Do New Episodes Come Out? 'Stick' Release Guide: When Do New Episodes Come Out?

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store