
What Is ‘Physical AI'? Inside The Push To Make AI Understand The Real World
What happens when AI enters the physical world — predicting actions, spotting risks and transforming ... More how machines understand real-time events?
For years, AI has been great at seeing things. It can recognize faces, label objects and summarize the contents of a blurry image better than most humans. But ask it to explain why a person is pacing nervously near a fence, or predict what might happen next in a crowded room — and suddenly, the illusion of intelligence falls apart.
Add to this reality the fact that AI largely remains a black box and engineers still struggle to explain why models behave erratically or how to correct them, and you might realize the big dilemma in the industry today.
But that's where a growing wave of researchers and startups believe the next leap lies: not just in faster model training or flashier generative outputs, but in machines that truly understand the physical world — the way it moves, reacts and unfolds in real time. They're calling it 'physical AI'.
The term was initially popularized by Nvidia CEO Jensen Huang, who previously has called physical AI the next AI wave, describing it as 'AI that understands the laws of physics,' moving beyond pixel labeling to bodily awareness — space, motion and interaction.
At its core, physical AI merges computer vision, physics simulation and machine learning to teach machines cause and effect. Essentially, it enables AI systems to not just recognize objects or people, but to understand how they interact with their surroundings — like how a person's movement might cause a door to swing open or how a ball might bounce off a wall.
At Lumana, a startup backed by global venture capital and growth equity firm Norwest, that phrase isn't just branding; it's a full-blown product shift. Known for AI video analytics, the company is now training its models not only to detect motion, but to recognize human behavior, interpret intent and automatically generate real-time alerts.
'We define physical AI as the next evolution of video intelligence,' Lumana CEO Sagi Ben-Moshe said in an interview. 'It's no longer just about identifying a red car or a person in a hallway — it's about inferring what might happen next, and taking meaningful action in real-world conditions.'
In one real-world deployment, Lumana's system flagged a possible assault after detecting unusual body language and close proximity between two men and a pair of unattended drinks, prompting an alert that allowed staff to step in before anything escalated. In another case, it caught food safety violations in real time, including workers skipping handwashing, handling food without gloves and leaving raw ingredients out too long. These weren't issues discovered after the fact, but ones that the system caught as they unfolded. This kind of layered inference, Ben-Moshe explained, transforms cameras into 'intelligent sensors.'
It's no coincidence that Huang has also previously used the term 'physical AI,' linking it to embodied intelligence and real-world simulation. It reflects a broader shift in the industry about creating AI systems that better understand the laws of physics and can reason more intelligently. Physics, in this context, is shorthand for cause and effect — the ability to reason about motion, force and interaction, not just appearances.
That framing resonated with investors at Norwest, who incubated Lumana during its earliest phase. 'You can't build the future of video intelligence by just detecting objects,' said Dror Nahumi, a general partner at Norwest. 'You need systems that understand what's happening, in context and can do it better than a human watching a dozen screens. In many cases, businesses also need this information in real-time.'
Norwest isn't alone. Other players, from Hakimo to Vintra, are exploring similar territory — using AI to spot safety violations in manufacturing, detect loitering in retail, or prevent public disturbances before they escalate.
For example, Hakimo recently built an autonomous surveillance agent that prevented assaults, identified vandalism and even saved a collapsed individual using live video feeds and AI. At Nvidia GTC in March, Nvidia even demoed robotic agents learning to reason about gravity and spatial relationships directly from environment-based training, echoing the same physical reasoning that Lumana is building into its surveillance stack.
And just yesterday, Meta announced the release of V- JEPA 2, 'a self-supervised foundation world model to understand physical reality, anticipate outcomes and plan efficient strategies.' As Michel Meyer, group product manager at the Core Learning and Reasoning arm of the company's Fundamental AI Research, noted on LinkedIn yesterday quoting Meta chief AI scientist Yann Lecun, 'this represents a fundamental shift toward AI systems that can reason, plan, and act through physical world models. To reach advanced machine intelligence, AI must go beyond perception and understand how the physical world works — anticipating dynamics, causality, and consequences. V‑JEPA 2 does just that.'
When asked what the real-world impact of physical AI might look like, Nahumi noted that it's more than mere marketing. 'Anyone can detect motion, but if you want real AI in video surveillance, you must go beyond that to understand context.' He sees Lumana's full-stack, context-driven architecture as a foundation and not a vanity pitch.
'We think there's a big business here and the technology is now reliable enough to augment and outperform humans in real time,' told me.
The reality is that the success of physical‑AI systems will not be just about the technology. As AI continues to advance, it's becoming much clearer that the success of most AI systems largely hinges on ethics, trust and accountability. Put in a different way, trust is the currency of AI success. And the big question that companies must continue to answer is: Can we trust your AI system to be safe?
In a security context, false positives can shut down sites or wrongly accuse innocent people. In industrial settings, misinterpreted behavior could trigger unnecessary alarms.
Privacy is another concern. While many physical AI systems operate on private premises — factories, campuses, hotels — critics warn that real-time behavior prediction, if left unchecked, could drift into mass surveillance. As Ben-Moshe himself acknowledged, this is powerful technology that must be used with guardrails, transparency and explicit consent.
But, according to Nahumi, Lumana's multi-tiered model delivers actionable alerts, but also protects privacy and supports seamless integration into existing systems. 'Lumana engineers systems that layer physical AI on current infrastructure with minimal friction,' he noted, 'ensuring operators aren't overwhelmed by false positives.'
Despite these questions, demand is accelerating. Retailers want to track foot traffic anomalies. Municipalities want to prevent crime without expanding staff. Manufacturers want safety compliance in real time, not post-event reviews. In every case, the challenge is the same: too many cameras, too little insight.
And that's the business case behind physical AI. As Norwest's Nahumi put it, 'We're seeing clear ROI signals — not just in avoided losses, but in operational efficiency. This is no longer speculative deep tech. It's a platform bet.'
That bet hinges on systems that are scalable, adaptable and cost-effective. Lumana's approach, which layers physical AI on top of existing camera infrastructure, avoids the 'rip-and-replace' problem and keeps adoption friction low. Nahumi pointed to rising enterprise demand across retail, manufacturing, hospitality and public safety — fields where video footage is ubiquitous, but analysis remains manual and inefficient.
And even across boardrooms and labs, the appetite for machines that 'understand' rather than 'observe' is growing. That's why companies like Norwest, Nvidia, Hakimo and Lumana are doubling down on physical AI.
'In five years,' Ben-Moshe envisions, 'physical AI will do more than perceive — it will suggest actions, predict events and give safety teams unmatched visibility.' This, he noted, is about systems that not only see, but also act.
Ultimately, the goal of physical AI isn't just to help machines see better — it's to help them understand what they're seeing. It's to help them perceive, understand and reason in the messy physical world we inhabit.
Ben-Moshe envisions a future where physical AI suggests actions, prevents escalation and even predicts incidents before they unfold. 'Every second of video should generate insight,' he said. 'We want machines to reason about the world as a system — like particles tracing possible paths in physics — and highlight the most likely, most helpful outcome.'
That's a far cry from today's basic surveillance. From thwarting crime and preventing accidents to uncovering new operational insights and analyzing activity trends, reasoning engines over cameras promise real, demonstrable value.
But scaling them is where the real work is. It'll require systems that are accurate, ethical, auditable and trustworthy. If that balance is struck, we could enter a world where AI won't just help us see what happened, but help us know what matters most.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


CNET
30 minutes ago
- CNET
Score $250 Off Google's Flagship Pixel 9 Pro XL While You Still Can
When you're shopping at the high end of the Android phone market, the Google Pixel 9 Pro XL is one option you should definitely consider. It's fast, has a huge display, and gets all the latest Android updates as soon as possible. And right now, Best Buy is offering the flagship phone with a huge $250 discount. This deal means that you can now get the 128GB model for just $849, rather than the usual $1,099. Need more space? The 256GB version is just $949, down from the usual $1,199. Both deals are part of a larger Android savings event that Best Buy is running, so check the full collection out if you don't want to go the Pixel route. The Google Pixel 9 Pro XL comes with a huge 6.8-inch display that looks stunning whether you're inside or outdoors, even in bright sunlight. It packs a 42-megapixel front-facing camera for great selfies, while the main rear camera uses a 50-megapixel sensor. That's backed up by 48-megapixel telephoto and ultrawide cameras. Wireless connectivity includes support for the latest Wi-Fi 7 technology, while both wired and wireless charging are supported. As we've come to expect from modern flagship phones, the Pixel 9 Pro XL is IP68 dust and water resistant, so you need not worry about trips to the beach or pool. Hey, did you know? CNET Deals texts are free, easy and save you money. Fans of AI will enjoy a ton of Google features, not least the Gemini assistant. AI-powered features can also be found throughout Android, including phone call screening and advanced image-editing options. This phone is, of course, unlocked so can be used on all major US carriers, including Verizon, AT&T and T-Mobile. Alternatively, you can use it with prepaid carriers including Cricket Wireless, MetroPCS, Google Fi, Total Wireless, Tracfone and Mint. Why this deal matters Getting a Pixel phone means you'll get all of the latest Android improvements and features first. The Pixel 9 Pro XL is also at the top of its game with fast performance, a stunning display and all of the AI features that you could possibly want. The chance to pick one up for just $849 shouldn't be ignored, especially if you're in the market for a new, premium phone already.


Bloomberg
30 minutes ago
- Bloomberg
Rapidan's McNally Sees $3-5 Upside Risk to Oil
Bob McNally, Rapidan Energy Group Founder & President, says he easily sees $3 to $5 upside risk in oil prices, with the tensions between Iran and Israel likely to continue and potentially expand. He added that the US regime is concerned about the inflationary pressure of elevated oil prices, especially if they hit around $120 a barrel, something he described as potentially as "devastating price spike." He spoke with Dani Burger on Bloomberg Brief. (Source: Bloomberg)

Business Insider
31 minutes ago
- Business Insider
I took a chaotic, surreal robotaxi ride through central London. It left me impressed, but with one big question.
Self-driving taxis are now a common sight on the streets of San Francisco and Shanghai — but in London, the city's famous black cabs still rule the roost. That may not be the case for much longer. Earlier in June, Uber announced it would begin trialling driverless taxis in the UK's capital, in partnership with Microsoft and Nvidia-backed self-driving startup Wayve. As the two companies prepare to launch their robotaxi service, I took a test drive through the busy streets of central London in one of the self-driving Ford Mach-Es Wayve has been testing in the city. The robotaxi dealt with everything London's chaotic driving environment could throw at it, with Wayve CEO Alex Kendall, who joined for the ride, saying it was one of the more eventful trips he'd taken. Robot, take the wheel After a quick look around Wayve's Kings Cross base, we hopped in a Ford Mach-E equipped with the company's "AI driver" self-driving system. Wayve's upcoming robotaxi service with Uber will be fully driverless, but for now, the company is testing with safety drivers who take over if the robotaxi runs into problems. I sat in the front seat, alongside a big red button that disengages the self-driving system (the safety driver on my ride said he'd never had to use it). The driver drove us out of the gates of Wayve's HQ and then, with a sharp buzzing noise, he removed his hands from the wheel, and the robotaxi took over. London's streets are famously complex, so much so that taxi drivers are required to learn thousands of them off by heart in a 150-year-old exam known as "The Knowledge." The city's hodgepodge of Roman and Victorian roads are a mess of cycle lanes and pedestrian crossings, with complex road layouts that often serve more as a rough guide than a rulebook for the millions of drivers passing through the city each day. For Wayve, that complexity is the point. The company says its AI driver — which runs on an end-to-end AI model, an approach also adopted by Tesla — is capable of generalizing and reacting to the physical world in the same way a human would, unlike rivals like Waymo, which rely on high-definition maps and sensors. Kendall said that this allows Wayve's software to drive anywhere, even places it hasn't seen before, and deal with the kind of unexpected encounters that are an everyday occurrence on the streets of a major city like London. "I can't wait to see another autonomy company come into London because I think it's extremely challenging," said Kendall. "The advantage of starting in London is that we've been forced to develop a system that can operate on complex roads and deal with all of these unexpected scenarios," he added. On the road In the first few minutes of our drive, we encountered multiple jaywalkers, including several who darted out across the street without warning in front of the robotaxi. We also had to inch through narrow gaps between rows of parked cars. Perhaps understandably, Wayve's robotaxi was more cautious than your average Uber driver. On one occasion, the robotaxi came to a jerky stop at an amber light when it may have had enough time to go through. Sometimes, that caution was too much for London's impatient commuters. One maneuver that saw the car slowly navigate a tight gap between a set of roadworks and a row of traffic drew an irate honk from the car behind, which was clearly keen for us to hurry up. Wayve has been testing its autonomous vehicles on the streets of London for years, but they still run into the odd surprise. Around halfway into our drive, we came across a roadblock as a team of workers trimmed an overhanging tree. In the seat behind me, Kendall broke off from answering my question and leaned forward to see how the car would deal with the falling branches and the worker directing traffic — but the robotaxi barely paused, successfully reading hand signals to stop and proceed and accelerating past the quizzical stares of the maintenance crew. "I've never seen that before," said Kendall as the robotaxi exited the roadblock, also dodging a tractor reversing along the road. Our route skirted the edge of Soho, London's central nightlife district, and went past the entrance to the British Museum, a narrow street packed with tourists, taxi drivers, and roadworks. Perhaps the most impressive moment of the drive came on this road, one of London's busiest. The robotaxi came to a stop at a crosswalk in near constant use as a huge crowd of pedestrians streamed toward the Museum. Working out the safe moment to move forward would be daunting even for a human driver, but the car spotted an opening and moved through the zebra crossing quickly, avoiding traffic building up behind us. "Lots of autonomy systems, I think, would be stuck here for hours," said Kendall, explaining that moments like this required the robotaxi's AI software to track and predict the activity of dozens of pedestrians. A glimpse into the future? As the drive went on, the activity around us occasionally passed into the bizarre. While coming to a stop at a red light, we narrowly avoided being hit by a small piece of metal that fell off the truck in front as it accelerated away. A pedestrian crossing at the red light kindly picked up the debris and moved it to the side of the road, but it's hard to imagine anyone thinking to simulate a situation so surreal. For Kendall, incidents like this are proof that robotaxis need to get on as many roads as possible. After opening offices in Canada, the US, and Japan over the last year, Wayve has now tested its vehicles in 90 cities in the past 90 days, from alpine roads in Switzerland and dirt tracks in rural Italy to bustling cities like Tokyo and Vancouver. Kendall said for some of those places, Wayve had little to no training data, demonstrating the ability of its software to apply what it has learnt to completely new cities and driving environments. "If robotaxis are just going to be in affluent areas like the Bay Area, then we've failed as an industry. We've got to make sure this technology is rolled out globally," Kendall said. The launch of a driverless ride-hailing service with Uber in London, which will be the first time Wayve has taken the safety driver out of its vehicles, is a big step toward that vision. Both companies declined to say when the driverless pilots would begin or which automaker would supply the vehicles for the robotaxi fleet, but Kendall described it as a huge moment for the young British startup and the broader industry. "It's exciting, it's my life's work. Over 99% of road accidents are due to human error, and this technology can eliminate those," he said. A long road ahead My ride in Wayve's robotaxi ended with the car smoothly negotiating a double roundabout. The safety driver took back control to pull into the company's headquarters, the first time he'd touched the wheel since we'd left an hour ago. After 60 minutes of driving around the centre of one of the world's biggest and most chaotic cities without a hitch, I was impressed, if not completely sold. While self-driving cars are now a reality, the history of the industry is littered with broken promises, high-profile failures, and impressive tech demos that have failed to pan out. Saber Fallah, professor of safe AI and autonomy at the University of Surrey, told BI that while Wayve's AI-focused approach to self-driving was "bold and innovative," he was skeptical about how soon the company would be able to build a robotaxi network in a city as complex as London. "Unlike cities like Phoenix, where many Level 4 trials have taken place, London presents an extremely dynamic and unpredictable environment," said Fallah, adding that the vehicles would need to understand informal social cues and context between drivers and reason about their intent. Fallah said that one issue with Wayve's approach is that the "black box" nature of end-to-end AI systems might make it difficult to understand why robotaxis based on this tech make decisions, potentially undermining public trust. Despite calling the planned robotaxi pilot an "exciting step" for self-driving cars in the UK, Fallah warned that Uber and Wayve would have to navigate challenging liability and regulatory issues while building public confidence in the frontier technology. "Achieving consistent, safe operation without fallback drivers in central London will be an enormous hurdle," he added. Whether Uber and Wayve can run rides like the one I experienced in London a hundred times a day, in rain, sun, fog, and hail, will be the ultimate test. Until then, Londoners will have to stick to the city's black cabs and iconic red buses to get around.