logo
What Is ‘Physical AI'? Inside The Push To Make AI Understand The Real World

What Is ‘Physical AI'? Inside The Push To Make AI Understand The Real World

Forbes13-06-2025
What happens when AI enters the physical world — predicting actions, spotting risks and transforming ... More how machines understand real-time events?
For years, AI has been great at seeing things. It can recognize faces, label objects and summarize the contents of a blurry image better than most humans. But ask it to explain why a person is pacing nervously near a fence, or predict what might happen next in a crowded room — and suddenly, the illusion of intelligence falls apart.
Add to this reality the fact that AI largely remains a black box and engineers still struggle to explain why models behave erratically or how to correct them, and you might realize the big dilemma in the industry today.
But that's where a growing wave of researchers and startups believe the next leap lies: not just in faster model training or flashier generative outputs, but in machines that truly understand the physical world — the way it moves, reacts and unfolds in real time. They're calling it 'physical AI'.
The term was initially popularized by Nvidia CEO Jensen Huang, who previously has called physical AI the next AI wave, describing it as 'AI that understands the laws of physics,' moving beyond pixel labeling to bodily awareness — space, motion and interaction.
At its core, physical AI merges computer vision, physics simulation and machine learning to teach machines cause and effect. Essentially, it enables AI systems to not just recognize objects or people, but to understand how they interact with their surroundings — like how a person's movement might cause a door to swing open or how a ball might bounce off a wall.
At Lumana, a startup backed by global venture capital and growth equity firm Norwest, that phrase isn't just branding; it's a full-blown product shift. Known for AI video analytics, the company is now training its models not only to detect motion, but to recognize human behavior, interpret intent and automatically generate real-time alerts.
'We define physical AI as the next evolution of video intelligence,' Lumana CEO Sagi Ben-Moshe said in an interview. 'It's no longer just about identifying a red car or a person in a hallway — it's about inferring what might happen next, and taking meaningful action in real-world conditions.'
In one real-world deployment, Lumana's system flagged a possible assault after detecting unusual body language and close proximity between two men and a pair of unattended drinks, prompting an alert that allowed staff to step in before anything escalated. In another case, it caught food safety violations in real time, including workers skipping handwashing, handling food without gloves and leaving raw ingredients out too long. These weren't issues discovered after the fact, but ones that the system caught as they unfolded. This kind of layered inference, Ben-Moshe explained, transforms cameras into 'intelligent sensors.'
It's no coincidence that Huang has also previously used the term 'physical AI,' linking it to embodied intelligence and real-world simulation. It reflects a broader shift in the industry about creating AI systems that better understand the laws of physics and can reason more intelligently. Physics, in this context, is shorthand for cause and effect — the ability to reason about motion, force and interaction, not just appearances.
That framing resonated with investors at Norwest, who incubated Lumana during its earliest phase. 'You can't build the future of video intelligence by just detecting objects,' said Dror Nahumi, a general partner at Norwest. 'You need systems that understand what's happening, in context and can do it better than a human watching a dozen screens. In many cases, businesses also need this information in real-time.'
Norwest isn't alone. Other players, from Hakimo to Vintra, are exploring similar territory — using AI to spot safety violations in manufacturing, detect loitering in retail, or prevent public disturbances before they escalate.
For example, Hakimo recently built an autonomous surveillance agent that prevented assaults, identified vandalism and even saved a collapsed individual using live video feeds and AI. At Nvidia GTC in March, Nvidia even demoed robotic agents learning to reason about gravity and spatial relationships directly from environment-based training, echoing the same physical reasoning that Lumana is building into its surveillance stack.
And just yesterday, Meta announced the release of V- JEPA 2, 'a self-supervised foundation world model to understand physical reality, anticipate outcomes and plan efficient strategies.' As Michel Meyer, group product manager at the Core Learning and Reasoning arm of the company's Fundamental AI Research, noted on LinkedIn yesterday quoting Meta chief AI scientist Yann Lecun, 'this represents a fundamental shift toward AI systems that can reason, plan, and act through physical world models. To reach advanced machine intelligence, AI must go beyond perception and understand how the physical world works — anticipating dynamics, causality, and consequences. V‑JEPA 2 does just that.'
When asked what the real-world impact of physical AI might look like, Nahumi noted that it's more than mere marketing. 'Anyone can detect motion, but if you want real AI in video surveillance, you must go beyond that to understand context.' He sees Lumana's full-stack, context-driven architecture as a foundation and not a vanity pitch.
'We think there's a big business here and the technology is now reliable enough to augment and outperform humans in real time,' told me.
The reality is that the success of physical‑AI systems will not be just about the technology. As AI continues to advance, it's becoming much clearer that the success of most AI systems largely hinges on ethics, trust and accountability. Put in a different way, trust is the currency of AI success. And the big question that companies must continue to answer is: Can we trust your AI system to be safe?
In a security context, false positives can shut down sites or wrongly accuse innocent people. In industrial settings, misinterpreted behavior could trigger unnecessary alarms.
Privacy is another concern. While many physical AI systems operate on private premises — factories, campuses, hotels — critics warn that real-time behavior prediction, if left unchecked, could drift into mass surveillance. As Ben-Moshe himself acknowledged, this is powerful technology that must be used with guardrails, transparency and explicit consent.
But, according to Nahumi, Lumana's multi-tiered model delivers actionable alerts, but also protects privacy and supports seamless integration into existing systems. 'Lumana engineers systems that layer physical AI on current infrastructure with minimal friction,' he noted, 'ensuring operators aren't overwhelmed by false positives.'
Despite these questions, demand is accelerating. Retailers want to track foot traffic anomalies. Municipalities want to prevent crime without expanding staff. Manufacturers want safety compliance in real time, not post-event reviews. In every case, the challenge is the same: too many cameras, too little insight.
And that's the business case behind physical AI. As Norwest's Nahumi put it, 'We're seeing clear ROI signals — not just in avoided losses, but in operational efficiency. This is no longer speculative deep tech. It's a platform bet.'
That bet hinges on systems that are scalable, adaptable and cost-effective. Lumana's approach, which layers physical AI on top of existing camera infrastructure, avoids the 'rip-and-replace' problem and keeps adoption friction low. Nahumi pointed to rising enterprise demand across retail, manufacturing, hospitality and public safety — fields where video footage is ubiquitous, but analysis remains manual and inefficient.
And even across boardrooms and labs, the appetite for machines that 'understand' rather than 'observe' is growing. That's why companies like Norwest, Nvidia, Hakimo and Lumana are doubling down on physical AI.
'In five years,' Ben-Moshe envisions, 'physical AI will do more than perceive — it will suggest actions, predict events and give safety teams unmatched visibility.' This, he noted, is about systems that not only see, but also act.
Ultimately, the goal of physical AI isn't just to help machines see better — it's to help them understand what they're seeing. It's to help them perceive, understand and reason in the messy physical world we inhabit.
Ben-Moshe envisions a future where physical AI suggests actions, prevents escalation and even predicts incidents before they unfold. 'Every second of video should generate insight,' he said. 'We want machines to reason about the world as a system — like particles tracing possible paths in physics — and highlight the most likely, most helpful outcome.'
That's a far cry from today's basic surveillance. From thwarting crime and preventing accidents to uncovering new operational insights and analyzing activity trends, reasoning engines over cameras promise real, demonstrable value.
But scaling them is where the real work is. It'll require systems that are accurate, ethical, auditable and trustworthy. If that balance is struck, we could enter a world where AI won't just help us see what happened, but help us know what matters most.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Copyrighted Books Are Fair Use For AI Training. Here's What To Know.
Copyrighted Books Are Fair Use For AI Training. Here's What To Know.

Forbes

time34 minutes ago

  • Forbes

Copyrighted Books Are Fair Use For AI Training. Here's What To Know.

The use of AI systems has become part of our daily lives. The sudden presence of generative AI systems in our daily lives has prompted many to question the legality of how AI systems are created and used. One question relevant to my practice: Does the ingestion of copyrighted works such books, articles, photographs, and art to train an AI system render the system's creators liable for copyright infringement, or is that ingestion defensible as a 'fair use'? A court ruling answers this novel question, and the answer is: Yes, the use of copyrighted works for AI training is a fair use – at least under the specific facts of those cases and the evidence presented by the parties. But because the judges in both cases were somewhat expansive in their dicta about how their decisions might have been different, they provide a helpful roadmap as to how other lawsuits might be decided, and how a future AI system might be designed so as not to infringe copyright. The rulings on Meta and Anthropic's respective cases require some attention. Let's take a closer look. More than 30 lawsuits have been filed in the past year or two, in all parts of the nation, by authors, news publishers, artists, photographers, musicians, record companies and other creators against various AI systems, asserting that using the authors' respective copyrighted works for AI training purposes violates their copyrights. The systems' owners invariably assert fair use as a defense. They provide a helpful roadmap as to how other lawsuits might be decided, and how a future AI system might be designed so as not to infringe copyright. The Anthropic Case Anthropic planned to create a central library of "all the books in the world." The first decision, issued in June, involved a lawsuit by three book authors, who alleged that Anthropic PBC infringed the authors' copyrights by copying several of their books (among millions of others) to train its text generative AI system called Claude. Anthropic's defense was fair use. Judge Alsup, sitting the Northern District Court of California, held that the use of the books for training purposes was a fair use, and that the conversion of any print books that Anthropic had purchased and converted to digital was also a fair use. However, Anthropic's use of pirated digital copies for purposes of creating a central library of 'all the books in the world' for uses beyond training Claude, was not a fair use. Whether Anthropic's copying of its central library copies for purposes other than AI training (and apparently there was some evidence that this was going on, but on a poorly developed record) was left for another day. It appears that Anthropic decided early on in its designing of Claude that books were the most valuable training materials for a system that was designed to 'think' and write like a human. Books provide patterns of speech, prose and proper grammar, among other things. Anthropic chose to download millions of free digital copies of books from pirate sites. It also purchased millions of print copies of books from booksellers, converted them to digital copies and threw the print copies away, resulting in a massive central library of 'all the books in the world' that Anthropic planned to keep 'forever.' None of this activity was done with the authors' permission. Significantly, Claude was designed so that it would not reproduce any of the plaintiffs' books as output. There was not any such assertion by the plaintiffs, nor any evidence that it did so. The assertions of copyright infringement were, therefore, limited to Claude's ingestion of the books for training, to build the central library, and for the unidentified non-training purposes. Users of Claude ask it questions and it returns text-based answers. Many users use it for free. Certain corporate and other users of Claude pay to use it, generating over one billion dollars annually in revenue for Anthropic. The Anthropic Ruling Both decisions were from the federal district court in Northern California, the situs of Silicon ... More Valley. To summarize the legal analysis, Judge Alsup evaluated each 'use' of the books separately, as it must under the Supreme Court's 2023 Warhol v. Goldsmith fair use decision. Turning first to the use of the books as training data, Alsup found that the use of the books to train Claude was a 'quintessentially' transformative use which did not supplant the market for the plaintiffs' books, and as such qualified as fair use. He further found that the conversion of the purchased print books to digital files, where the print copies were thrown away, was also a transformative use akin to the Supreme Court's 1984 Betamax decision in which the court held that the home recording of free TV programming for time-shifting purposes was a fair use. Here, Judge Alsup reasoned, Anthropic lawfully purchased the books and was merely format-shifting for space and search capability purposes, and, since the original print copy was discarded, only one copy remained (unlike the now-defunct Redigi platform of 2018). By contrast, the downloading of the over seven million of pirate copies from pirate sites, which at the outset was illegal, for central library uses other than for training purposes could not be held to be a fair use as a matter of law, because the central library use was unjustified and the use of the pirate copies could supplant the market for the original. Anthropic Is Liable For Unfair Uses – The Cost of Doing Business? The case will continue on the issue of damages for the pirated copies of the plaintiffs' books used for central library purposes and not for training purposes. The court noted that the fact that Anthropic later purchased copies of plaintiffs' books to replace the pirated copies will not absolve it of liability, but might affect the amount of statutory damages it has to pay. The statutory damages range is $750 per copy at a minimum and up to $150,000 per copy maximum. It tempts one to wonder about all those other millions of copyright owners beyond the three plaintiffs – might Anthropic have to pay statutory damages for seven million copies if the pending class action is certified? Given the lucrativeness of Claude, could that be just a cost of doing AI business? The Meta Case Meta's decision to use shadow libraries to source books was approved by CEO Mark Zuckerberg. The second decision, issued two days following the Anthropic decision, on June 25, involves thirteen book authors, most of them famous non-fiction writers, who sued Meta, the creator of a generative AI model called Llama, for using the plaintiffs' books as training data. Llama (like Claude), is free to download, but generates billions of dollars for Meta. Like Anthropic, Meta initially looked into licensing rights from book publishers, but eventually abandoned those efforts and instead downloaded the books it desired from pirate sites called 'shadow libraries' which were not authorized by the copyright owners to store their works. Also like Claude, Llama was designed not to produce output that reproduced its source material in whole or substantial part, the record indicating that Llama could not be prompted to reproduce more than 50 words from the plaintiffs' books. Judge Chhabria, also in the Northern District of California, held Meta's use of plaintiffs' works to train Llama was a fair use, but he did so very reluctantly, chiding the plaintiff's lawyers for making the 'wrong' arguments and failing to develop an adequate record. Chhabria's decision is riddled with his perceptions of the dangers of AI systems potentially flooding the market with substitutes for human authorship and destroying incentives to create. The Meta Ruling Based on the parties' arguments and the record before him, like Judge Alsup, Judge Chhabria found that Meta's use of the books as training data for Llama was 'highly transformative' noting that the purpose of the use of the books - for creating an AI system - was very different than the plaintiffs' purpose of the books, which was for education and entertainment. Rejecting plaintiff's argument that Llama could be used to imitate the style of plaintiffs' writing, Judge Chhabria noted that 'style is not copyrightable.' The fact that Meta sourced the books from shadow libraries rather than authorized copies didn't make a difference; Judge Chhabria (in my opinion rightly) reasoned that to say that a fair use depends on whether the source copy was authorized begs the question of whether the secondary copying was lawful. Although plaintiffs tried to make the 'central library for other purposes than training' argument that was successful in the Anthropic case, Judge Chhabria concluded that the evidence simply didn't support that copies were used for purposes other than training, and noted that even if some copies were not used for training, 'fair use doesn't require that the secondary user make the lowest number of copies possible.' Since Llama couldn't generate exact or substantially similar versions of plaintiffs' books, he found there was no substitution harm, noting that plaintiffs' lost licensing revenue for AI training is not a cognizable harm. Judge Chhabria's Market Dilution Prediction Judge Chhabria warns that generative AI systems could dilute the market for lower-value mass market ... More publications. In dicta, clearly expressing frustration with the outcome in Meta's favor, Judge Chhabria discussed in detail how he thought market harm could – and should - be shown in other cases, through the concept of 'market dilution' - warning that a system like Llama, while not producing direct substitutes for a plaintiff's work, could compete with and thus dilute the plaintiff's market. There may be types of works unlike award-winning fictional works more susceptible to this harm, he said, such as news articles, or 'typical human-created romance or spy novels.' But since the plaintiffs before him didn't make those arguments, nor presented any record of the same, he said, he could not make a ruling on the same. This opportunity is left for another day. AI System Roadmap For Non-Infringement The court decisions provide an early roadmap as to how to design an AI system. Based on these two court decisions, here are my take-aways for building a roadmap for a non-infringing generative AI system using books:

Is Go Hub Capital Berhad's (KLSE:GOHUB) Recent Performance Underpinned By Weak Financials?
Is Go Hub Capital Berhad's (KLSE:GOHUB) Recent Performance Underpinned By Weak Financials?

Yahoo

time35 minutes ago

  • Yahoo

Is Go Hub Capital Berhad's (KLSE:GOHUB) Recent Performance Underpinned By Weak Financials?

With its stock down 6.9% over the past three months, it is easy to disregard Go Hub Capital Berhad (KLSE:GOHUB). We decided to study the company's financials to determine if the downtrend will continue as the long-term performance of a company usually dictates market outcomes. Specifically, we decided to study Go Hub Capital Berhad's ROE in this article. Return on Equity or ROE is a test of how effectively a company is growing its value and managing investors' money. In short, ROE shows the profit each dollar generates with respect to its shareholder investments. This technology could replace computers: discover the 20 stocks are working to make quantum computing a reality. The formula for ROE is: Return on Equity = Net Profit (from continuing operations) ÷ Shareholders' Equity So, based on the above formula, the ROE for Go Hub Capital Berhad is: 6.1% = RM3.7m ÷ RM60m (Based on the trailing twelve months to March 2025). The 'return' is the amount earned after tax over the last twelve months. Another way to think of that is that for every MYR1 worth of equity, the company was able to earn MYR0.06 in profit. View our latest analysis for Go Hub Capital Berhad Thus far, we have learned that ROE measures how efficiently a company is generating its profits. Based on how much of its profits the company chooses to reinvest or "retain", we are then able to evaluate a company's future ability to generate profits. Generally speaking, other things being equal, firms with a high return on equity and profit retention, have a higher growth rate than firms that don't share these attributes. When you first look at it, Go Hub Capital Berhad's ROE doesn't look that attractive. A quick further study shows that the company's ROE doesn't compare favorably to the industry average of 12% either. Go Hub Capital Berhad was still able to see a decent net income growth of 15% over the past five years. So, there might be other aspects that are positively influencing the company's earnings growth. For example, it is possible that the company's management has made some good strategic decisions, or that the company has a low payout ratio. We then compared Go Hub Capital Berhad's net income growth with the industry and found that the company's growth figure is lower than the average industry growth rate of 30% in the same 5-year period, which is a bit concerning. The basis for attaching value to a company is, to a great extent, tied to its earnings growth. The investor should try to establish if the expected growth or decline in earnings, whichever the case may be, is priced in. This then helps them determine if the stock is placed for a bright or bleak future. If you're wondering about Go Hub Capital Berhad's's valuation, check out this gauge of its price-to-earnings ratio, as compared to its industry. While the company did pay out a portion of its dividend in the past, it currently doesn't pay a regular dividend. We infer that the company has been reinvesting all of its profits to grow its business. In total, we would have a hard think before deciding on any investment action concerning Go Hub Capital Berhad. While the company has posted decent earnings growth, the company is retaining little to no profits and is reinvesting those profits at a low rate of return. This makes us doubtful if that growth could continue, especially if by any chance the business is faced with any sort of risk. Up till now, we've only made a short study of the company's growth data. You can do your own research on Go Hub Capital Berhad and see how it has performed in the past by looking at this FREE detailed graph of past earnings, revenue and cash flows. — Investing narratives with Fair Values Suncorp's Next Chapter: Insurance-Only and Ready to Grow By Robbo – Community Contributor Fair Value Estimated: A$22.83 · 0.1% Overvalued Thyssenkrupp Nucera Will Achieve Double-Digit Profits by 2030 Boosted by Hydrogen Growth By Chris1 – Community Contributor Fair Value Estimated: €14.40 · 0.3% Overvalued Tesla's Nvidia Moment – The AI & Robotics Inflection Point By BlackGoat – Community Contributor Fair Value Estimated: $384.84 · 0.2% Overvalued View more featured narratives — Have feedback on this article? Concerned about the content? Get in touch with us directly. Alternatively, email editorial-team (at) article by Simply Wall St is general in nature. We provide commentary based on historical data and analyst forecasts only using an unbiased methodology and our articles are not intended to be financial advice. It does not constitute a recommendation to buy or sell any stock, and does not take account of your objectives, or your financial situation. We aim to bring you long-term focused analysis driven by fundamental data. Note that our analysis may not factor in the latest price-sensitive company announcements or qualitative material. Simply Wall St has no position in any stocks mentioned. Sign in to access your portfolio

2025 NBA free agency: Deandre Ayton addition slightly improves Los Angeles Lakers' title odds
2025 NBA free agency: Deandre Ayton addition slightly improves Los Angeles Lakers' title odds

Yahoo

time35 minutes ago

  • Yahoo

2025 NBA free agency: Deandre Ayton addition slightly improves Los Angeles Lakers' title odds

The Los Angeles Lakers finally have a new center after agreeing to a two-year deal with free agent and former No. 1 overall pick Deandre Ayton, and oddsmakers have bumped up their odds to win next year's NBA championship as a result. The Lakers moved from 16-1 to 14-1 to win the 2025-26 NBA championship at BetMGM, the seventh-best odds of any team and the fourth-best odds for any Western Conference team behind favorite Oklahoma City (+275), Denver (+700) and Houston (+800). Advertisement "They filled an obvious need with a previously unexpected option," Jeff Sherman, vice president of risk at the Westgate Las Vegas SuperBook told Yahoo Sports via text. "We went from 16-1 to 14-1 for perception purposes. We already have — and will continue to have — Lakers liability." The Lakers made the playoffs as the third seed in the West last season after trading for Luka Dončić in the middle of the season, but lost to the Minnesota Timberwolves in the first round in five games. As of Tuesday morning, the Lakers were the second-biggest title liability at BetMGM behind the Dallas Mavericks.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store