EXECUTIVE SUMMARY

The AI conversation is about to change. For three years, we've focused on what AI can say and generate. The next decade will be defined by what AI can do—physically, in the real world, with hands, wheels, and sensors that interpret touch, force, and motion.

For enterprise leaders, the strategic question has shifted: not whether AI will enter the physical world, but whether your organization is positioned to benefit when it does.

Lunch and Learn: AI Workflow Engines with Claude, Gemini, & ChatGPT

Most professionals use AI assistants the same way they'd use a search engine—one question at a time, starting from scratch with every conversation. They're leaving 80% of the value on the table.

The real power of Claude Projects, ChatGPT CustomGPTs and Projects, and Google Gemini Gems isn't having a smarter assistant—it's building persistent, menu-driven systems that codify your expertise and automate your workflows.

In this session, you'll learn how to transform general-purpose AI into purpose-built tools that replicate complex tasks with consistency and precision.

Join us on January 27, 12:00 PM EST and leave with practical frameworks you can use immediately!

MORE FROM THE ARTIFICIALLY INTELLIGENT ENTERPRISE NETWORK

🎙️ AI Confidential Podcast - Are LLMs Dead?

 📚 AIOS - This is an evolving project. I started with a 14-day free AI email course to get smart on AI. But the next evolution will be a ChatGPT Super-user Course and a course on How to Build AI Agents.

AI DEEPDIVE

Physical AI Is the Next Frontier

Why the shift from language models to sensory models will determine which enterprises lead—and which get left behind

The evolution of AI follows a clear trajectory. First came language models—systems trained on text that could predict the next word. Then came multimodal models—systems that could process images, audio, and video alongside text. Now comes physical AI: systems that can perceive, reason, and act in the three-dimensional world.

Each stage requires capabilities that the previous stage lacked. A language model can write instructions for assembling furniture. A multimodal model can watch a video of someone assembling furniture and describe what's happening. A physical AI system can actually assemble the furniture—picking up pieces, applying appropriate force, adjusting when something doesn't fit.

That last capability—acting in physical space with real-time sensory feedback—represents the frontier that matters for enterprise.

What Is Physical AI

Physical AI refers to artificial intelligence systems designed to operate in and interact with the physical world. Unlike traditional AI that processes digital inputs and produces digital outputs, physical AI must handle the messy, continuous, unpredictable nature of reality.

The key differentiator is sensory integration. A physical AI system needs to process multiple streams of real-world data simultaneously: visual input from cameras, depth information from lidar, force feedback from tactile sensors, proprioceptive data about its own body position, and often much more. It then needs to translate high-level goals into low-level motor commands that account for physics, safety, and uncertainty.

This is fundamentally different from generating text or images. When ChatGPT produces an incorrect answer, you can ask it to try again. When a robot applies incorrect force while handling a component, it may damage equipment, injure someone, or destroy itself. The margin for error in physical systems is measured in millimeters and milliseconds.

NVIDIA's Jensen Huang captured the shift in his CES 2026 keynote: "The ChatGPT moment for robotics is here. Breakthroughs in physical AI—models that understand the real world, reason, and plan actions—are unlocking entirely new applications."

What makes this moment different from previous robotics hype cycles is the convergence of three capabilities: foundation models that can reason about physical scenarios, simulation environments that can generate training data at scale, and hardware platforms that can run these models in real-time on moving systems.

How Physical AI Works in Business Contexts

The clearest example of physical AI at scale is Waymo. The Alphabet subsidiary now provides over 450,000 paid robotaxi rides per week across five U.S. cities—Phoenix, San Francisco, Los Angeles, Austin, and Atlanta. By the end of 2026, Waymo aims to reach 1 million rides per week. In 2025 alone, the service completed 14 million trips.

These numbers matter because they represent real commercial deployment, not demonstrations. Waymo vehicles navigate construction zones, respond to ambulances, handle San Francisco's steep hills and dense pedestrian traffic—all without a human safety driver. The company's safety data shows its robotaxis are involved in five times fewer injury-causing accidents per million miles than human drivers.

What makes this possible isn't just cameras and lidar. It's the AI stack that processes sensor data into situational understanding. Waymo's vehicles don't just detect objects—they predict behavior, model uncertainty, and make thousands of decisions per second about speed, trajectory, and risk. They understand that a ball rolling into the street probably means a child is about to follow. They recognize a pedestrian's body language when they're about to jaywalk.

This kind of contextual physical reasoning is what separates true physical AI from earlier automation. And it's now being applied beyond vehicles.

Figure AI completed an 11-month deployment at BMW's Spartanburg plant in South Carolina, with its Figure 02 humanoid robots running 10-hour shifts loading sheet metal into welding fixtures. The robots contributed to the production of over 30,000 vehicles. More importantly, they demonstrated the minimum viable product for factory humanoids: performing a limited set of tasks reliably, over extended periods, with measurable results.

The technical requirements for physical AI are divided into three categories:

  • Perception: Processing camera feeds, depth sensors, force sensors, and proprioceptive data into a unified model of the environment and the system's position within it.

  • Reasoning: Understanding context, predicting outcomes, and planning sequences of actions that achieve goals while respecting physical constraints.

  • Action: Translating high-level plans into precise motor commands that account for dynamics, friction, balance, and real-time feedback.

NVIDIA's new model releases target each layer. Cosmos Reason 2 handles perception and reasoning—it's a vision-language model that enables machines to see, understand, and contextualize physical environments. Isaac GR00T N1.6 handles action—it's a vision-language-action model purpose-built for humanoid robots that translates reasoning into whole-body control. Together with simulation frameworks like Isaac Lab-Arena, they form a complete stack for training, testing, and deploying physical AI systems.

How to Implement Physical AI

For most enterprises, physical AI implementation will follow the pattern set by Waymo and factory robotics: start narrow, prove value, expand scope.

Phase 1: Identify Structured Physical Tasks

Physical AI today excels at tasks that are repetitive, rule-governed, and occur in controlled environments. Warehouse picking. Assembly line loading. Last-mile delivery in geofenced areas. Inspection routes that follow predictable paths.

The questions to ask: What physical tasks in your operation are high-volume, well-defined, and currently performed by humans? What's the cost of errors in those tasks? What data exists about how those tasks are currently performed?

Figure AI's success at BMW came from choosing sheet metal loading—a classic pick-and-place task with clear success criteria (cycle time under 84 seconds, placement accuracy within tolerance, minimal interventions). The task is boring, repetitive, and ergonomically challenging for humans. It's perfect for a first deployment.

Phase 2: Establish Measurement Infrastructure

Physical AI deployment requires granular measurement that most organizations lack. You need to know: How long does the task currently take? What's the error rate? What causes variations? What safety incidents occur?

Waymo's advantage comes partly from the 100+ million miles of autonomous driving data it has accumulated. Every edge case, every unusual scenario, every near-miss feeds back into training. Organizations deploying physical AI need similar feedback loops.

This often means instrumenting processes before automation. Cameras, timers, quality sensors, and incident tracking create the baseline against which AI performance will be measured—and the data that will improve it.

Phase 3: Pilot with Clear Success Metrics

Physical AI pilots should have quantified targets before deployment begins. Figure AI's BMW project tracked three metrics: cycle time (37-second target for sheet metal loading), placement accuracy, and the number of human interventions required. Everything else was secondary.

Conservative success criteria matter. Physical systems that fail can cause injury, damage equipment, or halt production. A pilot that achieves 80% of human performance while demonstrating safety and reliability is more valuable than one that occasionally matches human performance but introduces unpredictable failures.

Phase 4: Build for Fleet Learning

The economics of physical AI improve with scale because learning compounds. When one Waymo vehicle encounters an unusual situation—a parade, a flooded intersection, an aggressive driver—the learnings propagate to the entire fleet. When one factory robot masters a tricky assembly sequence, the skill transfers to every robot in the network.

This requires architecture decisions upfront. How will data flow from deployed systems back to the training infrastructure? How will model updates reach deployed robots? What testing validates that updates improve performance without introducing regressions?

Organizations that treat physical AI deployment as a one-time project will find themselves constantly re-implementing. Those who build for continuous learning will see compounding returns.

Common Missteps

  • Expecting immediate human replacement: Physical AI augments human workers before replacing them. Figure's robots at BMW worked alongside humans on the line, handling specific tasks while humans managed exceptions, quality verification, and coordination. Full autonomy in unstructured environments remains years away for most applications.

  • Underestimating edge cases: Physical environments generate edge cases at a rate that surprises organizations accustomed to software deployment. A new pallet configuration, a humidity change that affects sensor readings, a slightly different part from a new supplier—each creates potential failure modes. Budget for extensive testing and ongoing tuning.

  • Ignoring the data infrastructure: Physical AI requires data pipelines that most organizations haven't built. Sensor feeds, video storage, annotation workflows, training infrastructure, model deployment systems—each requires investment. Organizations that rush to deploy hardware without this foundation find themselves unable to improve performance or diagnose failures.

  • Treating robots as capital equipment: Traditional industrial robots are bought, installed, and operated for years with minimal changes. Physical AI systems are more like software products—they require updates, monitoring, and continuous improvement. Organizations that budget for purchase but not operation will underinvest in ongoing capability development.

Business Value

The economic case for physical AI rests on three factors: labor arbitrage, capability extension, and data generation.

  • Labor arbitrage is the obvious case. U.S. warehouses face persistent labor shortages, with turnover rates exceeding 100% annually at some facilities. Manufacturing struggles to fill positions that require repetitive physical work. Physical AI systems don't call in sick, don't quit, and can work continuous shifts. Waymo's robotaxis operate 24/7 in ways human drivers cannot.

  • Capability extension is less obvious but often more valuable. Physical AI can operate in environments too dangerous, too small, or too remote for human workers. Inspection robots can enter confined spaces. Autonomous vehicles can operate during severe weather when human drivers would stay home. Factory robots can work in clean rooms or high-temperature environments.

  • Data generation may ultimately prove most valuable. Every physical AI deployment creates detailed records of real-world operations. Waymo knows more about urban driving patterns than any transportation agency. Factory robots generate quality data that can identify supplier issues, environmental effects, and process improvements invisible to human observation.

  • The Morgan Stanley analysis of humanoid robot economics suggests unit costs will fall dramatically as production scales. Tesla's Optimus Gen 2 bill of materials runs approximately $46,000 with Chinese supply chains; estimates suggest this could fall below $20,000 at scale. Combined with 24/7 operation capability, the cost-per-task comparison with human labor becomes compelling for high-volume physical work.

  • Competitive implications: The physical AI capability gap between early movers and laggards will compound. Organizations that deploy now accumulate operational data, develop integration expertise, and build institutional knowledge about what works. Those that wait will find themselves competing against rivals whose systems have learned from millions of operational hours.

What This Means for Your Planning

The shift from language AI to physical AI will reshape enterprise technology strategy over the next five years. Three planning considerations deserve immediate attention.

First, the workforce strategy must account for augmentation before replacement. Physical AI systems will initially work alongside humans, handling specific tasks while humans manage exceptions, quality, and coordination. Organizations that frame robotics as immediate headcount reduction will underinvest in the human capabilities—supervision, maintenance, process optimization—that successful deployment requires.

Second, data infrastructure becomes a competitive asset. Every physical AI deployment generates operational intelligence. Organizations with substantial data collection, storage, and analysis capabilities will extract greater value from robotics investments and improve more quickly. Those without will find themselves dependent on vendor learning curves rather than their own operational data.

Third, the geographic dimension matters. China's 5-to-1 patent advantage in humanoid robotics and its cost leadership in robot components signal where manufacturing capability is concentrated. U.S. organizations face choices about supply chain resilience, technology partnerships, and build-versus-buy decisions that have geopolitical implications.

The question is no longer whether AI will operate in the physical world. Waymo's 450,000 weekly rides settle that. The question is whether your organization will be among those defining how physical AI integrates into your industry—or among those adapting to standards set by others.

Author’s note: This week’s complete edition—including the AI Toolbox and a hands-on Productivity Prompt—is now live on our website. Read it here.

ALL THINGS AI 2026

Join us at All Things AI 2026 in Durham, North Carolina, on March 23–24, 2026.

This is where AI gets real. No sales pitches—just 4,000 builders, engineers, operators, and execs sharing how AI actually works in practice, from hands-on workshops to real-world sessions, right in the heart of RTP and Duke’s AI ecosystem.

Early registration is open, and prices go up soon.

AI TOOLBOX

Tools for organizations exploring physical AI deployment and autonomous systems.

  • NVIDIA Isaac Sim — Robotics Simulation - High-fidelity simulation environment for training and testing physical AI systems. Integrates with NVIDIA's Cosmos and GR00T models for end-to-end robot development. Generates synthetic training data across diverse scenarios.

  • RoboDK — Robot Programming - Offline programming and simulation for industrial robots. Supports 75+ robot brands with unified interface. Enables testing robot programs before deployment to physical hardware.

  • Formant — Robot Fleet Management - Cloud platform for managing and monitoring robot fleets across locations. Real-time telemetry, remote intervention capabilities, and fleet analytics. Vendor-agnostic integration.

  • Viam — Robotics Development Platform - Modular platform for building smart machines and robots. Handles hardware abstraction, data management, and ML model deployment. Open-source with enterprise support options.

PRODUCTIVITY PROMPT

Prompt of the Week: Physical AI Readiness Assessment

Organizations considering physical AI deployment lack structured frameworks for evaluating readiness. Operations leaders see robotics announcements and wonder whether the technology applies to their context, but the analysis tends toward either dismissive skepticism or uncritical enthusiasm. Neither serves planning needs.

This prompt applies a structured evaluation framework to specific operational contexts. By requiring concrete details about current processes, it forces a realistic assessment rather than abstract speculation. The multi-dimensional scoring surfaces both opportunities and obstacles in a format that supports executive briefings.

The Prompt

You are a physical AI strategy consultant evaluating automation readiness for enterprise operations. Your task is to assess a specific operational area for physical AI potential.

## Context
Physical AI refers to robotic and autonomous systems that perceive, reason about, and act in physical environments. Current capabilities are strongest for structured, repetitive tasks in controlled environments. Technology is advancing rapidly but edge cases and exceptions still require human oversight.

## Input
[DESCRIBE YOUR OPERATIONAL AREA: Include task types, volume, current staffing, error rates, safety considerations, and physical environment]

## Assessment Dimensions
Evaluate each dimension 1-5 and explain your reasoning:

1. Task Structure (5 = highly repetitive and rule-governed; 1 = highly variable and judgment-dependent)
2. Environment Control (5 = controlled indoor space; 1 = unpredictable outdoor/public environment)
3. Error Tolerance (5 = errors easily corrected; 1 = errors cause safety incidents or major damage)
4. Data Availability (5 = extensive operational data exists; 1 = no systematic data collection)
5. Integration Complexity (5 = standalone process; 1 = deeply integrated with human workflows)

## Output Format
Provide:
- Dimension scores with justification
- Overall readiness score (sum of dimensions / 25 as percentage)
- Top 3 physical AI use cases for this operation (be specific)
- Critical prerequisites before pilot deployment
- Recommended timeline: Now / 12-24 months / 3-5 years / Not suitable
- Key risks and mitigation approaches

## Constraints
- Be specific to the described operation, not generic
- Acknowledge current technology limitations honestly
- Prioritize safety considerations
- Identify what data or infrastructure would improve readiness

Example Use Case

A distribution center operations manager evaluating automated picking would describe the following: current pick volume (50,000 units/day), product mix (5,000 SKUs), pick accuracy (99.2%), picker turnover (85% annually), and facility layout (traditional rack aisles). The prompt would return specific scoring, identify high-potential zones (uniform product areas, heavy-item sections), flag prerequisites (WMS integration, aisle width modifications), and recommend a phased approach starting with robotic goods-to-person systems before advancing to mobile manipulation.

I appreciate your support.

Your AI Sherpa,

Mark R. Hinkle
Publisher, The AIE Network
Connect with me on LinkedIn
Follow Me on Twitter

Reply

or to participate

Keep Reading

No posts found