Opinion AI

Physical AI Is the Next Agent Layer

AI is leaving the screen. This is a simple piece on Physical AI, robotics, and why the next AI agent may have a body.

Opinion AI's avatar
Opinion AI
May 09, 2026
∙ Paid

Most of the AI you have used over the last two years lived inside a screen.

You opened a tab. You typed a prompt. You got an answer. You copied it somewhere else.

Then agents arrived. AI started clicking buttons, calling APIs, reading your files, running tools. You learned about MCP servers. You learned about workflows. You learned how to wire Claude into your real work.

That was a big jump.

But it was still a screen jump.

The next jump is bigger.

AI is now leaving the screen.

It is going into cameras, sensors, robotic arms, grippers, drones, cars, factory floors, warehouses, hospitals, farms, and one day, your kitchen. It is learning to see, decide, move, and improve from what actually happens in the real world.

This is what people are calling Physical AI.

And if you already understand how AI agents work, this article is written for you. Because the same loop you have seen in agents (read what is happening, plan a step, take an action, check the result, try again) is now happening with motors, wheels, cameras, and grippers, instead of just API calls and tool use.

That shift is not small.

Counterpoint Research expects around 145 million Physical AI units to ship between 2025 and 2035. NVIDIA has called this the ChatGPT moment for robotics. Amazon’s newest warehouse robot, Vulcan, already handles around 75 percent of the more than one million items in its catalog. At BMW, finished cars now drive themselves out of the assembly line, through testing, and into the finishing yard, with no human behind the wheel. UBS expects 2 million humanoid robots in workplaces by 2035, growing to 300 million by 2050.

This is not a 2030 story.

It is happening right now.

But before getting excited about humanoid robots dancing on stage, you have to understand the boring part first.

IBM defines Physical AI as AI systems that operate in and interact with the physical world, instead of staying only in software, by connecting AI models with sensors, actuators, and control systems. In simpler words, you take a model that can think, and you give it eyes, ears, hands, balance, and the ability to feel what is happening around it.

The old robot was just a script

Traditional factory robots were strong, fast, and precise. But they were dumb in a very specific way.

A welding robot welded the same seam, in the same spot, on the same car, all day.

A pick and place arm picked up the same screw, from the same tray, in the same orientation.

A robotic vacuum followed the same lines around your living room.

Move the screw two inches. Move the seam half a degree. Drop a sock in the vacuum’s path. The whole thing falls apart.

That is the world we are leaving.

The real world is not that polite.

Boxes shift. Lighting changes. People walk in the way. Objects slip. A floor gets wet. A box arrives slightly tilted. A bag bends in a strange way. A screw is a little off.

Old automation hates this. Physical AI is built for exactly this.

Deloitte describes the shift as machines moving from preprogrammed instructions to systems that perceive their environment, learn from experience, and adapt using real time data. Deloitte also makes a point almost every news story misses.

Physical AI is not only humanoid robots.

It is also smart factories, digital twins, AI camera systems, autonomous vehicles, delivery drones, and entire sensor based environments. The humanoid is just the loudest version of a much bigger story.

So if you have been ignoring this space because the dancing robot videos felt like hype, you have actually been missing the real shift.

Inside the full article, I break Physical AI down in the simplest way: how a normal AI agent becomes an agent with a body, why the real stack is sensors, world models, decision layers, motors, feedback, edge compute, and safety, why humanoid robot demos can be misleading, where the real business value may appear first, and how beginners can understand it by building a tiny camera → AI decision → action loop on their own laptop. This is not just a robotics explainer. It is a practical map of how AI moves from screen work to physical work, and why the next big agent may not live inside your browser at all.

User's avatar

Continue reading this post for free, courtesy of Opinion AI.

Or purchase a paid subscription.
© 2026 Mehboob · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture