If you’ve had the chance to ride in a Waymo, you’ve likely emerged from the vehicle amazed by its abilities. Since Alphabet launched the project, in 2009, Waymo has developed a fleet of 2,500 driverless robotaxis now on the road in San Francisco, Miami, Phoenix, and other cities, where they’ve completed more than 20 million trips. The vehicles do more than zip passengers around at up to 65 miles an hour. They can respond to verbal instructions (“Play some ’80s rap on Spotify”) or answer questions (“What time is the Giants game?”) while switching lanes to avoid a double-parked delivery van. Early customer service data and comments on the Waymo One app show that riders are thrilled by the experience.
Waymo is a specific use case of a technology that’s maturing rapidly and set for significant deployment: robots powered by generative AI. Many companies are already using gen AI chatbots, agents, and related technologies to automate and scale up customer service, but in most of these cases customers interact with the technology on screens. Embedding gen AI into robots gives companies the chance to reinvent their interactions with customers in physical settings—restaurants, hotels, hospitals, retail stores, and other brick-and-mortar locations—where service has remained stubbornly human. Using large language models (LLMs), large behavioral models (LBMs), and agentic AI, this new generation of robots can better understand context, make inferences, and provide personalized experiences. They can converse like competent employees—following logic across conversational turns, clarifying ambiguity, and explaining complex ideas simply. In the past a patient asking a robot, “Is this going to hurt? How long will it take? And what happens if I feel dizzy?” would overwhelm the limits of its script. But today’s LLM-powered service robots can unpack those concerns and offer plain-language responses. A robot named Robin is already doing just that, providing emotional support in 30 pediatric units and nursing homes across the country. It moves around autonomously to greet children and answer questions. Nurses can give Robin verbal commands, such as “Go to room 517 for 20 minutes, and then go to room 516 for 10 minutes.” It also comes loaded with games children can play using spoken responses.
This probably isn’t the first article you’ve read claiming that robots will transform the service sector. Admittedly, their impact has grown more slowly than enthusiasts expected. The global market for professional service robots, which includes models for logistics, healthcare, cleaning, and other sectors, grew roughly 9% in 2024, reaching almost 200,000 units sold. But many pilot projects stalled or underperformed. McKinsey research shows that 71% of companies say that high up-front costs are a major challenge when adopting robots, and 61% point to a lack of experience with automation as another top hurdle. Maintenance and reliability remain ongoing challenges, as does customer and employee acceptance; many people still prefer human interaction, especially in situations that are complex or emotionally charged. When companies have deployed customer-facing robots, most implementations have been confined to narrowly scripted tasks such as delivering room service or baggage in a hotel. These service robots are not much better than elaborate mobile vending machines. They dependably follow preprogrammed routes, read barcodes, and answer FAQs but largely have failed to deliver the scale or returns that early adopters had hoped for.
Nonetheless, virtually every major robot manufacturer is integrating gen AI into its offerings—and some of the early results show real promise. I’ve been researching advanced robotics for more than a decade, and over the past 18 months I’ve traveled to Europe, Asia, and North America to observe deployments of physical service robots powered by gen AI in 14 organizations operating in financial services, healthcare, education, hospitality, and more. In this article I’ll outline how firms can use this new technology to create value, mitigate risks, and build the organizational muscle for its success.
What Are Gen AI Robots?
Gen-AI-enabled robots rely on a mix of technologies—some that are familiar and some that are not.
By now most executives have a grasp of LLMs and agentic AI. In a robot, LLMs allow for conversation, and agentic AI adds memory, planning, execution, and reflection. Using those technologies, a robot can remember a returning customer, reason through trade-offs (Is early check-in possible if housekeeping says the room is almost ready?), plan a sequence of tasks, execute steps across digital systems and physical spaces, and then reflect on what worked and what didn’t. The difference between traditional, script-bound robots and those powered by agentic AI is night and day. A script-bound system might recognize that a delivery must be prioritized but still follow a predetermined route. An agentic system can assign the task to a human, reroute the delivery, and reallocate resources to enable it. While a robot has domain-specific intelligence (one in a hotel will be trained very differently from one in a hospital), it can make complex decisions to execute high-level instructions, such as “check in guests quickly” or “replenish the inventory of IVs by the end of the shift.”
LBMs are a less familiar technology. They are trained on large sets of behaviors, just as LLMs are trained on a seemingly infinite supply of text. LBMs help robots deal with the fact that service in a physical environment is messy. Trays tilt. Floors get slick. Customers hand over fragile items. As such, programming robots for every contingency is impossible. Instead, developers teach robots to learn using LBMs for any contexts that are required. LBMs are what allow Waymo vehicles to drive around double-parked vans.
Gen AI robots use cameras, microphones, and sensors to learn by observing humans, by asking questions, and through trial and error. This training can initially be done in the real world. Robots can acquire behaviors by observing a handful of demonstrations (of how to carefully pick up a glass of wine, for example) and then experimenting with millions of microvariations of speed, grip, and trajectory, using the metaverse or a digital twin to perfect their approach. The behaviors can also be transferred across contexts. If a robot learns to handle a fragile glass in a café, elements of that skill will carry over to handling vials in a clinic or delicate merchandise in a boutique.
Other technologies facilitate gen AI robots’ ability to learn. No-code programming and fleet learning (sharing instruction across a group of robots), for instance, make implementation and improvements easier than they were a decade ago. In the past any adjustment to robots required a ticket to IT or a visit from a vendor. No-code training (made possible by gen AI) lets frontline workers adjust a robot’s behavior by asking why it took a certain approach, describing a better one, and physically demonstrating it. This ease of improvement compresses the cycle time for operational enhancements from months to days. These capabilities turn robots into adaptive systems that integrate conversation, cognition, and physical action—at scale and personalized for each customer.
Consider the robots at work inside BMW’s automobile assembly plant in Spartanburg, South Carolina. The auto industry was an early adopter of robotics; factories have used simple robotic arms to do repetitive tasks (such as spot welding) since the 1980s. But in 2024 BMW began piloting Figure 02, a humanoid robot that represents a sharp break from traditional industrial automation. Unlike conventional factory robots, Figure 02 can move autonomously through the plant using six onboard cameras, interpret what it sees, and reason about how objects should be used, drawing on a large base of automotive and general knowledge. Powered by OpenAI models, it listens to and processes human speech, infers intent from even vague instructions, asks clarifying questions when needed, and learns from its mistakes over time. During an 11-month deployment, Figure 02 contributed to the production of roughly 30,000 BMW vehicles. It acted as a high-precision pair of hands in the body shop, carrying and placing fragile sheet-metal parts and lining components up so that the welding robots could build car frames. BMW is now upgrading to Figure 03, a lighter, taller successor designed to apply these capabilities beyond the factory floor. In a promotional video, Figure 03 performs tasks such as washing dishes, folding laundry, serving drinks, and playing fetch—highlighting that gen AI is giving robots the ability to take on an expanding range of functions.
How to Deploy Gen AI Robots
My work with companies shows that the most exciting uses of gen AI robots involve frontline and customer- and employee-facing tasks. Bringing a robot into a workplace is more complicated than just unboxing a gadget, because many workplaces are unpredictable—waiters carry trays, doctors and nurses hustle from room to room, and so on. To convert potential into performance, leaders must carefully select use cases, communicate to customers and employees why and how they’re using robots, and set up guardrails. Four critical steps, based on my observations in the field, can guide them.
1. Start with use cases that address labor constraints.
Robots are most effective when applied to repeatable, economically valuable tasks that provide measurable returns. Many of the early experiments I observed involved roles in industries that face chronic labor shortages. That makes sense: The robots are reducing not only costs but also the difficulty of recruiting hard-to-find workers. Once you’ve identified target roles, start by examining the specific tasks those jobs require. Are they repeatable enough for a robot to learn quickly? Is the payoff from turning them over to a robot immediate in terms of speed, efficiency, consistency, or freeing up employees to do work where they add more value?
Good candidates for experiments include check-in and checkout in hotels, order modification and delivery in quick-service restaurants, and logistics in hospital wards—settings where there’s constant pressure on frontline capacity and a clear operational handoff the robot can own.
Once you’ve chosen the use case, design the pilot so that frontline employees can actively improve performance in the flow of work. An employee who notices a robot carrying out a suboptimal sequence should be able to explain and demonstrate a better method—without waiting weeks for a software update.
Done well, this approach expands what robots can do over time, enhances their performance in the domain, and makes their implementation more effective—while shifting frontline roles toward more-skilled work and easing labor shortages where they’re most acute.
2. Design robot interactions for customer acceptance.
Most resistance to robots begins at the customer touchpoint, not with the technology itself. Traditional kiosks, self-service technologies, and scripted bots often force customers into rigid, unnatural sequences. This is a design problem, and LLM-enabled robots undo that burden. Customers and employees can speak naturally to them and—critically—the robots can follow through in the physical environment.
Consider hotel check-in. A human employee typically navigates multiple systems—reservation records, loyalty profiles, housekeeping status, payment, and preferences—one screen at a time. A gen-AI-powered robot can connect to those systems in parallel, reconcile conflicts, and deliver a room key in seconds while interacting with and welcoming the guest. To the customer, the experience is a warm, efficient conversation—and often proves more consistent than human service under pressure.
Early deployments highlight both the promise and the pitfalls. At Henn na Hotel in Tokyo, robot receptionists guide guests through identity verification, room assignment, and payment. The initiative has helped with labor costs and shortages, but not everything has worked as planned. At times the robots have struggled with accents, background noise, and unexpected requests, sometimes increasing rather than reducing staff workload.
Designing for customer acceptance means testing these interactions in real environments with real customers, learning where friction arises, and ensuring the robot’s conversational competence is matched by reliable physical execution.
3. Position robots as service enhancers, not workforce replacements.
How robots are introduced—and how their role is explained—strongly shapes how people perceive them. Acceptance of AI varies by demographics and context, and today many customers still prefer human interaction for encounters that require warmth, empathy, or judgment. Employees, meanwhile, are worried about losing their jobs to the technology.
To help address those preferences and fears, companies should position gen AI robots not as replacements for frontline employees but as tools that improve accessibility, speed, and reliability while freeing humans to focus on emotionally charged, ambiguous, or high-stakes interactions. In my observations the most successful deployments happen where convenience dominates customers’ priorities and the value of automation is obvious.
