Your PMS Has AI Now. Here's What It Still Doesn't Do

Your PMS Has AI Now. Here's What It Still Doesn't Do

If you manage 20 STR units or more, you've probably tried at least one AI chatbot vendor by now. There's a decent chance it didn't end well — long onboarding, integrations that never quite worked, the bot replying after a human took over the conversation, a contract you couldn't get out of. If that's your experience, you're not alone, and the conclusion isn't "AI doesn't work in hospitality." It's that the category has been selling a version of AI that doesn't match how operations actually run.

That picture has changed in the last year. PMS-native AI is now real. Guesty ships ReplyAI Autopilot, Copilot, AI Task Creation, ReviewSmart, and an AI Revenue Agent. Hospitable, Hostaway, and Cloudbeds are tracking the same direction. If you're paying for a major PMS in 2026, inbox automation is increasingly included. Auto-tasks from messages happen. Listing copy writes itself. None of this is a differentiator anymore.

Good news for operators. It also means the question has moved. It's no longer can the AI reply to a guest. It's can the AI operate across the real stack you already run — phone, multiple PMSs, multiple channels, your SOP, your team's escalation rules.

For operators managing 20–100+ STR units, the gap between those two questions shows up gradually. Past 50 units, it gets hard to ignore. Across the 25 hospitality properties we run AI agents for, where Una has handled over 100,000 guest conversations across phone, widget, WhatsApp, and OTA inbox, five gaps consistently show up. None of them are about model quality. They're about what your PMS — and most third-party chatbot vendors — don't try to do.

Gap 1: Your AI runs, but it doesn't tell you what to fix.

This is the gap costing operators the most time, and the one nobody puts on a marketing page.

Open Copilot. Ask "how am I doing this month." You'll get a table. The table is correct. It does not tell you what to do next. Open the Autopilot dashboard. You'll see how many messages got auto-answered. It does not tell you which units consistently produce escalations, or which task types your KB hasn't covered, or which OTA channel is producing the worst guest sentiment in week four.

The information exists. It's somewhere in your PMS data. The agent doesn't surface it because answering messages and prescribing operational fixes are different products. Your PMS AI is doing the first; nobody is doing the second by default.

This is where ops time disappears at scale. Your AI is closing 70% of inbound. The other 30% escalates to humans, and you don't have a clean view of why. So you can't fix the underlying cause. So next month, your escalation rate is the same. So you hire another VA.

The fix isn't a better dashboard. It's an agent that reads your operational data the way it reads guest messages, and tells you in plain language what's structurally broken this week. "Three units in Barcelona generated 40% of complaints. Check-in instructions look outdated." That's the output operators actually use.

Gap 2: Voice for STR is still largely unsolved.

PMS-native AI doesn't do voice. Read Guesty's AI page carefully — Autopilot, Copilot, ReviewSmart, the Revenue Agent are all text. No voice agent, no telephony integration, no phone-handling layer. Voice is a different stack from text. It needs telephony, speech-to-text, text-to-speech, and a runtime that responds in roughly half a second to feel natural. None of that lives in a PMS.

The third-party vendors with native phone voice today are mostly built for hotels — branded enterprise hotels with reservation desks and call centers. Very few integrate deeply with Hospitable or with the operational shape of an STR portfolio.

For an STR operator, the practical issue is simple: the phone calls that still matter sit outside the tools you already use. As text automation matures across your inbox channels, FAQ migrates out of the phone. What's left is the late-night arrival who can't find the lockbox, the guest who locked themselves out, the noise complaint from a neighbour, the booking extension that needs to be priced. These are the moments where service quality is decided. They're also the moments where you're paying overnight staff or letting calls go to voicemail.

By the time you're at a hundred units, you probably have one phone number per cluster, plus office lines, plus whatever's printed on the lockbox and the Airbnb "contact host" button. None of that traffic flows through PMS-native AI, and few STR-targeting AI vendors handle it either. It's the largest channel still without a real answer for STR operators.

Gap 3: Your PMS's AI sees only the units in that PMS.

If you grew organically and stayed on one stack, this gap doesn't exist for you. Skip ahead.

Past 50 units, most operators are not on one stack. They have sixty units in Hospitable from a property they acquired, thirty in Guesty because that's where the original portfolio lives, ten in Hostaway because a partner brought them in. Normal outcome of growing through acquisition or running mixed inventory types.

Guesty's AI works inside Guesty. Its Copilot answers questions about Guesty data. Its Autopilot replies to messages routed through Guesty's inbox. It doesn't see, know about, or act on units in your other systems.

This isn't a feature gap — it's how PMS vendors are built. They don't write AI that reads competitors' data, because doing so would invert their business model.

If your portfolio is split across Guesty, Hospitable, Mews, or Apaleo, logo coverage on a vendor's website is not enough. What you need to know is whether the agent can pull state from all of them, route the guest to the right action, and escalate cleanly when judgment is needed — when a guest in your Mews unit asks about a stay extension, when WhatsApp on a Hospitable unit needs an answer, when a Guesty booking gets cancelled and downstream tasks need to be cleaned up. A single agent that runs that workflow across all your PMSs has to be built outside the PMS layer to exist at all.

Gap 4: One agent across channels, or five tools that almost talk to each other.

Guesty Autopilot covers OTA threads and WhatsApp routed through Guesty's inbox. That's a lot. But guests reach you through more channels than that. They call. They write to office email addresses that don't sync to the PMS. They use the chat widget on your direct booking site. They send WhatsApp voice notes, not text. They reply to automated check-in emails from a different system.

You can stitch this with five different tools. Each will work. They will not share context, brand voice, or escalation rules. A guest who started in your widget, switched to phone, and ended up in WhatsApp will get three different agents handling three fragments of one conversation. None of them will know what the others said.

The alternative is one agent exposed through several channels. Phone, widget, WhatsApp voice, OTA inbox — same agent, same property knowledge, same escalation logic. The transport changes; the agent doesn't. Hard to retrofit, impossible to do from inside a PMS, because the PMS only owns one of those channels.

Gap 5: What breaks at scale — and what other operators are saying about it.

This is the section we most wish someone had written for our customers three years ago.

When AI gets deployed across a small portfolio, almost everything works. The KB is fresh because it was just written. The data is clean because it was just imported. The agent has ten properties to remember and gets them right. Then the operator scales. The failure patterns that show up are remarkably consistent. They're also written, in plain language, across public review platforms — Trustpilot, HotelTechReport, Capterra. Anyone evaluating a guest-communication AI vendor in 2026 should read those reviews before signing.

Four failure patterns repeat across public operator reviews and the sales conversations we have every week:

The bot keeps messaging after a human takes over. A guest asks something complex. Your team takes over the conversation. The AI is supposed to step back. It doesn't — it keeps generating replies, sometimes contradicting your team in front of the guest. We hear this one weekly in sales conversations, and it shows up in public reviews. It's a product commitment, not a technical limitation.

The chatbot makes false promises about policy or processes bookings that don't exist. Hallucinated cancellation rules. Confirmed availability that isn't there. In the worst public cases, "accepted and processed bookings" the property cannot honor. This pattern shows up in operator reviews, sometimes from verified hoteliers who paid for the product.

Integration that doesn't actually integrate. Tasks that don't make it to the PMS. Email channels that work in the demo and silently break in production. WhatsApp activation that takes months while billing continues. The pattern is almost identical across vendors: the integration was sold as plug-and-play; the reality required quarters of support tickets.

Contracts you can't get out of. "We were forced to stay in for an entire year." "I am trapped in a contract with them." A pattern operators flag repeatedly, especially against vendors who don't publish their pricing. Many vendors here require "talk to sales" pricing — and some of those have public reviews from customers paying while service wasn't delivered.

None of these are model quality failures. They're product, integration, and commercial-design failures — exactly the layer where PMS-native AI also doesn't help, and exactly where most third-party vendors got into trouble in the first wave of deployments.

What to put on the dashboard.

Stop tracking Average Handle Time. It was designed to minimize human labor cost — useful when every minute is a paid agent's minute. For AI, it's the wrong signal.

Track instead: escalation rate by PMS (which integration is degrading), escalation rate by issue type (which task types are failing), resolution rate by channel (phone vs widget vs OTA inbox), and the precision of escalations — when a human gets a handoff, was it actually one that needed a human?

These tell you whether the system is working. AHT tells you whether your calls are short, which is not the same thing.

What to ask before you sign anything.

If you've been burned once, the question isn't whether to use AI. It's how to evaluate the next vendor without repeating the mistake. Five questions worth asking before you sign:

  • Does the AI actually stop replying when my team takes over a conversation? Show me, in a demo, how that handoff works.
  • When a guest needs something done — extension, late check-in, lock code, payment fix — does the agent complete the workflow, or does it just hand off? Walk me through one workflow end-to-end. How many of the steps does the agent do, and at what point does a human get involved?
  • Can you handle phone calls natively, or does "voice" mean voice notes from WhatsApp?
  • What's your published pricing? If it's "talk to sales," what's the minimum contract length, and what happens if the integration doesn't work in month two?
  • Show me what the agent does when a guest cancels their booking. Then show me how it handles "I can't find my check-in instructions" at 3 AM — there are at least five different reasons this happens, and the agent should figure out which one before replying.

If a vendor can't answer these clearly, the risk usually sits with you, not with them. The category has been overpromising for two years, and operator reviews are the public record of what happens when claims don't match implementation.

For a deeper evaluation framework — including how to test state-dependent overnight problems and the kind of escalations that decide whether the product actually works — see The 3 AM Test.

When you don't need any of this.

If you're on one PMS, your phone volume is low, and your guests communicate almost exclusively through OTA inbox threads — Guesty Autopilot or its equivalent on Hostaway or Hospitable will cover most of what you need. Don't add another tool. The integration cost and the operational complexity aren't worth it. Spend your time on KB hygiene and check-in instructions instead.

The five gaps above start to matter when you're multi-PMS, when phone is a real channel for you, when you've been burned by a vendor whose dashboard didn't catch a failure mode, or when your portfolio is past the point where one person can hold every property's quirks in their head.

The conference noise will keep cycling — chatbots don't work, agents are the future, interfaces will win, voice is the next thing. Most of it is true and most of it is incomplete. The question for an operator running real units is narrower: what does your PMS now do for free, what doesn't it do, and which of those gaps are costing you enough to be worth solving.

About Polydom. Polydom runs voice-first AI agents for hospitality operators. Una works across phone, WhatsApp, widget, and inbox, pulls state from PMS, payment, and access systems, and escalates with full context when human judgment is needed.

We integrate with Mews, Apaleo, Guesty, Hospitable, Zeevou, Clock PMS, and Profitroom at workflow depth — not just messaging passthrough. Native phone voice ships in production.

Over the past year and a half in production, Una has handled 100,000+ guest conversations and nearly a million messages.

We don't sell self-serve. Every deployment starts with a guided pilot on one real operational use case — typically after-hours phone, stay extensions, or multi-channel inbox unification — configured around your systems and SOP, so you can see whether it fits your operation before any wider rollout. Single-property deployments take a day. Multi-property, one to three days. The gating factor is usually how fast your team hands over access and property data, not the integration itself.

Talk to me live