The things AI agents still can’t do

The things AI agents still can’t do

6 min read
    Tags:

    AI agents are getting better at reasoning, but most still struggle to act inside the real systems where work happens. The biggest constraint in agentic AI today isn’t intelligence. It’s execution across inboxes, calendars, and identity systems.


    AI agents are getting better at reasoning, which is why interest in agentic AI systems has exploded over the past year. They can draft emails, summarize meetings, extract next steps, and plan multi-step workflows. In controlled demos they increasingly appear autonomous.

    In production environments, however, most agents still run into the same constraint: they can’t reliably interact with the systems where people actually coordinate work. Most agentic AI systems still rely on integrations with real email APIs, calendar APIs, and identity systems to take action inside real workflows.

    That limitation rarely appears in demos. It becomes obvious the moment an agent attempts to operate inside real software used by real teams.

    Because most business workflows don’t end with reasoning. They end with communication. Someone sends an email, schedules a meeting, updates a system of record, or replies to a conversation already in progress.

    Until an agent can operate inside those communication systems, it can’t actually execute the workflow it reasoned about.

    That distinction is quietly becoming one of the most important constraints shaping the next phase of agentic AI.

    Why reasoning is no longer the hard problem in agentic AI

    Large language models have dramatically lowered the cost of reasoning. Generating a follow-up email, summarizing a conversation, or identifying the next action in a workflow is now relatively straightforward, and many applications already do this well.

    Execution is harder.

    An agent that drafts an email isn’t the same as an agent that can send it from a real inbox, maintain the thread across replies, and react when the recipient responds.

    An agent that suggests meeting times isn’t the same as one that can coordinate availability across calendars, handle time zones, send invitations, and update the event when someone reschedules.

    Reasoning about a workflow is one problem. Acting inside the systems where that workflow actually happens is another.

    Most agents today can do the first. And far fewer can do the second.

    Where work actually happens

    If you look closely at where work moves forward inside orgs, you’ll notice a clear pattern.

    The coordination layer is almost always communication infrastructure.

    Deals advance through email conversations, support issues move through shared inboxes, interviews are coordinated through calendars, and projects move forward through meetings and follow-ups.

    Even highly automated SaaS products still rely on these systems because they connect software to people.

    That means an agent can’t fully execute most real workflows unless it can operate inside those environments — not simulated ones, but real inboxes, real calendars, and real identities.

    Where AI agents break in production

    This gap becomes obvious when teams attempt to deploy agents in production.

    Take a sales workflow as an example. An agent might correctly identify the next step after a customer call. It drafts a follow-up email, suggests a meeting, and recommends updating the CRM.

    Without access to the actual inbox and calendar involved in that conversation, none of those actions can happen automatically. The system can only make suggestions.

    The same pattern appears in support operations. Classifying an incoming message is easy. Routing it, responding from the correct shared inbox, tracking replies, and escalating issues requires interacting with the real mailbox where those conversations live.

    Recruiting coordination exposes the same limitation. Suggesting interview times is trivial. Actually scheduling interviews across multiple calendars, handling time zones, managing reschedules, and sending confirmations requires operating inside live calendar systems.

    Even meeting intelligence products follow the same pattern. Generating a transcript is now table stakes. Triggering recap emails, scheduling follow-ups, and updating downstream systems requires acting within the communication infrastructure where the meeting participants already operate.

    Across these workflows, the useful part isn’t reasoning about what should happen next. It’s executing the next step inside the systems where the conversation continues.

    The real constraint in agentic AI

    This is why many teams find that moving from a working demo to a reliable production agent is harder than expected.

    The challenge is rarely intelligence. More often, it’s integration.

    Agents need to operate inside real inboxes, real calendars, and real identity boundaries. They need to maintain thread continuity, respond to events, and respect permission models across providers.

    Those requirements introduce complexity that doesn’t appear in model benchmarks or demo environments.

    But they are exactly what make the difference between a system that suggests actions and one that actually performs them.

    What this means for builders

    For teams building agentic workflows, this shifts where the real work happens.

    The hard problem is no longer just improving reasoning. It’s building systems that allow agents to interact safely and reliably with the environments where human coordination occurs.

    That means connecting to live communication systems, respecting identity and permission boundaries, maintaining context across conversations, and reacting to real-time events such as replies or calendar updates.

    In other words, enabling agents to act inside the infrastructure where work already moves.

    Because an agent that can’t communicate with people through the systems they already use isn’t truly autonomous; it’s simply suggesting what someone else should do next.

    The next phase of agentic AI

    LLMs made reasoning widely accessible.

    The next phase of agentic AI will be defined by something more practical: which systems allow agents to reliably take action in the real workflows where work happens.

    That means interacting with inboxes, calendars, meetings, and the identity layers behind them — not simulated data or isolated APIs, but real communication environments.

    Until agents can operate there, their intelligence will remain partially constrained.

    Once they can, the difference between suggesting a workflow and executing one begins to disappear.


    Agentic AI guides from Nylas

    Industry insights

    Build guides

    Related resources

    How agentic AI adoption is scaling from the inside out

    Public conversations about agentic AI focus on what customers will see. Autonomous support agents. AI…

    The production stack for agentic AI

    Agentic AI has moved past the demo stage. In public conversations, the focus often remains…

    The infrastructure race behind agentic AI

    Much of the conversation around agentic AI focuses on models: reasoning improvements, autonomy levels, and…