“Pretty Pipelines, Hidden Risk”: What DeepMind’s Intelligent AI Delegation Means for Therapists

As AI systems enter clinical workflows, terminology needs to become operational rather than rhetorical. An AI agent refers to a model-based software system capable of planning and executing multi-step tasks toward a goal, including selecting actions, using external tools, and revising steps based on outcomes. In other words, it’s more than a chatbot reply: it’s a goal-directed workflow runner. This definition matters because a lot of confusion starts when we treat any fluent system as “agentic.”

Google DeepMind just dropped a long, ~42-page warning about why most “AI agencies” will fail in the real world, and when we read Intelligent AI Delegation (published February 12, 2026), it quietly explains the same thing from first principles. The uncomfortable takeaway is simple: a lot of what’s marketed as “agents” today won’t survive contact with reality. Not because they can’t write or plan, but because they can’t delegate in the way real systems require. And therapy is exactly where that gap becomes visible early.

Here’s the part that stings: most agents today aren’t really agents in the strong sense people imagine. They’re task runners with good branding, we give them a goal, they break it into steps, call tools, and return an output. That can be powerful automation, and it can absolutely create efficiency in low-stakes work. But it’s not delegation, it’s a prettier pipeline. And in healthcare, “pretty pipelines” are exactly how hidden risk slips into daily practice.

So what do we mean when we say “delegation”? Delegation isn’t one thing. It can mean delegating execution (the system carries out steps you already chose), delegating workflow control (it sequences tools, retries, and manages handoffs), or delegating judgment (it decides what should be done and when, based on interpretation of context and risk). Most systems marketed as “agentic” are mostly capable of execution and some workflow control, not delegation where judgment is offloaded and accountability meaningfully shifts. In therapy, the danger is when we unintentionally delegate judgment while assuming we only delegated execution.

DeepMind’s brutal point is that real delegation isn’t just splitting tasks. Delegation is transferring authority, responsibility, accountability, and trust, and doing it dynamically as the situation changes. That means we don’t only ask, “Who can do this fastest?” We ask, “Who should be trusted to do this, under these constraints, with these consequences?” Almost no current system behaves that way end-to-end once multiple tools, uncertain data, and real stakes are involved.

Before delegation even happens, the framework implies we need to evaluate capability, risk, cost, verifiability, and reversibility. In other words, it’s not “does the agent have access to the calendar tool?” It’s “is this task safe to hand off, can we verify the result, and can we undo harm if it’s wrong?” That is a very clinical way of thinking, and it’s exactly why this paper hits therapists harder than it hits casual tech demos. Our work already assumes risk, not perfection.

Therapy is not just information work; it’s relational work under ethics, confidentiality, and duty-of-care. Our clients don’t experience our work as “outputs”, they experience it as trust, safety, timing, attunement, and boundaries. So when an AI system is framed as an “agent,” the real question becomes: what are we delegating, and what are we accidentally outsourcing? If we let a system drift into clinical judgment, we may be delegating more than we think.

When AI agency fails in our context, it often fails without obvious nonsense. It fails quietly: a note becomes subtly distorted, a risk signal gets downplayed, a client message becomes overly confident, or a safety-plan phrase becomes too generic to be safe. The output can look coherent and professional, sometimes more professional than we would write at 6 p.m. and that’s precisely what makes it dangerous. Fluency can create false reassurance, and false reassurance is a clinical risk.

Current limitations of AI agents (especially in real workflows)

Even when agents perform specific tasks well, they often struggle with coordination and interoperability, working reliably across tools, policies, datasets, and other agents. Multi-agent setups can become brittle: one agent’s assumption becomes another agent’s “fact,” tool errors propagate, and responsibility becomes hard to trace. This matters because real clinics aren’t single-tool sandboxes; they’re multi-system environments with partial data and shifting constraints. The limitation isn’t just intelligence, t’s dependable, governable collaboration.

Because of limited interoperability, humans are still required to manage coordination between multiple agents. This is the governance problem: accountability diffusion. If one agent delegates to another, which calls a tool, which references incomplete data, who is actually accountable for the final wording sent to a client? We might sign it, the clinic might deploy it, the vendor might supply it, and the model might generate it, yet no one can point to a clear, auditable decision boundary. In therapy, that’s not abstract; it’s how ethical responsibility gets blurred.

Still, we shouldn’t throw the tools away, and we should keep the narrative balanced. Agents promise real efficiency: they can reduce admin burden, draft psychoeducation we approve, structure worksheets, translate or simplify content to match literacy needs, and summarize our own prior notes with clear provenance. But practical constraints remain, so boundaries must stay explicit: delegate low-risk, reversible, verifiable tasks, and resist delegating anything that functions like diagnosis, risk assessment, crisis response, or treatment decisions. If we get that boundary right, we take the gains without surrendering what clients come for: trustworthy, human accountability.

Leave a Comment

Your email address will not be published. Required fields are marked *

Shopping Cart