
Clinical and research work rarely fails because we cannot locate information; it fails because we cannot convert information into usable outputs quickly enough. The friction lives in the in‑between tasks: preparing a background brief for a case conference, turning scattered notes into a coherent protocol amendment, cleaning citations before a resubmission, or drafting a patient-facing handout that is both readable and accurate. What is newly compelling about the idea of an “AI computer” is that it aims at this connective tissue of work, not simply at conversation.
Perplexity describes Computer as a unified, cloud-based system that can “research, design, code, deploy, and manage” projects end-to-end, breaking a goal into sub-tasks and routing them across specialized components. In the public description, it can orchestrate work across 19 models in parallel, “matching each task to the best model,” while also remembering prior context and connecting to external services. The ambition is not subtle: it is a bid to turn the browser into an operational workspace where the user specifies intent and the system performs the multi-step work.
The attraction is easy to understand in real workflows. We already practice constrained delegation every day: we assign parts of a project to trainees, research assistants, administrators, or colleagues, and we integrate, verify, and sign off. An agentic system promises a similar pattern, but at a different speed and scale. If it performs reliably, it may create time for what remains stubbornly human: therapeutic presence, clinical judgment under uncertainty, nuanced supervision, and careful interpretation of evidence.
That tension becomes sharper when orchestration is invisible. If Computer decomposes a task into steps, chooses tools, and merges results, then provenance matters: which sources were used, which model generated which claim, and what was the system’s uncertainty at each step? In research, these details determine whether a literature synthesis is reproducible; in clinical settings, they determine whether a handout, policy memo, or documentation aid stays within the boundaries of evidence-based practice. The more autonomous the workflow, the more we need systems that make their reasoning legible rather than merely impressive.
A second tension is provenance. In research, we need to know what sources were used, how claims were derived, and what uncertainty remains, because the credibility of a synthesis depends on traceable reasoning. In clinical environments, provenance is equally important, though less often formalized: we need to know whether an output is grounded in guidelines, high-quality trials, local policy, or merely plausible generalizations. Agentic tools can compress steps so efficiently that they also compress our visibility into where a conclusion came from.
Cost pulls these questions from theory into daily decision-making. At roughly $240 per month, this is not an impulsive subscription for most clinicians; it is closer to a staffing tradeoff. Paying that amount implicitly assumes that time saved is both substantial and dependable, and that the time we spend verifying the output does not quietly re-inflate the workload. In clinical settings, the “true cost” includes not only money, but also the cognitive burden of oversight and the reputational risk of errors.
From a practice perspective, the safest near-term uses are those that keep identifiable data out of the system and keep verification firmly human. Drafting non-identifying psychoeducation templates, creating training materials for interns, turning internal procedures into clearer language, or generating first-pass outlines for research documents can be sensible, provided we treat outputs as drafts and insist on source checking. The risk profile changes sharply when we move toward identifiable case details or highly specific clinical recommendations, especially in small communities or rare presentations where re-identification can be easier than we like to admit.
We also need to acknowledge a quieter limitation that experienced researchers recognize: these tools can accelerate the appearance of scholarship. They can produce coherent framing, persuasive prose, and tidy synthesis even when the evidence base is mixed or contested. The danger, then, is not only “hallucination” in the headline sense; it is routine overconfidence, particularly under deadline pressure, fatigue, or institutional incentives that reward speed over carefulness.
Ethically, we should treat agentic systems as a new layer of professional delegation that demands transparency and documentation habits. If AI materially shaped an output that informs care (a clinic policy, a patient handout, a decision support memo), the clinician’s responsibility is not reduced; it is reconfigured. We owe patients and colleagues a disciplined stance on what data entered the system, what sources were relied upon, and how claims were checked. This is consistent with broader AI-risk frameworks emphasizing lifecycle governance: mapping likely failures, setting boundaries, and building habits of verification rather than relying on good intentions.
Looking ahead, the question is not whether “AI computers” will become more common; they likely will. The more important question is whether they become legible: systems that make their sources, assumptions, and limitations visible enough for clinical and research cultures that depend on auditability and trust. If we adopt them thoughtfully, starting with low-risk tasks, measuring time saved against time spent verifying, and maintaining strict boundaries around sensitive data, we can treat these tools as assistants rather than authorities, and preserve the integrity of our work while reducing avoidable friction.
