AI Strategy & Consulting

Digital workers are here. Your workflows are not ready for them.

Agentic AI platforms can now take a business goal, decompose it into tasks, and run those tasks across your software stack without a human directing every step. That is not an upgrade on Copilot. It is a different class of adoption, and most mid-market organisations are not ready for it.

Discuss this article with AI

I have watched the same failure pattern play out across four technology shifts: dotcom, mobile, cloud, and now AI. What changes is the technology. What doesn't change is the readiness gap. The tools arrive before the workflows are designed to use them. With agentic AI, the consequences of that gap are no longer occasional errors. They are systematic ones.

In 2026, the first generation of agentic platforms has moved from enterprise-only pilots to mid-market availability. Perplexity Computer is the clearest current signal that digital worker technology has arrived for mid-market use. It takes a goal, spins up sub-agents to execute each task, and coordinates them across your applications and data. The question for a COO is not whether this technology works. It's whether the work you would hand it is designed well enough to automate.

What does the shift from AI tools to agentic workflows mean for operations?

Agentic AI workflows are end-to-end business processes run by AI agents rather than humans: the agent receives a goal, breaks it into steps, executes each step using tools and data, and returns a result or escalates when it cannot proceed.

That single sentence contains the entire shift. In the AI tool era, a human ran the process, used the tool at a step and moved on. The agent was an assistant. With agentic AI, the relationship inverts: the agent runs the process. The human defines the goal, governs the exceptions, and approves the outputs.

The contrast is clearest in a concrete example. A COO asks the system to reconcile last week's project delivery against billing. In a tool-assisted workflow, a human opens the CRM, pulls the records, exports the delivery log, and compares them. In an agentic workflow, the agent receives the goal, queries both systems directly, runs the comparison logic, and produces a structured exception summary. Multi-agent orchestration handles the parallel task execution. The human receives the output, not the process.

When the agent owns the workflow, three things must exist that most organisations don't currently have: the workflow defined at step level, escalation rules explicit, and the audit trail designed in. Without these, the agent guesses. It guesses consistently.

What 'digital worker' means in practice

A digital worker is not a bot that responds to queries. It is an agent that receives an objective and executes toward it: coordinating tools, APIs, and data sources, making intermediate decisions, and producing a structured output or escalation.

Three elements define a digital worker: goal decomposition, tool coordination, and structured output with audit trail.

Finance and operations. An agent queries the CRM for client records and billing events, queries the project management platform for delivery milestones and hours logged, reconciles the two, auto-resolves discrepancies below a defined threshold, and escalates anything above with a structured exception summary ready for human decision.

Marketing. An agent pulls channel performance data on a daily cadence, compares spend against targets, identifies underperforming lines, and surfaces budget reallocation candidates. The CMO reviews structured signal, not raw dashboards.

Sales. An agent pulls CRM history, recent LinkedIn activity, and public news for a target account; assembles a structured brief; and drafts contextualised outreach for the rep to review before sending. The rep's time goes to judgement and relationship, not research.

Each example has the same architecture: goal in, structured output out, human at the decision layer rather than the execution layer. A Perplexity-class digital worker handles the high-frequency, low-value execution work, so the human team moves to the high-complexity, high-value decisions that require judgement. The question is whether the workflow they are joining has been designed for them.

Why is workflow design the bottleneck in agentic AI deployment?

The technology works. Most mid-market business processes are not designed for agent execution. A capable agent running a badly-designed workflow does not produce good results faster. It produces consistent, systematic errors at scale.

A badly-designed workflow run by a human produces occasional errors. The human notices something looks wrong, corrects it, moves on. The same workflow run by an agent produces the same error every time, at the speed of software, until someone notices. The design flaw is not the agent. It is the workflow.

Three questions most COOs cannot confidently answer about any of their core processes: Can you describe this process in atomic steps, with defined inputs and outputs at each stage? Do you have explicit escalation rules — what the agent decides alone versus what it refers back? What does the audit trail look like — can you reconstruct every decision after the fact?

Underneath those steps sits your data and knowledge layer. In practice, effective agentic AI means grounding the workflow in an enterprise knowledge graph: a coherent view of customers, products, and processes, rather than letting agents pull from scattered, inconsistent sources.

Three examples of how this design pattern works in practice:

Finance and operations. This isn't speculation. At Graph Digital, we run agentic workflows in our own operations. Our financial reconciliation process: an agent queries our CRM for client records and billing events; a second agent queries our project management platform for delivery milestones and resource allocation; a reconciliation agent compares the two and auto-resolves discrepancies below a defined threshold. Above threshold, a structured exception summary goes to human review. A reporting agent then compiles the weekly operational report. The human reviews exceptions and approves the report. Not because we outsourced the judgement, but because we designed the workflow first.

Marketing. An agent monitors campaign spend against performance targets daily, flags anomalies, and surfaces budget reallocation options. The CMO reviews an exception list, not a dashboard.

Customer service. Incoming tier-1 queries go to an agent that applies resolution logic against a knowledge base; queries within scope are resolved and closed; edge cases and emotionally complex interactions route immediately to a human. The escalation is not a failure state. It is a deliberate design feature: the explicit point where the human-agent handoff is built into the process rather than discovered as a failure mode.

The design pattern across all three: atomic steps, explicit escalation rules, human governance at output level. This is what the agentic operating model requires. Not a bigger investment. Not a different technology. A different design discipline applied to existing work.

In the workflows we design at Graph Digital, the single most common gap is not agent capability. It is the absence of explicit escalation rules.

Graph Digital is an AI workflow design advisory for UK mid-market organisations. We design and deploy the agentic operating models the five readiness questions point toward.

What IT and risk will flag, and why they're right

Shadow AI risk, data access scope, agent observability requirements, and auditability are not technology objections. They are the governance questions every agentic deployment must answer before it goes live.

Shadow AI risk. Shadow AI is the real one. Employees who discover Perplexity Computer or similar tools will wire them into CRM data, finance records, and customer information without IT visibility or governance. The agentic version has a larger blast radius than standard AI tools because agents take actions, not just answers.

Over-access. AI agents granted broad permissions can trigger payments, data writes, and communications that no one intended to automate. Defining the access boundary is a governance decision, not an IT configuration detail.

Agent observability. The more insidious risk is silent failure: a human who gets a wrong result usually notices something is off. An agent returns an incorrect result with equal confidence and will continue doing so, at scale, until someone reviews the output. Traditional system logging captures API calls and timestamps, not prompts, tool chains, or intermediate decisions. When a compliance team needs to reconstruct what happened in a regulated process, that gap becomes an audit risk by design.

Regulatory exposure. In every agentic deployment I have worked through, the IT teams who raised governance questions earliest produced the most reliable outcomes. Not because they slowed the project, but because they caught the permission and escalation gaps before the agent found them. The governance work and the workflow design work are not separate workstreams. They are the same work.

Five readiness questions before you deploy anything

Before you deploy Perplexity Computer or any other digital worker on a business process, these five questions will tell you whether the workflow is ready.

1. Can you describe this process in atomic steps, with defined inputs and outputs at each stage?

If the answer is "roughly" or "the team knows how it works," the process is not ready. Agents execute against explicit definitions, not institutional knowledge. That documentation is the design work the agent requires before it can be trusted to run.

2. Do you have explicit escalation rules — what the agent decides alone versus what it refers to a human?

Without a designed escalation boundary, the agent either decides too much (producing systematic errors on edge cases) or too little (generating constant interrupts). Escalation rules must be explicit and testable before the agent runs in production.

3. Do you have an audit trail design — a record of every decision the agent makes?

Every run should produce a structured record: what was processed, what was decided, what was escalated, and what the human decided on each exception. For regulated or customer-facing processes, this is not optional. It is hardest to retrofit after deployment.

4. Have you defined the access scope — what systems and data the agent can read or write?

Access scope is a governance decision with commercial consequences. An agent with broad access can trigger unintended actions across the business. Define the boundary before the first run.

5. Is there a named governance owner — not "the IT team", but a specific person accountable for outcomes?

"The IT team owns it" is not governance. There must be a named person who reviews outputs, approves threshold changes, and is accountable when the process produces an error.

If you cannot confidently answer all five for any process you are considering for agent deployment, the workflow is not ready. That is not a problem. It is a design brief. Answering them is the foundation of an agentic operating model and the work that makes every subsequent deployment reliable.

The design decision before deployment

The organisations that get durable value from digital workers are not the ones that deploy fastest. They are the ones that design first.

Most mid-market operations have workflows built around human execution: institutional knowledge fills gaps in process documentation, escalation happens informally, and the audit trail lives in email threads. Deploying a digital worker onto that foundation produces the failures the process has always contained, running faster and more consistently than before.

The AI Portfolio Review maps which of your processes are ready for agent execution now, which need redesign before they can be automated reliably, and which should not be automated at all, because the human judgement in them is not replaceable by a threshold rule. It is the design decision made before deployment, not the diagnostic run after something goes wrong.

Key takeaways

  • Agentic AI runs business processes end-to-end. It does not assist with steps; it replaces the human as executor of the workflow, which changes what governance is required.
  • A badly-designed workflow run by an agent produces consistent, systematic errors at scale, not the occasional errors that human-run processes absorb.
  • The bottleneck in agentic AI deployment is not agent capability. It is workflow design: the absence of atomic steps, explicit escalation rules, and a designed audit trail.
  • IT's concerns about agentic AI (shadow AI risk, over-access, agent observability gaps, regulatory exposure) are legitimate governance requirements that belong in the workflow design.
  • The organisations that deploy agentic workflows most successfully treat workflow governance as part of the design process from the start, not a compliance step added afterward.
  • Five readiness questions determine whether a workflow is ready: atomic steps defined, escalation rules explicit, audit trail designed, access scope bounded, governance owner named.

Frequently asked questions

What is an agentic workflow?

An agentic workflow is an end-to-end business process run by an AI agent rather than a human. The agent receives a goal, breaks it into discrete tasks, executes each task using tools and data, applies decision rules at defined points, and returns a structured output or escalates to a human when it cannot proceed. The distinguishing feature is that the agent owns the execution of the process, not a single step within it.

What is a digital worker?

A digital worker is an AI agent designed to execute business objectives rather than respond to queries. It coordinates tools, APIs, and data sources, makes intermediate decisions according to defined rules, and produces structured outputs with an audit trail. Perplexity Computer is a current example: it takes a goal, spins up sub-agents to execute component tasks, and completes the workflow rather than returning a single answer. A digital worker operates at the process level, not the prompt level.

How is agentic AI different from Copilot or ChatGPT?

Unlike Copilot, which operates within a single prompt-response cycle where a human directs each step, an agentic workflow takes a goal and executes autonomously across multiple steps, tools, and decisions. The human receives an output rather than directing each step. The practical difference: Copilot helps you draft an email; an agentic workflow queries the CRM, identifies accounts requiring follow-up, drafts outreach, and routes it for rep approval, without the human managing the intermediate steps.


Stefan Finch — Founder, Graph Digital

Stefan is an AI strategy advisor to leaders in complex B2B organisations. With 26 years across enterprise and mid-market companies, he advises boards and leadership teams on AI initiatives, sequencing, and roadmap, and builds the agentic infrastructure to execute it.

Connect with Stefan: LinkedIn

Graph Digital provides AI strategy and consulting for mid-market B2B companies in the UK, Europe, and the US — helping executive leaders move from scattered pilots to a prioritised AI roadmap and measurable commercial outcomes. AI strategy and advisory →