← Back to all blogs

Louise Cermak | 27 January 2026

What ‘Good AI’ Actually Looks Like Inside a Real Workflow

Most Enterprise AI is currently expensive shelf-ware.

Enterprise AI

A growing number of custom and experimental AI initiatives are failing to move beyond the pilot stage.

CFOs and CTOs are reviewing adoption metrics for general-purpose, out-of-the-box copilots and seeing little sustained usage. After the initial novelty of a conversational assistant wears off, the reality becomes clear – many teams are not using these tools in the flow of their day-to-day work.

This is not a technology problem. It is a workflow problem.

Good AI is not defined by what it can generate. It is defined by whether it becomes part of how work actually gets done.

In an enterprise organisation, AI rarely fails because the models are weak. It fails because it never survives contact with a real Tuesday morning. The moment a tool forces someone to stop what they are doing, open a new tab, or second-guess an answer, it has already broken the workflow.

For operators in regulated industries including pharma, financial services, or government, novelty is not the objective. The priority is reducing the friction that quietly consumes delivery capacity, slows decisions and inflates cost.

This is where most AI initiatives fall short.

The Antagonist. Why general purpose copilots fail the Tuesday morning test

Many organisations have fallen into the ‘licence trap’, assuming that a global, out-of-the-box large language model with a Copilot badge will solve their knowledge and decision-making challenges.

In high-stakes environments, that assumption rarely holds.

Generic AI fails in regulated workflows for three structural reasons

1. The context gap

A general purpose model may know everything about the public internet, but it knows nothing about your internal, version-controlled source of truth. Ask it how to structure an NDA and it will give you a reasonable answer. In a regulated enterprise, ‘reasonable’ is not enough. You need the approved, current, legally governed version.

2. Portal fatigue

Most AI deployments create another destination – another URL, another login, another interface. But work does not happen in portals. It happens in collaboration tools and delivery platforms. If AI is not embedded where decisions are made, adoption will stall.

3. The hallucination risk

In regulated workflows, ‘mostly right’ is a liability. General purpose models are optimised for fluency, not assurance. When people cannot trust the answer or its source, they fall back to manual processes. The AI becomes a novelty, not an operational tool.

The hidden cost of operational friction

Before defining what good AI looks like, it’s worth understanding the cost of the status quo.

Across large enterprises, operators – engineers, regulators and programme leads, spend a significant part of their day simply trying to find the information they need to move work forward. They navigate fragmented document libraries, shared drives and message threads to locate the one policy, standard, or decision record that unlocks the next step.

This is the search Tax.

It shows up as delayed decisions, broken momentum, duplicated work and growing frustration across delivery teams. It quietly erodes productivity and inflates the cost of every initiative.

Many organisations attempt to solve this by launching a new ‘knowledge portal’. In practice, this rarely helps. A portal is just another destination; another place for information to accumulate, fragment and go stale.

The real problem is not access. It is workflow.

Until trusted guidance is available in the moment decisions are made, the search tax will continue to grow.

The Contrast. Product AI vs workflow AI

To move from strategy decks to operational reality, it is important to distinguish between building an AI product and embedding AI into a workflow.

Product-led AI typically lives in a separate portal or application. It requires users to change behaviour, attend training and remember to use it. In practice, adoption is inconsistent and value is limited.

Workflow-led AI is embedded into the tools teams already use. It operates in the background, supporting decisions in real time and becomes part of how work gets done.

The difference is not technical. It is operational.

System-first. The three pillars of effective AI

If AI is to deliver meaningful return on investment, it must be built on three system-first principles. Invisibility, authority and augmentation.

1. Invisibility. Respecting where work happens

Effective AI respects the reality of modern work. If teams spend most of their day in collaboration and delivery platforms, that is where AI must live.

In a recent engagement with a global pharmaceutical organisation, an earlier attempt at AI failed because it was delivered as a standalone web application. It required users to leave their workflow to interact with it.

By embedding a Knowledge Agent directly into their existing collaboration environment, the AI became part of the day-to-day workflow. Adoption followed naturally, without formal change programmes or behaviour change initiatives.

2. Authority. Building trust through provenance

In regulated environments, the value of AI is not creativity. It is assurance.

Effective AI does not simply provide answers. It shows where those answers come from. Every response is grounded in an approved internal document, with a clear audit trail back to the source. When an answer cannot be found in authorised policy, the system must be able to say so.

This approach turns AI from an experimental tool into a defensible enterprise capability.

3. Augmentation. Supporting decisions in the moment

Poorly designed AI interrupts work. Effective AI supports it.

When a project lead is under pressure to approve a supplier, they should not need to search through document libraries or policy PDFs. They should be able to ask a question in the same place they are already working and receive a clear, referenced answer in seconds.

The workflow continues. The decision improves. The friction disappears.

Case Study. From document overload to decision confidence

To see this in practice, consider the Knowledge Agent framework we deployed for a highly regulated global enterprise.

The organisation was not suffering from a lack of information. It was suffering from an excess of it. More than 20 years of legacy policy documents sat across multiple systems, with overlapping and sometimes conflicting guidance.

In this environment, a general purpose copilot would have been a serious risk. Without a clear understanding of which documents represented the current, approved position, it could easily surface outdated or incorrect guidance.

The strategy shift

Rather than deploying a standalone AI tool, we built a governed data pipeline that prioritised authoritative sources. The system was designed to recognise which document estates contained the current approved policy and which represented historical archives.

The result

Search time did not simply reduce; the nature of work changed. The AI became the first point of reference for policy and standards queries, cutting the search tax and enabling teams to stay focused on delivery.

The real measure of AI success. Adoption in the flow of work

If an AI system requires formal training sessions before it can be used, it is already introducing friction.

In large organisations, complexity is the primary driver of shadow IT. People gravitate towards tools that are intuitive, fast and easy to use. In a real workflow, effective AI feels less like a system and more like an extension of the team – available when needed, invisible when not.

The objective is not to turn employees into AI specialists. It is to help them do their jobs better.

If a tool requires people to learn new interaction models, master prompt techniques, or change how they work, adoption will be limited. The system should adapt to human behaviour, not the other way around.

This is what separates experimental AI from operational AI.

Your commercial lever

When evaluating AI initiatives, leaders should look beyond feature lists and focus on operational impact.

A simple way to assess value is to audit the workflow using three questions:

  1. Input. Where does trusted data live today? Is it fragmented across document libraries, shared drives and email?
  2. Friction. How much effort does it take for someone to find the current, approved guidance they need to make a decision?
  3. Output. Does the AI simply provide a conversational response, or does it support a real operational decision?

If an AI initiative is focused on chat, its impact will be limited. If it is focused on decision support, it becomes a commercial lever.

Final thought. AI as an operational asset

The capability of modern AI models is advancing rapidly and is increasingly accessible. Power alone is no longer the differentiator.

The more important question is whether your AI respects the reality of how work actually gets done.

If it requires people to learn a new system, adoption will be limited. If it asks them to trust answers without evidence, it will not be used.

But if it lives where work happens, provides authoritative guidance and removes friction from everyday decisions, it becomes a genuine operational asset.

This is what separates experimental AI from enterprise AI.

See how trusted AI is deployed in practice

The KnowledgeAgent case study shows how a global pharmaceutical organisation transformed a fragmented document estate into a secure, workflow-embedded decision capability.

It details the operating model, architecture and adoption approach behind a trusted AI system designed for regulated environments.

Download the case study to see how governed, workflow-led AI delivers measurable operational impact.Pharma Leader Supercharges Knowledge Access with KnowledgeAgent