The Hidden Cost of AI

Posted on by Adam Wespiser

Most AI tools promise to save time. Many of them instead accumulate attention debt.

The cost of AI is not compute. It is not data. It is not headcount.

The real cost of AI is attention.

You can think of this as attention debt.

Every prompt, evaluation, correction, and context switch adds a small cognitive obligation that someone eventually has to pay. But the most revealing symptom of this problem appears when organizations try to deploy multiple AI agents.

They often build systems that predict which AI agent should handle a problem.

At first glance this seems sensible.

But if a system can reliably predict the correct agent, and that agent can reliably solve the problem, then routing the human to that agent is already the wrong design.

The system already knows enough to execute the solution directly.

Predicting the right tool but still requiring the human to invoke it is simply moving the cognitive burden around rather than removing it.

This reveals the deeper issue: most AI systems optimize for capability, not attention cost.


The Real Problem With AI Agents

Most discussions about AI agents focus on capability.

Does it work?
Is it accurate?
Is the problem big enough?

Those questions matter, but they are not sufficient. The real stack of problems looks more like this:

AI Agent Concern Stack

        ┌───────────────────────────────┐
        │ 3. Emergence Layer            │
        │ Will it remain sticky?        │
        │ Does it survive real use?     │
        │ Emergent behavior • Drift     │
        └──────────────▲────────────────┘
                       │
        ┌──────────────┴────────────────┐
        │ 2. Adoption Layer             │
        │ Will users even try this?     │
        │ Fear • Avoidance • Habit      │
        │ Workflow friction             │
        └──────────────▲────────────────┘
                       │
        ┌──────────────┴────────────────┐
        │ 1. Capability Layer           │
        │ Does it work?                 │
        │ Correctness • Utility         │
        │ Technical feasibility         │
        └───────────────────────────────┘

Most engineering effort focuses on Layer 1: capability.

But in real organizations, systems often fail at Layer 2 (adoption) or Layer 3 (emergent behavior).

Even a system that technically works can fail if users avoid it, or if real‑world usage patterns slowly erode trust and usefulness.

Each layer also increases attention cost:

The second and third layers are frequently ignored because they aren’t purely technical problems.

AI resistance is real. A Gallup poll found that 64% of people report fear or concern about AI. Fear cannot always be solved through better product design when the technology challenges people’s identity, expertise, or job security. Even if fear disappears, a deeper constraint remains: human attention.


Attention Is the Real Scarce Resource

Attention behaves very differently from time or compute. It is:

Every time a worker switches attention they pay several cognitive penalties:

Context reconstruction
What was I doing before?

Semantic evaluation
What does this new input mean?

Trust calibration
Can I rely on this system?

Modern corporate environments are already oversubscribed with attention demands:

AI adds a new surface area of interruption.

Even if AI accelerates a single local task, the system can still decrease overall productivity if it increases the global cognitive load.


Why Many AI Tools Fail in Practice

AI tools often demonstrate their capabilities clearly in controlled settings but can be harder to integrate smoothly into real workflows.

This happens because most AI tools require users to actively manage them.

They are often:

Interactive

Users must wait for responses, evaluate results, and continue the exchange.

Conversational

Users must remember what the system can do and how to phrase requests.

Flexible

Users must determine whether the tool applies to the situation at all.

Numerous

Organizations deploy many agents, each with slightly different capabilities.

The result is a new cognitive tax:

“Which AI should I use for this problem?”

Ironically, the more human-like the interface becomes, the more it competes for human attention.

Human conversation is one of the most expensive cognitive interfaces we have.


The Organizational Cost of Attention

Attention loss isn’t just an individual productivity issue. It becomes a direct cost center at organizational scale.

Interruption density scales roughly with headcount.

If a tool saves five minutes locally but introduces attention overhead across dozens of workers, the total system cost can quickly exceed the benefit.

This is particularly visible in internal AI systems.


Example: AI Agents for Support Engineers

Consider a common pitch:

“The AI agent will resolve support tickets faster.”

In reality, support engineers often need to:

Each of these steps consumes attention.

This helps explain why AI tools can perform well in demonstrations yet require additional design work to fit naturally into production workflows.

This leads to a deeper question.

Many organizations attempt to solve this by building systems that predict which AI agent should be used for a given problem.

But if we can reliably predict the correct agent, and that agent can reliably solve the problem, then routing the human to that agent is already the wrong design.

At that point the system already knows enough to execute the solution directly.

Predicting the right tool but still requiring the human to invoke it is simply moving the cognitive burden around rather than removing it. In other words, an AI router that still requires human activation is not automation. It is closer to workflow triage than to full automation.

The missing variable isn’t accuracy.

Typical AI Workflow

Human
  ↓
Identify Problem
  ↓
Choose Agent
  ↓
Prompt Agent
  ↓
Evaluate Output
  ↓
Fix Errors
Attention‑Aware AI Workflow

Problem Appears
  ↓
AI Identifies Pattern
  ↓
AI Executes Solution
  ↓
Human Notified Only If Needed

This contrast illustrates the real design shift: removing humans from the operational loop rather than giving them better tools to manage it.

It’s cognitive overhead.


Designing Attention-Aware AI Systems

If attention is the constraint, AI systems should be designed to minimize the cognitive load they impose.

An internally facing agent should:

Minimize interruptions

The system should run silently whenever possible.

Prefer execution over conversation

Automation should replace interactive prompting.

Escalate selectively

Only interrupt humans when:

Batch cognitive demand

Aggregate decisions instead of requesting constant feedback.

Align with natural human checkpoints

Provide summaries during moments where attention is already available.

Be legible at a glance

Binary states outperform narrative explanations.

Good status indicators include:

Probabilistic narratives force additional cognitive evaluation.


The Best AI Systems Are Invisible

Good infrastructure is rarely noticeable.

The best systems are boring.

They operate silently, reliably, and predictably.

Success should go unnoticed and failures should be crisp, bounded, and recoverable.

AI systems should aim for the same design philosophy.


Rethinking AI Through the RACI Framework

Another useful way to think about AI agents is through the RACI framework:

Many AI tools today occupy the Consulted role.

Humans remain responsible and accountable, while the AI provides suggestions.

This is precisely the configuration that maximizes attention cost.

The real opportunity is to shift AI left in the RACI model.

AI agents should increasingly take the Responsible role — executing the primary work — while humans remain accountable and consulted only when necessary.

In other words:

Don’t ask the AI for advice.

Ask it to do the work.

Then notify humans when an expensive or irreversible decision must be made.


The Real Goal of AI

The goal of AI is not to make humans faster at operating software.

The goal is to make software operate without requiring human attention.

The organizations that succeed with AI will not be the ones with the most advanced models or technically impressive demos.

They will be the ones that design systems where AI quietly performs work in the background, while humans focus their limited attention on the few decisions that truly matter.