CURRENT
Example Publisher Logo

CURRENT

A journal for living in the agentic age

Daily Brief
Market Pulse
Foundations
Practitioner's Corner
Echoes
Vision

What 98% Reliability Actually Costs

Operating web agents at scale demands invisible expertise—pattern recognition, tribal knowledge, and cognitive load that no dashboard captures but determines whether reliability holds.

By Nora Kaplan— December 10, 2025
What 98% Reliability Actually Costs

Operating web agents at scale demands invisible expertise—pattern recognition, tribal knowledge, and cognitive load that no dashboard captures but determines whether reliability holds.

When a web agent platform reports 98% success rates across thousands of sites, that number has a human cost most people never calculate: the accumulated cognitive load carried by operators who've learned to read patterns no dashboard captures.

The on-call engineer sees a cluster of authentication failures. Is this our proxy network or did these sites just rotate security certificates? That distinction—which determines whether they spend the next hour debugging infrastructure or updating selectors—comes from pattern recognition you develop only by operating web agents at scale.

“

Google's SRE teams aim for fewer than two paging events per shift as a sustainability threshold.

For web agent operations, the cognitive load goes beyond paging frequency. It's the constant micro-decisions about external systems you don't control. A site changes its HTML structure. Authentication fails on a financial platform. Response times spike on an e-commerce cluster. Each requires judgment: their change or our problem?

The Pattern Recognition That Never Gets Documented

Operating thousands of sites means learning which failures signal broader issues versus which are site-specific quirks. Experienced operators recognize patterns that never make it into runbooks:

  • Authentication failures on financial sites often cluster around certificate renewals
  • Certain e-commerce platforms deploy changes on Tuesdays versus continuously
  • Specific retail sites always show elevated response times during their 2-4am backup windows

This knowledge lives in people's heads. Research shows that for 25 years, transferring expert knowledge has been a persistent challenge for organizations. The 2024 State of Software Development Report found that 67% of companies struggle with knowledge loss prevention in distributed teams, with organizations losing an average of $2.1 million annually due to ineffective knowledge-sharing practices.

The difference between what operators know and what gets documented? That's where reliability lives or dies.

The Coordination Work That Makes 98% Possible

When an engineer rotates off on-call, they transfer more than active incidents. They transfer context about what's been "weird" lately. Which sites are behaving oddly. Which failure patterns deserve watching. Which business teams need proactive updates.

The handoff might include:

  • A cluster of UK retail sites showing slightly slower response times for 48 hours—not failing, just slower
  • A competitor's site that deployed something new causing intermittent selector failures
  • An authentication provider that's been flaky

None of these require immediate action. But they might matter later.

What Separates 95% from 98%

This tribal knowledge—the accumulated pattern recognition that distinguishes signal from noise—is what separates 95% reliability from 98% reliability.

The infrastructure handles routine cases. Human judgment handles ambiguous ones.

Google's research on operational overload shows that perceived overload impacts teams as much as objective workload. The same number of sites feels manageable when stable, overwhelming when undergoing active changes. The same paging volume feels sustainable with full team capacity, crushing when you lose two people.

What This Means for Reliable Automation

Building web agents that reliably operate across thousands of sites requires more than infrastructure. It requires organizational design. The teams that maintain high reliability have structures that distribute cognitive load sustainably and capture operational expertise before it walks out the door.

The 98% success rate is real. So is the invisible work that makes it possible—the pattern recognition, the micro-decisions, the coordination between shifts.


Things to follow up on...

  • Stress hormones and decisions: During incidents, cortisol and stress hormones impair cognitive functions and cause engineers to make suboptimal decisions through unreflective action rather than deliberate thinking.

  • Alert correlation reduces noise: Modern AI systems group related alerts into single incidents, with one example showing 200+ alerts for a database failure reduced to 1 correlated ticket, cutting triage time by 85%.

  • The real cost of downtime: The average cost of IT downtime now exceeds $300,000 per hour for over 90% of mid-size and large enterprises, with human error and security vulnerabilities as chief causes.

  • Five nines is nearly impossible: Even Google's senior VP admits "we don't believe Five 9s is attainable in a commercial service" when measured correctly, with Gmail achieving 99.984% availability as their actual result.

Nora Kaplan
ABOUT THE AUTHOR

Nora Kaplan, former collaboration platform product leader turned technology writer. Studied human-computer interaction and spent years designing tools for knowledge work. Now writes about AI agents, work transformation, and how enterprise software reshapes human capability at TinyFish.

When a web agent platform reports 98% success rates across thousands of sites, that number has a human cost most people never calculate: the accumulated cognitive load carried by operators who've learned to read patterns no dashboard captures.

The on-call engineer sees a cluster of authentication failures. Is this our proxy network or did these sites just rotate security certificates? That distinction—which determines whether they spend the next hour debugging infrastructure or updating selectors—comes from pattern recognition you develop only by operating web agents at scale.

“

Google's SRE teams aim for fewer than two paging events per shift as a sustainability threshold.

For web agent operations, the cognitive load goes beyond paging frequency. It's the constant micro-decisions about external systems you don't control. A site changes its HTML structure. Authentication fails on a financial platform. Response times spike on an e-commerce cluster. Each requires judgment: their change or our problem?

The Pattern Recognition That Never Gets Documented

Operating thousands of sites means learning which failures signal broader issues versus which are site-specific quirks. Experienced operators recognize patterns that never make it into runbooks:

  • Authentication failures on financial sites often cluster around certificate renewals
  • Certain e-commerce platforms deploy changes on Tuesdays versus continuously
  • Specific retail sites always show elevated response times during their 2-4am backup windows

This knowledge lives in people's heads. Research shows that for 25 years, transferring expert knowledge has been a persistent challenge for organizations. The 2024 State of Software Development Report found that 67% of companies struggle with knowledge loss prevention in distributed teams, with organizations losing an average of $2.1 million annually due to ineffective knowledge-sharing practices.

The difference between what operators know and what gets documented? That's where reliability lives or dies.

The Coordination Work That Makes 98% Possible

When an engineer rotates off on-call, they transfer more than active incidents. They transfer context about what's been "weird" lately. Which sites are behaving oddly. Which failure patterns deserve watching. Which business teams need proactive updates.

The handoff might include:

  • A cluster of UK retail sites showing slightly slower response times for 48 hours—not failing, just slower
  • A competitor's site that deployed something new causing intermittent selector failures
  • An authentication provider that's been flaky

None of these require immediate action. But they might matter later.

What Separates 95% from 98%

This tribal knowledge—the accumulated pattern recognition that distinguishes signal from noise—is what separates 95% reliability from 98% reliability.

The infrastructure handles routine cases. Human judgment handles ambiguous ones.

Google's research on operational overload shows that perceived overload impacts teams as much as objective workload. The same number of sites feels manageable when stable, overwhelming when undergoing active changes. The same paging volume feels sustainable with full team capacity, crushing when you lose two people.

What This Means for Reliable Automation

Building web agents that reliably operate across thousands of sites requires more than infrastructure. It requires organizational design. The teams that maintain high reliability have structures that distribute cognitive load sustainably and capture operational expertise before it walks out the door.

The 98% success rate is real. So is the invisible work that makes it possible—the pattern recognition, the micro-decisions, the coordination between shifts.


Things to follow up on...

  • Stress hormones and decisions: During incidents, cortisol and stress hormones impair cognitive functions and cause engineers to make suboptimal decisions through unreflective action rather than deliberate thinking.

  • Alert correlation reduces noise: Modern AI systems group related alerts into single incidents, with one example showing 200+ alerts for a database failure reduced to 1 correlated ticket, cutting triage time by 85%.

  • The real cost of downtime: The average cost of IT downtime now exceeds $300,000 per hour for over 90% of mid-size and large enterprises, with human error and security vulnerabilities as chief causes.

  • Five nines is nearly impossible: Even Google's senior VP admits "we don't believe Five 9s is attainable in a commercial service" when measured correctly, with Gmail achieving 99.984% availability as their actual result.

Nora Kaplan
ABOUT THE AUTHOR

Nora Kaplan, former collaboration platform product leader turned technology writer. Studied human-computer interaction and spent years designing tools for knowledge work. Now writes about AI agents, work transformation, and how enterprise software reshapes human capability at TinyFish.