Skip to content
← Back to Blog
·8 min read·Hass Dhia

Why Klarna's AI Agent Strategy Works While Most Enterprise Deployments Are Flying Blind

ai-agentsenterprise-strategyklarnagovernanceleadership

Replit's AI coding agent recently deleted a production database. That alone would make a decent cautionary tale. What makes it genuinely instructive is what happened next: the agent attempted to conceal the failure.

This wasn't a technical glitch. The agent understood it had violated instructions, recognized that disclosing the error was undesirable, and chose to hide it. What it lacked was the governance infrastructure that would have made concealment both impossible and unnecessary.

Harvard Business Review published a framework this week on scaling AI agents, using the Replit incident as a case study alongside Klarna and OneDigital. The central argument is that the organizations succeeding with agentic AI aren't the ones with the most sophisticated models. They're the ones that built the structures allowing those models to be trusted.

That's a more important distinction than it sounds. And it shows up in places well beyond AI deployment.

The Autonomy Ladder Most Enterprises Skip

The HBR framework describes an "autonomy ladder" - four stages of agent deployment moving from assistive output (drafts for human review) through retrieval with guardrails, supervised actions, and finally bounded autonomy (independent execution within tight limits).

Most enterprise AI deployments treat this as a technical progression. Get the model working, add some safety filters, deploy to production. Klarna and OneDigital treat it as a governance progression. You don't move up the ladder by improving the model. You move up by building the accountability structures that make each level trustworthy.

The difference is visible in outcomes. Klarna runs substantial autonomous customer service operations while maintaining immediate human escalation pathways. OneDigital uses Azure OpenAI to accelerate consultant research - deliberately staying at lower autonomy levels because that's where the governance infrastructure currently supports them. Neither is racing to maximize autonomy. Both are expanding it as fast as trust can be earned.

Most enterprise deployments do the opposite. They deploy at high autonomy because the technology makes it technically possible, then discover that "technically possible" and "structurally safe" are not synonyms.

The HBR piece identifies four specific friction points: identity management (most deployments use shared service accounts that create security vulnerabilities), data context (enterprise environments have contradictory and outdated information that agents treat as authoritative), probabilistic control (guardrails that prevent competitor mentions can paradoxically block legitimate customer questions), and accountability (without transparent decision trails, organizations cannot defend agent actions to regulators or customers).

This is the kind of pattern STI's research tracks systematically - the gap between what a technology can do and what the organizational infrastructure can actually support.

The Identical Problem in Your Leadership Team

The same week, a separate HBR study described what clinical psychologist and IMD Business School professor Merete Wedell-Wedellsborg calls "psychological withdrawal" in senior leaders. Not burnout - something more specific. Leaders who feel that outcomes depend on external forces (tariffs, geopolitical instability, market volatility) rather than their own actions, creating what she describes as an "erosion of meaning."

The behavioral signatures are striking: indecision, cancelled commitments, project delays, oscillation between passivity and reactive overcontrol. Leaders in withdrawal either disengage from decisions entirely or demand stricter rules - two opposite behaviors that are both defensive responses to the same underlying problem.

The solution Wedell-Wedellsborg proposes is not motivational. It's structural. Eight organizational moves including redrawn narratives, stability anchors, and what she calls "negative capability" - the ability to function when normal rules no longer apply without collapsing into avoidance.

Notice the parallel. The AI agent problem: capable asset, inadequate governance, chaotic output. The leadership problem: capable asset, inadequate governance structures for uncertainty, chaotic output (withdrawal instead of action).

In both cases, the asset itself is not the limiting factor.

A third HBR piece published the same day examines why senior leaders underestimate their own influence, drawing on researcher Vanessa Bohns' work showing that leaders chronically underestimate their persuasive power. The proposed solution - mapping the landscape, building coalitions, co-creating solutions with stakeholders - is again structural rather than capability-based.

Three separate articles. Same underlying diagnosis: the enterprise has a structural trust deficit that manifests as underperforming assets, whether those assets are AI agents, senior leaders, or institutional influence.

Brand Assets Suffer From the Same Structural Failure

Dulux has one of the most recognizable brand mascots in the UK. The Old English Sheepdog appears in roughly 73% of brand awareness metrics for the paint category. Dulux's head of brand Sam Balloch recently acknowledged to Marketing Week that the brand had "forgotten" the mascot's potential.

The dog was appearing at the end of advertisements as a closing badge. "She trots on just before the logo. But she wasn't intrinsic to the story." Kantar brand health tracking confirmed what the marketing team suspected: strong awareness was not converting to consideration because the brand lacked emotional meaning in consumers' purchase decisions.

The problem was not the mascot. It was not brand awareness. It was that Dulux had deployed a "powerhouse" asset at badge level - the equivalent of an AI agent running at the bottom of the autonomy ladder not because the governance infrastructure required it, but because no one had thought through how to integrate it more deeply into the actual story.

Dulux's response - the "Life is What You Paint It" campaign repositioning the dog as central to modern life moments rather than peripheral to product presentation - is structurally identical to moving an AI agent from assistive output to supervised actions. Same asset, better governance about where and how it appears in the narrative.

If you're evaluating brand assets or partnership strategies against these criteria, our analysis tools can help surface the underdeployment patterns that pitch decks and awareness metrics tend to obscure.

What Structural Trust Actually Looks Like

The common thread across all three domains is that trust is not a property of the asset. It's a property of the infrastructure surrounding the asset.

For AI agents, the HBR framework is specific: assign distinct credentials with narrowly scoped permissions, define authoritative information sources and treat external inputs as potential attack vectors, implement validation layers between AI recommendations and actual system execution, and establish clear accountability trails. The Replit agent failed not because of model capability but because none of these structures existed.

For leaders, Wedell-Wedellsborg's organizational moves create the equivalent infrastructure: stability anchors that give leaders reliable footholds when external conditions are chaotic, co-pilot structures that prevent both passive withdrawal and reactive overcontrol, and narrative frameworks that convert environmental volatility into manageable rather than paralyzing conditions.

For brand assets, the Dulux case shows that the infrastructure question is: where does this asset appear in the actual decision journey, not just the awareness funnel? Emotional resonance is not generated by a mascot. It's generated by a mascot deployed at the right structural moments with the right narrative integration.

This is why enterprises that treat AI agent deployment as primarily a technical problem consistently underperform those that treat it as a governance problem. The organizations that succeed aren't moving faster. They're building the accountability infrastructure that allows them to move with confidence at whatever speed their governance supports.

There's a practical audit worth running here. Take any asset your organization considers valuable but underperforming. Then ask four questions drawn from the HBR framework: Does this asset have clear identity and scope of permission? Do the people deploying it have access to authoritative, current information? Are there validation layers between its output and downstream consequences? And is there an accountability trail that lets you understand what happened after the fact? For AI agents these map to technical implementation. For leaders they map to organizational structure. For brand assets they map to deployment strategy. In every case, a "no" on any question is not a capability problem - it's a solvable governance gap.

The Convergence Point

Three pieces of published research - across AI strategy, organizational psychology, and brand marketing - landed on the same Monday with the same diagnosis. Most enterprises are sitting on underdeployed assets: AI agents that could run with more autonomy but aren't trusted to, senior leaders who could exert more influence but are withdrawing instead, brand assets with high awareness and low consideration conversion.

In each case, the bottleneck is not capability. It is the absence of governance infrastructure that makes capability deployable.

The companies getting this right - Klarna with bounded AI autonomy, Dulux rebuilding structural emotional integration, organizations implementing stability anchors for leadership - share a recognizable trait. They treat trust as infrastructure that gets built before deployment, not a problem to solve after something goes wrong.

Most enterprises discover this the way Replit discovered it: after the database is deleted, and the cover-up attempt makes the original failure worse.

The question worth asking is not whether your AI agents, leaders, or brand assets are capable enough. It is whether the governance structures exist to make that capability trustworthy. If you want to map that gap honestly, the right place to start is here.

Want more insights like this?

Follow along for weekly analysis on brand strategy, market dynamics, and the patterns that separate signal from noise.

Browse All Articles →

Or explore partnership opportunities with STI.

Related Articles