Intent engineering: the missing layer in your AI strategy
Klarna's AI agent handled 2.3 million customer conversations in its first month. Resolution times dropped from 11 minutes to two. The CEO projected $40 million in savings. Then customers started complaining. The AI worked brilliantly β and that was exactly the problem.
The story everyone misreads
The standard reading of Klarna's AI rollout is that AI cannot handle nuance. That was a comforting interpretation in early 2025. A more accurate one, looking back now, is that the agent was extraordinarily good at resolving tickets fast β and that was the wrong goal to give it.
Klarna's actual organizational intent was not "resolve tickets fast". It was "build lasting customer relationships in a highly competitive fintech market". Those are profoundly different objectives. A human agent with five years at the company knows the difference intuitively: when to bend a policy, when to spend an extra three minutes because the customer's tone signals they are about to leave. The AI agent had none of that. It had a prompt. It had context. It did not have intent.
By mid 2025, CEO Sebastian Siemiatkowski told Bloomberg that while cost was the predominant evaluation factor, the result was lower quality. Klarna began rehiring the human agents it had let go. The $60 million in cumulative savings had not been enough to cover the reputational damage from becoming the public example of AI that optimises for the measurable at the expense of the meaningful.
Three disciplines, one missing
The last three years of enterprise AI have produced two meaningful disciplines. Prompt engineering came first: individual, synchronous, session-based. You sit in front of a chat window, craft an instruction, iterate the output. It is a personal skill with personal value.
Context engineering followed. Anthropic defined it in September 2025 as the shift from crafting isolated instructions to designing the entire information state an AI system operates within. RAG pipelines, MCP servers, structured knowledge bases. It is where the industry is now, and it is genuinely important work.
But neither of these solves what Klarna ran into. Context engineering tells agents what to know. What comes next β intent engineering β tells agents what to want. It is the discipline of making organizational purpose machine-readable and machine-actionable: not as prose in a system prompt, but as structured parameters that shape how agents make decisions when running autonomously over days, weeks, or months.
What intent engineering actually means
OKRs were designed for people. They encode human-readable goals and assume human judgment about trade-offs, priorities, and exceptions. A manager can tell a direct report what matters this quarter and trust that the report will interpret that guidance through months of institutional context, professional norms, and tacit knowledge built over time.
Agents have none of that. An agent does not know your company's OKRs unless you put them in the context window. It does not know which trade-offs your leadership team prefers unless you encode those preferences in a way it can act on. It does not know when to escalate versus when to decide autonomously unless you define that boundary explicitly.
When a human joins a company, alignment happens through a hundred informal mechanisms: the wiki, the onboarding, Slack conversations, watching senior people handle ambiguous situations. None of that works for agents. Agents need explicit alignment, and they need it before they start working β not six months later.
This means organizations need to produce something most have never had to create: machine-readable expressions of organizational intent. Not "increase customer satisfaction" β that is a human-readable aspiration. An agent needs a structured definition: which signals indicate customer satisfaction in this specific context, which data sources contain those signals, which actions the agent is authorized to take, which trade-offs it can make, and where the hard limits are.
The three layers no one has built
Intent engineering operates across three infrastructure layers, each at a different altitude.
The first is unified context infrastructure β the systems that make organizational knowledge agent-accessible. Most teams build this ad hoc: one team pipes Slack through a custom RAG pipeline, another exports Google Docs into a vector store, a third built an MCP server connecting to Salesforce but not Jira, and a fourth does not know the other three exist. The Model Context Protocol, which Anthropic introduced in late 2024 and donated to the Linux Foundation in 2025, is the most promising attempt at standardisation β but protocol adoption is not organizational implementation. Which systems become agent-accessible? Who decides what context an agent can see across departments? How do you version knowledge so agents are not operating on stale information? A standard does not answer those questions.
The second layer is workflow alignment β a shared understanding of which processes are agent-ready, which require a human in the loop, and which remain human-only. Right now, every employee rolls their own AI workflow. One uses Claude for research and another tool for drafting. A third built a custom agent chain. A fourth pastes text into a chat window. None of them can describe their workflow in a way that is transferable or improvable. The difference between individual AI use and organizational AI leverage is the difference between one good hire and a system that makes everyone better.
The third layer is intent encoding proper: translating human-readable organizational objectives into agent-actionable parameters. Decision boundaries. Escalation thresholds. Value hierarchies for resolving trade-offs. Feedback loops that measure alignment drift over time. This is what almost no company has built. It is also what determines whether a long-running agent operates in line with what the organization actually needs β or optimises its way into a public relations problem.
The race is no longer about the model
Deloitte's 2026 state of AI in enterprise survey found that 84% of companies have not redesigned jobs around AI capabilities and only 21% have a mature model for agent governance. These are not technology gaps. They are intent gaps.
Meanwhile, 74% of companies globally report they have yet to see tangible value from AI. McKinsey found 30% of AI pilots failed to reach scaled impact. The investment numbers say something different: 57% of digital transformation budgets flowing into AI automation, averaging $700 million for companies with $13 billion in revenue. Large investment, mixed outcomes. The pattern is not that the models are weak. The models are not the bottleneck. A mediocre model with clear, structured, goal-aligned intent infrastructure will outperform a frontier model in an organization with fragmented, inaccessible, unaligned knowledge every time.
The AI race has been framed as an intelligence race for three years β who has the best model, the largest context window, the highest benchmark score. That framing made sense when models were the bottleneck. They are not the bottleneck any more. The race is now an intent race.
What this means in practice
The companies that get this right will treat context infrastructure the way they treated data warehouse strategy: as a core strategic investment, not an IT project. They will build a living organizational AI capability map β a shared understanding of which workflows are agent-ready, which are agent-augmented with human oversight, and which remain human-only. And they will develop goal translation infrastructure that converts human-readable organizational objectives into agent-actionable parameters.
For back office operations, the implication is direct. An AI agent running procurement, order management, or customer service without intent infrastructure does exactly what Klarna's agent did: optimises for the measurable at the expense of the meaningful. Resolution time goes down. Relationship cost goes up.
Building that intent layer is the work that determines whether AI automation delivers on its promises of cost reduction and scale β or simply moves cost from one column to another. The models work. The context pipelines are getting better. What remains is the organizational infrastructure that connects AI capability to organizational purpose.
That is not an engineering problem. It is a strategy problem that requires engineering to implement. And it is the gap that separates organizations running AI tools from organizations where AI operates as genuine infrastructure.