
Intent-Driven Engineering and the Myth of “AI Friction Metrics”
- Mark Kendall
- 5 hours ago
- 2 min read
Intent-Driven Engineering and the Myth of “AI Friction Metrics”
Introduction
Recently, a new idea has started circulating in AI discussions: measuring system quality through concepts like “WTF moments per week.”
At first glance, it sounds clever—even relatable. But beneath the surface, it exposes something deeper:
A misunderstanding of how modern AI systems actually work.
If we’re serious about building enterprise-grade AI systems, we need to move beyond reaction-based metrics and start focusing on intent clarity, system design, and user responsibility.
What Is Intent-Driven Engineering?
Intent-Driven Engineering is the practice of designing systems where outcomes are governed by clearly defined intent—not trial-and-error interaction.
Instead of relying on users to “figure it out,” the system is structured so that:
Intent is explicitly defined upfront
Context is preserved and reusable
Behavior is predictable and repeatable
Outputs align with expectations by design—not chance
In this model, AI is not guessing.
It is executing.
The Problem with “Friction-Based Metrics”
Metrics like “WTF moments per week” attempt to measure user frustration.
But frustration is not a root cause—it’s a symptom.
Here’s the issue:
A “WTF moment” could come from unclear prompts
It could come from missing context
It could come from poor system design
Or simply from a user not understanding how to interact with AI
Lumping all of that into a single emotional metric doesn’t improve the system—it obscures the real problem.
Worse, it shifts responsibility away from engineering discipline and into vague user experience complaints.
Where the Responsibility Actually Lies
In advanced AI systems—especially when using tools like structured prompting, agents, or orchestration layers—the majority of failures are not model failures.
They are intent failures.
If the system produces an unexpected result, we should ask:
Was the intent clearly defined?
Was the context complete and structured?
Was the system designed to guide correct usage?
Because when intent is correct, outcomes stabilize.
When intent is vague, variability increases—and frustration follows.
From “Guessing” to “Execution”
The industry is still transitioning from:
Prompting AI → hoping for a good result
to:
Engineering intent → guaranteeing a result
This is the shift that separates:
Hobbyist usage from enterprise systems
Experimentation from production reliability
Friction from flow
In a properly designed system:
There are no “WTF moments.”
There are only signals that something in the intent layer needs refinement.
Why This Matters
If we normalize vague metrics like frustration counts, we risk:
Designing systems around emotion instead of structure
Blaming tools instead of improving engineering practices
Slowing down adoption of truly reliable AI systems
But if we focus on intent:
Systems become predictable
Teams move faster
Users gain confidence instead of frustration
And most importantly:
We move from reacting to problems…
to eliminating them at the source.
Key Takeaways
“WTF moments” are symptoms—not useful engineering metrics
Most AI failures are caused by unclear or incomplete intent
Intent-Driven Engineering replaces trial-and-error with structured execution
The future of AI is not better prompts—it’s better system design
If you’re experiencing friction with AI, the answer isn’t to measure frustration.
It’s to refine intent.
Comments