Let’s start somewhere we can all agree.
Software teams want to build better products. Faster. With less friction. And if we’re honest, a little less pain along the way.
That’s exactly why tools like automation, and now AI, are so appealing. They promise leverage. They promise acceleration. They promise a kind of unfair advantage.
And sometimes… they actually deliver.
But here’s where things quietly go off the rails.
The Old Story, Rebranded
We’ve seen this movie before.
First, it was automation.
The narrative sounded familiar:
“If some automation is good… then more must be better.”
So we chased it.
We measured it.
We optimized for it.
And somewhere along the way, the goal quietly shifted:
From solving problems → to maximizing automation.
Now, we’re doing it again. With AI. Just faster. And louder.
When the Metric Becomes the Mission
Here’s the uncomfortable truth:
“Increase AI usage” is not a product goal. It’s a proxy. And a dangerously lazy one.
Because what happens next is predictable:
- Teams are pushed toward near-100% AI adoption
- Metrics become targets instead of signals
- Usage becomes performative rather than purposeful
And suddenly, you’re not asking:
“Did this improve the outcome?”
You’re asking:
“Did we use AI here?”
That’s not progress. That’s theater.
The Reality Nobody Wants to Say Out Loud
AI is powerful. But it’s not magic.
There are scenarios where:
- An experienced developer writes clearer, safer, more maintainable logic than AI ever could
- A tester spots edge cases AI doesn’t even know exist
- AI output, in inexperienced hands, becomes confidently wrong code at scale
And then there are scenarios where AI shines:
- Accelerating repetitive tasks
- Exploring solution spaces
- Enhancing—not replacing—expert judgment
So the real question isn’t:
“How much AI are we using?”
It’s:
“Where does AI actually improve this specific scenario?”
This Is Where Bridge Thinking Changes the Game
Bridge thinking doesn’t start with tools.
It starts with shared understanding:
- What problem are we solving?
- Who is affected?
- What does success actually look like in this scenario?
Only then do we ask:
“What’s the right combination of human expertise and AI support here?”
Scenario-Based Development (SBD) forces this discipline.
Instead of blanket adoption, we:
- Define the scenario clearly (actor, action, goal)
- Identify friction points
- Introduce AI where it meaningfully reduces that friction
- Evaluate outcomes—not activity
Now your KPI isn’t “AI usage rate.”
It becomes:
- Reduced cycle time in this scenario
- Improved defect detection in that workflow
- Higher confidence in delivery outcomes
That’s not anti-AI.
That’s pro-AI where it actually matters.
Metrics Should Follow Meaning—Not Replace It
Let’s be blunt.
If your KPI is “100% AI adoption,” you’ve already lost the plot.
Because:
- You’ll incentivize the wrong behavior
- You’ll reward usage over value
- And you’ll introduce a new layer of technical and product debt
Metrics should emerge after you understand what success looks like in a real scenario—not before.
Otherwise, you’re not measuring progress.
You’re manufacturing noise.
The Real Opportunity (That Most Teams Will Miss)
AI isn’t the problem.
Blind adoption is.
The teams that win won’t be the ones using AI everywhere.
They’ll be the ones who can say:
“Here’s exactly where AI creates value and here’s where it doesn’t.”
That level of clarity doesn’t come from dashboards.
It comes from thinking in scenarios.
It comes from bridging perspectives.
It comes from challenging the metric before optimizing for it.
Conclusion: Stop Worshipping the Tool. Start Solving the Problem.
Let’s call it what it is.
Chasing 100% AI adoption is just the latest version of a very old mistake:
Confusing activity with impact.
AI will absolutely change how we build software.
But it won’t replace the need for:
- judgment
- context
- experience
- and, most critically, shared understanding
If anything, it amplifies the cost of getting those wrong.
So here’s the shift:
Don’t ask how to use more AI.
Ask where AI earns its place.
Because in the end:
The best teams won’t be AI-first.
They’ll be problem-first and relentlessly precise about the tools they choose to solve them.

