Michael Zuo
March 13, 2026 · 5 min read

One Metric Is All You Need: Measuring AI Engineering Productivity

We've always used throughput to measure engineering productivity. To keep improving it, we built process metrics around the development workflow — optimize the process metrics, and throughput follows.

But in the age of AI coding, the workflow has fundamentally changed. The process metrics that once drove throughput are no longer the best leading indicators. So what's the new metric?

My answer is FPI.

Why FPI Matters

Here's the key insight: AI can run 24/7. It doesn't sleep, doesn't take PTO, doesn't context-switch between meetings. And you can scale it horizontally — more compute, more parallel agents, more throughput.

But only if it can work autonomously.

If every AI-generated feature requires a human to review, correct, or redo, then humans are still the bottleneck. Adding more AI agents just creates more work for the same engineers. You've scaled the cost without scaling the output.

High FPI changes the equation entirely. When AI can deliver features with minimal human intervention, productivity scales linearly with compute. Double the machines, double the output. That's not an incremental improvement — it's a fundamentally different cost curve. FPI is the metric that tells you how close you are to unlocking that leverage.

The Autonomous Driving Parallel

The self-driving car industry solved a version of this problem years ago with a metric called Miles Per Intervention (MPI): how far can the car drive before a human has to grab the wheel?

MPI doesn't measure speed. It doesn't measure distance. It measures trust — earned autonomy. A car with high MPI can handle more of the real world on its own. A car with low MPI is just an expensive cruise control.

Notice what MPI captures that "miles driven" does not:

  • A car that drives 10,000 miles but needs a human correction every 5 minutes is less capable than one that drives 1,000 miles with zero interventions.
  • More miles driven doesn't mean better driving. More miles between interventions does.
  • As MPI rises, the human role shifts from operating the vehicle to supervising the system.

This is exactly what's happening in your engineering org.

MPI and FPI parallel

Features Per Intervention (FPI)

FPI applies the same logic to AI coding:

FPI = Features Delivered / Human Interventions

A "feature" is a unit of shipped product value — whatever your team already counts as a deliverable (story, task, user-facing change). An "intervention" is any time a human has to step in — a correction, a rework, a takeover. Just like MPI counts every mile the same regardless of road type, FPI counts every intervention the same. Simple is better. What matters is the ratio and the trend.

FPI formula and progression

That's not just "more output." That's the AI earning trust — handling more of the work without human correction.

How to Read FPI

FPI trend interpretation

Rising FPI means your AI tooling is maturing. The team is learning to prompt better, the codebase is becoming more AI-friendly, and the AI itself is handling more complexity autonomously. You're building earned trust.

Flat FPI with rising output means you're scaling linearly — more features, but proportionally more hand-holding. You're getting volume, not leverage.

Falling FPI is a red flag. It means the AI is generating more work for humans, not less. Common causes: increased codebase complexity that the AI can't navigate, team members over-relying on AI for tasks it's bad at, or degraded prompt/context quality.

Questions to Ask Your Engineering Team

  1. "What's our FPI trend over the last three months?" — If they can't answer, you're flying blind.
  2. "Where are interventions happening most?" — Is it architecture decisions? Test failures? Integration bugs? The pattern tells you where AI tooling needs improvement.
  3. "How does FPI vary across teams?" — Team-level variance reveals which teams have figured out how to work with AI and which haven't. That's a coaching opportunity, not a headcount decision.

What to Do Tomorrow

Start tracking the data. The good news is you probably don't need to do it manually. AI coding tools already capture this — session logs, intervention points, completion rates. Tools like Claude Code's /insights can tell you exactly what happened in each coding session: what the AI handled autonomously and where a human had to step in. The raw data is already there. You just need to start reading it as FPI — and the trend will tell you everything you need to know.


Velocity told you how fast your team was moving. FPI tells you how much of that movement is earned — how much your AI tooling can handle on its own, and how much still needs a human hand on the wheel.

In the age of AI coding, the teams that win aren't the ones that generate the most code. They're the ones that intervene the least.