All posts
AI Adoption March 21, 2026 · 6 min read

How to Measure AI Adoption ROI (When Nobody Can Tell If It's Working)

By Dave O'Dell

I was on a call last month with a VP of Engineering at a mid-size fintech. They’d rolled out AI coding tools six months ago. When I asked how it was going, the answer was: “I think it’s going well? Some people like it.”

That’s a $200K annual investment being evaluated with vibes.

He’s not alone. 77% of organizations report they can’t effectively track ROI on their AI investments. Not “haven’t gotten around to it” — can’t. They don’t know what to measure, the data is messy, and every framework they’ve seen is built for ML model deployments, not developer tooling.

This is a solvable problem. You just have to stop measuring the things that don’t matter and start measuring the things that do.

What Everyone Measures (And Why It’s Wrong)

Tool Logins and “Active Users”

This is the metric vendors love because it always looks good. “You have 85% of seats showing activity this month!”

Activity is not adoption. Logging in is not using. I’ve seen dashboards where “active” means the engineer opened the tool once in 30 days — maybe by accident, maybe because IT auto-launched it, maybe because they were trying to figure out how to turn off the autocomplete that kept interrupting their flow.

If your measurement of AI ROI starts with “X% of seats are active,” you’re measuring your procurement success, not your adoption success.

Lines of Code Generated

This one’s seductive because it feels tangible. “Our AI tools generated 40,000 lines of code last month!”

Great. How much of it shipped to production? How much of it was deleted in code review? How much of it introduced subtle bugs that your team spent the next sprint tracking down?

Lines generated without lines accepted and deployed is a vanity metric. And even accepted lines can be misleading — if the AI generated 200 lines of boilerplate that a competent engineer would have abstracted into 20, you’ve actually made the codebase worse.

Time Saved (Self-Reported)

Survey your engineers and ask “how much time does AI save you per week?” You’ll get answers ranging from “zero” to “ten hours” and the average will be somewhere around three. Leadership will multiply that by headcount and hourly cost and produce a number that looks fantastic in a board deck.

The problem: self-reported time savings are wildly unreliable. People overestimate when they’re enthusiastic and underestimate when they’re skeptical. There’s social pressure to report that the expensive tool is working. And even accurate estimates don’t account for the time spent reviewing, correcting, and debugging AI output.

What Actually Matters: Four Metrics That Work

Here’s the framework we use. None of these require AI-specific tooling to measure — they use data you probably already have.

1. Cycle Time (From Commit to Deploy)

This is the single most important metric for AI adoption ROI, and most teams already track it through their CI/CD pipeline.

Why it works: If AI tools are genuinely accelerating your engineering team, cycle time should decrease. Not just “time writing code” — the whole cycle. Because if AI writes code faster but your CI pipeline takes 45 minutes and your code review queue is backed up, the net effect on delivery speed is negligible.

How to measure: Track median cycle time (commit to production deploy) monthly. Compare the 3-month trailing average before AI adoption to the current 3-month trailing average. Control for team size changes, major architectural work, and seasonal patterns.

What good looks like: A 15-25% reduction in cycle time within 6 months of meaningful AI adoption. If you’re not seeing movement, either adoption isn’t real or there are bottlenecks elsewhere in the pipeline eating the gains.

Watch out for: DORA metrics are showing something interesting right now — teams using AI are seeing a 91% increase in code review time and 154% increase in PR size. AI is shifting the bottleneck from writing to reviewing. If your cycle time isn’t improving despite AI usage, this is probably why.

2. Throughput (PRs Merged Per Engineer Per Week)

Raw throughput is a blunt instrument, but directionally useful.

Why it works: If engineers are using AI effectively, they should be completing more units of work. Not necessarily bigger PRs — in fact, smaller, more frequent PRs are a better signal. AI makes it easier to break work into smaller chunks because the overhead of setting up each change is lower.

How to measure: Track PRs merged per engineer per week, segmented by team. Look at the trend line, not individual weeks. Compare teams with high AI adoption to teams with low adoption.

What good looks like: A 20-40% increase in throughput for teams with genuine adoption. But only if quality metrics (next section) hold steady. Throughput without quality is just generating technical debt faster.

Important caveat: This is where the “two engineers doing the output of twenty” narrative can be misleading. It’s real in specific contexts — greenfield projects, well-documented codebases, teams with strong context infrastructure. But in legacy codebases with poor documentation, the gains are more modest. Don’t let anyone sell you a 10x productivity story without checking the conditions.

3. Defect Rate (Bugs Per Release)

This is the metric that separates real AI adoption from AI theater.

Why it works: If engineers are shipping faster but the defect rate is climbing, AI isn’t making them more productive — it’s making them faster at creating problems. The goal of AI-assisted development is to increase velocity while maintaining or improving quality. If you can’t demonstrate that, you don’t have ROI.

How to measure: Track defects per release (or per sprint) normalized by throughput. If you’re merging 30% more PRs, a 30% increase in absolute defect count means the rate is flat — which is acceptable. A decrease is excellent. An increase that outpaces throughput is a red flag.

What good looks like: Flat or improving defect rates alongside throughput increases. The research shows only 33% of developers fully trust AI-generated output — which is actually healthy. The teams that do best are the ones where engineers treat AI output as a first draft that needs review, not finished code.

The nuance: Early in adoption, defect rates sometimes spike as engineers learn to calibrate their trust in AI output. That’s normal. If it persists beyond 2-3 months, you have a training or process problem — your engineers aren’t reviewing AI output rigorously enough.

4. Adoption Depth (Weekly Active Usage by Tier)

This is the metric that replaces the useless “active users” number.

Why it works: It tells you how AI is being adopted, not just whether. 50% tool login rate tells you nothing. Knowing that 20% of your engineers use AI for code generation, 15% use it for code review, and 5% use it for architecture and planning — that tells you where you are on the adoption curve and what to focus on next.

How to measure: Define usage tiers based on depth of integration:

  • Tier 1 — Autocomplete: Engineer uses AI for inline suggestions. Low effort, low impact.
  • Tier 2 — Generation: Engineer uses AI to generate functions, tests, or boilerplate. Moderate effort, moderate impact.
  • Tier 3 — Workflow: Engineer uses AI across multiple steps — planning, coding, reviewing, debugging. High effort, high impact.

Survey monthly (keep it to 3 questions, not 30). Cross-reference with the quantitative metrics above.

What good looks like: Tier 1 is where most people start. You want to see movement from Tier 1 to Tier 2 within the first quarter, and Tier 2 to Tier 3 within two quarters. If people are stuck at Tier 1, they’re getting autocomplete value — which is real but modest. The big productivity gains come at Tier 3.

Putting It Together: The ROI Dashboard

Here’s what a real AI adoption ROI dashboard looks like:

MetricBaseline (Pre-AI)CurrentTrendTarget
Median Cycle Time4.2 days3.1 days↓ 26%< 3 days
PRs Merged / Engineer / Week3.14.4↑ 42%> 4.0
Defects Per Release2.82.5↓ 11%< 2.5
Tier 3 Adoption0%12%> 25%

That’s four numbers. You can put them on a single slide. Leadership can understand them. And they tell you something real about whether your AI investment is working.

When the Numbers Don’t Move

If you’re three months into AI adoption and these metrics are flat, here’s where to look — roughly in order:

1. Adoption is shallow. Most engineers are at Tier 1 (autocomplete) and haven’t gone deeper. This is usually a champion problem — you need peer-led enablement, not more training sessions.

2. The context layer is missing. AI tools are producing mediocre output because they don’t understand the codebase. Engineers try it, get bad results, and stop. Fix this with context infrastructure — architecture docs, conventions, patterns that AI tools can consume.

3. The bottleneck moved. AI accelerated code writing, but review, CI, and deployment are now the constraint. Your velocity engineering problem just shifted — and that’s actually progress, because now you know where to focus next.

4. You bought licenses without lift. You have the tools but none of the supporting infrastructure — no champions, no context, no pipeline improvements. This is the $100K shelfware problem and it’s the most common failure mode we see.

The Real ROI Question

The real ROI of AI adoption isn’t “how much time did we save?” It’s “can we ship more value to customers with the same team, at the same or better quality, without burning people out?”

That’s a harder question to answer, but the four metrics above get you close. Cycle time measures speed. Throughput measures volume. Defect rate measures quality. Adoption depth measures sustainability.

If all four are moving in the right direction, your AI investment is working — and you have the data to prove it. If they’re not, you know exactly where to focus.

Stop measuring logins. Start measuring outcomes.


Dave O’Dell is co-founder of App Vitals, where he and Dan McAulay help engineering teams build the measurement frameworks and infrastructure that turn AI tool investments into measurable velocity gains. If your ROI story is “I think it’s going well?” — we should talk.

Want to accelerate your engineering team?

Book a 30-minute discovery call to discuss your team's AI adoption strategy.

Get in Touch