
Claude Code Source Leak: What It Reveals About AI Agent Trust
.png)

Most teams adopt AI code review tools without baselines, making ROI impossible to measure. Here's a framework for tracking defect escape rates, review time, and the metrics that actually matter.
Your team just rolled out an AI code review tool. Two months later, someone asks: "Is it working?" Nobody can answer, because nobody measured anything before turning it on.
This happens constantly. Teams adopt AI review tools because the demos look compelling and the vendor quotes impressive accuracy numbers. But accuracy on a benchmark doesn't tell you whether the tool catches real bugs in your codebase, or whether your reviewers trust it enough to act on its suggestions. Without baselines, you're flying blind.
Here's a practical framework for measuring what actually matters, from defect escape rates to reviewer cognitive load, so you can tell whether your AI code review investment is paying off or just generating noise.
Most AI code review vendors will happily show you "comments generated per PR" or "lines of code scanned." These are vanity metrics. They measure activity, not outcomes. A tool that leaves 47 comments on every pull request isn't necessarily catching bugs. It might be drowning your reviewers in nitpicks about import ordering.
Four metrics actually tell you whether AI review is making your team better at shipping reliable software.
This is the one that matters most. Defect escape rate measures how many bugs make it past code review and into production (or staging, if that's your threshold). You calculate it as the number of post-release defects divided by total defects found, expressed as a percentage.
If your defect escape rate was 35% before AI review and drops to 22% after a month, that's a real signal. If it stays flat or climbs, the tool isn't catching the bugs that matter. Simple as that.
Track this by tagging bugs in your issue tracker with where they were discovered: in review, in QA, in staging, or in production. You probably should have been doing this already, AI tool or not.
One legitimate promise of AI code review is faster feedback loops. If a bot can flag obvious issues within seconds of a PR opening, the human reviewer spends less time on mechanical checks and can focus on architecture, logic, and design.
Measure the time from PR creation to the first substantive review comment (human or bot). Then separately track how long it takes for a human reviewer to leave their first comment. If AI review consistently provides useful initial feedback within minutes, you should see human reviewers starting their reviews faster too, because the easy stuff is already flagged.
Watch for the opposite effect, though. If AI comments are noisy, human reviewers may start ignoring the PR until the bot finishes spamming, which actually increases time to first human review.
This is the metric that determines whether your team will keep using the tool or silently start ignoring it. False positive ratio is the percentage of AI-generated review comments that get dismissed, resolved without action, or explicitly marked as unhelpful.
A false positive rate above 30-40% is a serious problem. Developers build habits quickly. If more than a third of AI comments are wrong or irrelevant, reviewers stop reading them. Once that trust erodes, even the tool's genuinely useful findings get ignored. You've created a boy-who-cried-wolf dynamic in your review workflow.
Track this by adding a simple thumbs-up/thumbs-down reaction convention on AI comments, or by monitoring which AI suggestions lead to code changes versus which get dismissed.
This one's harder to quantify but arguably just as important. Cognitive load is about how much mental effort your reviewers spend per PR. You can approximate it through a few proxy measurements: total review time per PR, number of review passes before approval, and the ratio of human comments to AI comments (are humans supplementing the AI's findings or duplicating them?).
The goal is for AI to reduce cognitive load by handling the mechanical checks (null pointer risks, missing error handling, unused imports) so humans can focus on higher-order concerns. If total review time per PR goes up after adopting AI review, something's wrong. Either the AI is creating more work by generating noise, or reviewers are spending time evaluating AI suggestions instead of reviewing the actual code.
A quick developer survey every quarter can fill in what the numbers miss. Ask reviewers directly: "Does the AI tool make reviews easier, harder, or about the same?" The qualitative signal matters.
The single biggest mistake teams make is skipping the baseline. You can't measure improvement without a starting point. And retrofitting a baseline after the tool is already running is unreliable at best.
Before enabling the AI review tool on any repos, spend two to four weeks collecting data on your current review process. Here's what to capture:
Collect this data for at least two full sprint cycles. One sprint isn't enough; you need to account for variation between feature work, bug fixes, and refactoring sprints.
AI code review vendors have a strong incentive to steer you toward metrics that make their tool look good. Here's how to tell the difference.
Vanity metrics measure tool activity: comments generated, lines scanned, PRs analyzed, suggestions offered. They go up and to the right no matter what. A tool that flags every single line of code as potentially problematic would score perfectly on these metrics while being completely useless.
Outcome metrics measure what changed in your engineering process: fewer bugs reaching production, faster review cycles, lower reviewer fatigue. These can go down, which is why vendors don't highlight them.
Here's a quick reference for which is which:
Here's a concrete timeline for measuring AI code review ROI in your org.
Pull existing metrics from your Git platform API, time-to-review, defect escape rates, bug discovery stages. Run a quick developer survey. One week of historical data plus API pulls is enough to establish a baseline.
Enable the AI tool on two or three repos, keep the rest as controls. Track the same baseline metrics plus false positive ratio and suggestion acceptance rate. Have devs flag AI comments as helpful/unhelpful with emoji reactions.
Compare pilot repos against baseline and control repos across your core metrics. Re-run the developer survey. You're looking for directionality, are things trending the right way? Two weeks of pilot data won't move defect escape rates dramatically, but you'll see signals in review speed, false positive rates, and developer sentiment.
Promising data? Roll out broadly and keep measuring. Ambiguous? Extend the pilot another two weeks. Clearly negative? Cut the tool. That's a valid, data-backed conclusion.
The patterns from teams that actually measured before and after tend to cluster around a few recurring themes.
AI review catches different bugs than humans. The most consistent finding is that AI tools are good at spotting mechanical issues: null reference risks, missing error handling, race conditions in well-known patterns, and security anti-patterns like SQL injection vectors. They're much weaker at catching logic errors, business rule violations, and architectural problems. This makes them genuinely complementary to human review when the false positive rate is manageable.
Time-to-first-review improves, but total merge time often doesn't. Developers get faster initial feedback, which is real value. But the total time from PR creation to merge often stays about the same, because the bottleneck was never "waiting for someone to spot a null check." It was waiting for a human with the right context to evaluate the overall approach.
The biggest ROI shows up in large, distributed teams. Teams spread across time zones see the most benefit, because AI review provides instant feedback during off-hours when human reviewers aren't available. A developer in Singapore opening a PR at 4 PM their time doesn't have to wait until their London colleagues wake up to get initial feedback. For co-located teams working the same hours, the speed advantage is less significant.
Configuration quality determines everything. Teams that invested time configuring their AI review tool (suppressing noisy rule categories, tuning severity thresholds, adding custom rules for their domain) saw dramatically better results than teams that turned it on with default settings. The out-of-box experience for most tools is mediocre at best. You should budget real engineering time for configuration and tuning, not just flip a switch and hope.
Not every team benefits from AI code review, and pretending otherwise doesn't help anyone. There are clear patterns where AI review becomes a net drag on your team.
High false positive rates erode trust. Once developers learn to ignore AI comments, they don't selectively start reading them again when the tool improves. The damage compounds. Teams that hit a 50%+ false positive rate in the first two weeks of a pilot tend to never recover engagement with the tool, even after tuning.
Alert fatigue cascades into human review quality. This is the insidious one. When a PR has 30 AI comments on it, human reviewers start skimming the whole PR, not just the AI comments. The cognitive overhead of distinguishing signal from noise contaminates the entire review. Counterintuitively, adding AI review can make human reviewers less thorough.
It creates a false sense of security. This one's the hardest to measure but the most dangerous. If your team starts thinking "the AI will catch it," they may unconsciously reduce their own review rigor. You might see this show up as shorter review sessions, fewer human comments per PR, or faster approvals. All of which look like efficiency gains until defect escape rate climbs.
GitClear's research found that code churn (lines reverted or substantially rewritten within two weeks) has been trending upward as AI coding tools see wider adoption. Their analysis of 211 million changed lines showed churn projected to double from its 2021 baseline. AI review tools are supposed to counteract this trend by catching low-quality code before it merges. If your churn rate isn't improving after adopting AI review, the tool may not be examining the right things.
Once you have baseline and post-rollout data, you can build an actual ROI calculation. The formula isn't complicated, but it requires honest inputs.
On the cost side, include the tool's license fees, engineering time spent on initial configuration and ongoing tuning, any increase in CI minutes from the tool running on every PR, and the cognitive cost of false positives (estimated as: false positive rate × average time to evaluate a comment × number of comments per week).
On the benefit side, estimate the value of bugs caught before production. One common approach: multiply the number of escaped defects prevented by the average cost of a production incident at your org. If your average production bug takes 4 hours of engineering time to diagnose, fix, and deploy, and the AI tool prevents 3 such bugs per month, that's 12 hours of engineering time saved. Compare that against the total cost of the tool.
Don't forget the second-order costs. If the tool generates enough false positives that reviewers spend 20 minutes per day evaluating and dismissing AI comments, that's about 7 hours per reviewer per month. For a team of eight, that's 56 hours. If the tool's license costs $500/month but it costs your team 56 hours of engineering time in false positive overhead, the math doesn't work regardless of how many bugs it catches.
The whole point of this framework is to replace gut feelings with data. After 12 weeks of structured measurement, you should be able to answer three questions:
If you get a clear yes on all three, expand the rollout. If it's mixed, invest in tuning before expanding. And if the tool fails on two or more, drop it. The worst outcome is paying for a tool that makes your reviews worse while everyone assumes it's helping because nobody's looking at the data.

Claude Code Source Leak: What It Reveals About AI Agent Trust

Reviewing AI-Generated Code: Different PRs Need Different Patterns
.png)
What to Look For When Reviewing Agent-Generated Code
Get Tenki
Change 1 line of YAML for faster runners. Install a GitHub App for AI code reviews. No credit card, no contract. Takes about 2 minutes.