The 39-Point Perception Gap: Why You Think AI Makes You Faster (When It Doesn't)
The 39-Point Perception Gap
You finished the feature in what felt like two hours. Maybe less. The AI agent scaffolded the boilerplate, generated the test stubs, even suggested the retry logic you would have written by hand. You merged the PR before lunch. Productive morning.
Except it was three hours and forty minutes. You know this because the time tracker was running. You checked it after lunch, expecting to see a number that matched the feeling in your chest - that light, fast, “I crushed it” feeling. The number didn’t match. Not even close.
You assumed the tracker was wrong. It wasn’t.
The Study That Broke the Narrative
In mid-2025, METR (Model Evaluation & Threat Research) published one of the most rigorous controlled studies ever conducted on AI-assisted development. Sixteen experienced open-source developers completed 246 tasks on their own repositories - projects they’d maintained for an average of five years. Each task was randomly assigned: AI tools allowed, or AI tools not allowed.
The developers used Cursor Pro with Claude 3.5 and 3.7 Sonnet. These aren’t toy setups. These are the tools most professional developers actually reach for.
The result:
When developers were allowed to use AI tools, they took 19% longer to complete issues. Not faster. Slower.
That alone would be noteworthy. But the second finding is what makes this study matter.
Before each task, developers predicted AI would speed them up by 24%. After completing the task - after experiencing the actual slowdown - they still believed AI had made them 20% faster. The perception gap between felt speed and measured speed was 39 percentage points.
Read that again. Developers who were measurably slower believed they were faster. Even after doing the work. Even after living through the experience.
Where the Time Goes
The METR researchers captured screen recordings. The data shows exactly where the hours evaporate.
AI-assisted sessions had more idle time. Not “waiting for the model” idle - actual no-activity-at-all idle. The likely explanation: coding with AI requires less cognitive effort moment to moment, making it easier to zone out or multitask. It feels effortless. Effort is how humans estimate time. Less effort, less perceived time. The clock disagrees.
But the real time sinks are more specific:
Reviewing AI suggestions. Every suggestion requires evaluation. Is this correct? Does it handle the edge case? Does it match the existing patterns in this codebase? Each evaluation is a micro-decision. Forty micro-decisions per hour doesn’t feel like work. It is work.
Debugging AI-generated code. The code looks right. The tests pass. But something is off at the boundary conditions, or the error handling is subtly wrong, or the AI hallucinated an API call that doesn’t exist. Finding these bugs takes longer than finding human bugs because the code has no “tells” - no personal style, no familiar mistake patterns to scan for.
Re-prompting after bad output. The first attempt misses the mark. You refine the prompt. The second attempt is better but introduces a new problem. You fix that and re-prompt. Three iterations later, you have code that works. You could have written it from scratch in the time it took to negotiate with the model.
Context-switching. This one is invisible. Every time you shift from “thinking about the problem” to “evaluating what the AI thinks about the problem,” you pay a cognitive switching cost. It’s small per instance. It compounds across a day. By 4 PM, you’re mentally exhausted but can’t explain why because you “barely wrote any code.”
The Enterprise Confirmation
If METR was an academic signal, Faros AI provided the enterprise-scale confirmation. Their study analyzed telemetry from over 10,000 developers across 1,255 engineering teams.
The individual numbers look great. Developers on high-AI-adoption teams merge 98% more pull requests. They touch 47% more PRs per day. They complete 21% more tasks.
The organizational numbers tell a different story. When Faros looked at DORA metrics - deployment frequency, lead time, change failure rate, mean time to recovery - there was no significant correlation between AI adoption and better delivery outcomes. Companies with heavy AI usage didn’t ship faster or more reliably than companies without it.
The gains at the individual level evaporated at the team level. Where did they go?
Into the review queue. PR review time increased 91%. PRs were 154% larger. Reviewers were drowning in AI-generated output that looked plausible but required careful inspection. Every hour a developer saved writing code became an hour (or more) someone else spent reviewing it.
This is Amdahl’s Law applied to your pipeline. Code generation got faster. Everything downstream - design review, code review, QA, integration, deployment - runs at the same speed. When one stage accelerates and the rest stays flat, you don’t get faster delivery. You get a pile-up at the next bottleneck.
Why You Can’t Trust the Feeling
The perception gap isn’t a fluke. It’s a well-documented cognitive phenomenon operating through multiple channels.
Effort heuristic. Humans estimate duration by effort. Hard things feel long. Easy things feel short. AI makes coding feel easier moment to moment, so you estimate you spent less time. The clock doesn’t care about your feelings.
Completion bias. AI helps you produce visible artifacts - files, functions, test stubs - quickly. The dopamine hit from “I finished something” arrives early and often. Your brain logs each hit as progress. More hits per hour = must be going faster. But progress and speed aren’t the same thing.
Variable ratio reinforcement. This is the slot machine mechanism in your IDE. Sometimes the AI nails the suggestion on the first try. Sometimes it takes five attempts. The intermittent reward pattern is the same one that keeps people pulling the lever in casinos. You remember the wins. You forget the re-prompts.
Anchoring. Before starting, you expected AI to make you faster. That expectation became an anchor. Post-task evaluation was biased toward confirming the anchor. Psychologists call this confirmation bias. Developers call it “intuition.” Here, it’s wrong.
The Stack Overflow 2025 Developer Survey found that trust in AI accuracy dropped to 29%, down from 40% the year before. But usage kept climbing. Developers are using the tools more while trusting them less. That’s not rational behavior. That’s a perception gap driving action.
The Metric That Lies to Your Manager
This problem doesn’t stay personal. It climbs the reporting chain.
Your manager sees more PRs merged. More commits per sprint. More story points closed. The dashboard looks great. The team adopted AI tools and “productivity” went up.
But the dashboard measures output volume. It doesn’t measure delivery outcomes. It doesn’t measure whether the features actually shipped to production faster. It doesn’t measure whether the code survived its first week in production without a hotfix.
InfoWorld called this “manufacturing liability” - producing code that creates more downstream work than it saves upstream. When PRs are 154% larger but review capacity hasn’t changed, you’re not moving faster. You’re just generating more material for the bottleneck.
The dangerous thing about the perception gap is that it’s self-reinforcing. Developer feels fast, reports feeling fast. Manager sees output metrics that confirm “fast.” Leadership doubles down on AI adoption. Nobody checks whether anything actually shipped sooner or better.
What to Measure Instead
The fix isn’t to stop using AI tools. The fix is to stop trusting feelings and start trusting measurements. Specifically, measurements of outcomes, not output.
DORA metrics. Deployment frequency. Lead time for changes. Change failure rate. Mean time to recovery. These measure whether software actually reaches users faster and more reliably. They’re resistant to inflation from AI because they track the whole pipeline, not just the code generation stage.
DX Core 4. A newer framework that adds effectiveness (developer experience), quality, and business impact to the speed dimension. The key insight: “diffs per engineer” is one metric out of four. Speed without quality, effectiveness, and impact is just motion.
Time-to-production, not time-to-PR. How long from “task started” to “running in production”? If AI cut your coding time by 40% but your time-to-production stayed flat, the gain evaporated somewhere in the pipeline. Find where.
Change failure rate. The Opsera 2026 benchmark found AI-generated PRs have a 32.7% acceptance rate versus 84.4% for human-written code. If your change failure rate is climbing post-AI adoption, you’re producing more but shipping less. That’s negative productivity.
Self-reported cognitive load. Ask yourself honestly, weekly: “How mentally drained am I at end of day, 1-10?” If the number is climbing while your “output” is climbing, you’re burning fuel faster, not running a better engine.
The 39 Points Between You and Reality
Here’s the uncomfortable truth. You probably read this article and thought: “Yeah, but I’m actually faster with AI. My situation is different.”
That’s the 39-point gap talking.
The developers in the METR study thought the same thing. They had five years of experience on these codebases. They were experts. They used top-tier tools. They still couldn’t tell they were slower.
This doesn’t mean AI tools are useless. It means the feeling of productivity is unreliable as a signal. And unreliable signals, left unchecked, drive bad decisions - for you, your team, and your organization.
The fix is boring. Measure outcomes. Track time honestly. Compare pre-AI and post-AI on metrics that matter (delivery speed, bug rate, time-to-production), not metrics that feel good (PRs merged, commits per day, story points).
And the next time you finish a feature and think “that was fast” - check the clock.
The perception gap is one of the six dimensions the OnTilt framework measures. It shows up in the “Loss of Control” and “Preoccupation” scales - patterns where your experience of a tool diverges from the tool’s actual impact on your work.
Take the Self-Check - 14 questions, 3 minutes, anonymous. It won’t tell you whether AI is making you faster. It’ll tell you whether you’d notice if it wasn’t.
Sources:
- METR. (2025). “Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity.” 16 developers, 246 tasks, randomized controlled trial. 19% slower with AI; 39-point perception gap. metr.org
- METR. (2026, February 24). “We are Changing our Developer Productivity Experiment Design.” Follow-up noting selection bias and recruitment challenges. metr.org
- Faros AI. (2026). “The AI Productivity Paradox Research Report.” 10,000+ developers, 1,255 teams. 98% more PRs merged, flat DORA metrics, 91% longer review times. faros.ai
- Opsera. (2026). “AI Coding Impact 2026 Benchmark Report.” AI PR acceptance rate 32.7% vs. human 84.4%; 4.6x longer review wait; 1.7x more bugs. opsera.ai
- Stack Overflow. (2025). Developer Survey. Trust in AI accuracy: 29% (down from 40%). stackoverflow.com
- Noda, A. & Tacho, L. (2025). “DX Core 4.” Unified developer productivity framework (speed, effectiveness, quality, impact). getdx.com
OnTilt is a research project studying behavioral patterns in AI-assisted work. The quiz is a self-check tool, not a diagnostic instrument. Read more on our About page.