← All thinking pages

The Velocity Trap

working Last updated 2025-12-28
CS Operations Metrics

The Velocity Trap is what happens when a CS organization mistakes activity volume for intervention accuracy. So the team ramps up touches, runs a tighter cadence, gets QBR completion rates where they should be, and the dashboard turns green while net retention stays flat.

I keep coming back to this pattern because the macro data makes it hard to ignore. Benchmarkit's 2025 SaaS Performance Metrics report shows median NRR declining from 105% in 2021 to 101% in 2024. GRR slid from 90% to 88% over the same period. And this is happening while CS teams have access to more proactive tooling than at any point in the industry's history. You end up with health scores feeding automated cadences feeding AI-generated playbooks, and the whole apparatus looks like progress. Capability keeps going up while results keep going down, and nobody can quite explain the gap. I'm not claiming the tooling caused the decline, but it certainly didn't prevent it, and that's worth sitting with for a second.

Median SaaS Retention Metrics, 2021 vs. 2024
Source: Benchmarkit 2025 SaaS Performance Metrics. NRR and GRR both declining despite increased CS tooling investment.

When a team increases velocity without improving targeting, they're spending more time on accounts they can't influence. A CSM builds a two-hour QBR deck for a customer who was already in procurement with a competitor, and nobody told them because the signal was in a Slack channel CS doesn't have access to. A health score fires a playbook against an account whose real problem sits upstream of your product. The CRM says every intervention was completed. The customer still churns.

Bain's "Customer Success at a Crossroads" report puts a number on it: CSMs spend more than half their time on low-value, repetitive tasks. Over 50% of a fully-loaded CSM's hours. Depending on market, that's $75-120K base plus benefits and tooling overhead, call it $55-75/hour fully loaded. Run that across a team of 8 and you're looking at roughly $880K-$1.2M annually pointed at activities that aren't moving anything.

What I keep coming back to is that the misses compound, and that's where the math turns ugly.

There's a finite number of interventions the team can run in a quarter, and every miss burns one. The mid-market account showing expansion signals hasn't heard from anyone because the team is buried in save attempts for accounts that were never saveable. When I've tagged intervention outcomes across client portfolios, hit rates land around 15% (more on that at the Outcome Hit Rate page). Eighty-five percent of CS activity is non-productive, interventions that consumed resources without influencing retention.

If you run the numbers: 200 interventions per quarter at 15% accuracy generates 30 meaningful outcomes. Cut to 120 interventions at 40% accuracy and you get 48, which is a 60% improvement in impact from a 40% reduction in volume. It's worth running the numbers here because they suggest the problem isn't volume.

Intervention Volume vs. Meaningful Outcomes
Fewer, better-targeted interventions produce more outcomes. The 40% reduction in volume yields a 60% increase in impact.

Staircase AI's research found that customers who receive regular, well-executed QBRs are twice as likely to renew. But the word "well-executed" is load-bearing there. A QBR aimed at an account you can't influence isn't just wasted. It's an opportunity cost against a QBR that could have moved something.

I've been thinking about how AI fits into this picture, and I'm increasingly concerned it makes the Velocity Trap cheaper to run without making it easier to detect. Not that AI will replace CSMs, but that it could make the whole pattern so inexpensive that nobody notices they're in it. Every CS vendor is demoing systems that fire playbooks faster than your team can read them. When you run the unit economics on poor targeting, the numbers get hard to ignore. Why qualify the intervention when you can blast everyone? ...until the retention number comes due.

Klarna is the case study. They claimed their AI assistant was doing the work of 700 agents, handling 75% of customer chats (2.3 million conversations). Headcount fell from 5,527 to 3,422. Then quality dropped, and now they're hiring humans again. Take the Klarna example. They saw a lot of people doing a lot of activity and concluded the constraint was velocity, when the actual constraint was targeting accuracy. Gartner predicts half the companies who cut service staff for AI will rehire by 2027, which seems right to me.

Meanwhile, roughly 70% of CS organizations haven't scaled AI beyond experimentation (also Bain). Most are layering automation onto legacy workflows, chasing micro-efficiencies. The few organizations I've seen get it right are doing something different. They're using efficiency gains to be more selective, not more prolific.

The problem would be manageable if it stayed flat. It doesn't. The direct cost of a miss is the time. The indirect cost is that it pulls capacity and focus away from the next intervention, which is now more likely to miss too. A CSM running six low-probability save attempts per week has less preparation time, less mental bandwidth, and less strategic clarity for the two accounts where the outcome is genuinely in play. And the degradation doesn't stay proportional. Each miss makes the next one more likely. The question I keep arriving at is whether velocity is better understood as a multiplier than a strategy. If the targeting is broken, more speed just compounds the waste, and I haven't seen a case where that dynamic resolved itself without explicitly addressing the targeting layer first.

If I had to point to one diagnostic, it's Outcome Hit Rate. And the structural change that's moved the needle most consistently is the Milestone-to-Intervention Model, which replaces calendar-based triggers with value-milestone triggers so each intervention lands on an account at a moment when the outcome can actually change.

Dillon Young is the founder of Customer Value Labs, where he builds and maintains revenue systems infrastructure for B2B SaaS teams. If your CS team is running hard and the retention number isn't moving, a conversation might help.