Forecast accuracy is the percentage difference between the revenue a sales org commits to (the forecast) and the revenue it actually books in a given period. Most B2B sales orgs track it weekly or monthly at the segment level and quarterly at the company level. Targets vary, but world-class is +/- 5% on the quarterly commit.
How it’s measured
Forecast Accuracy = 1 - |Actual − Forecast| / Forecast
A team that commits to $4M and books $3.8M is at 95% accuracy (5% miss). The same team committing to $4M and booking $4.5M is also at 88% accuracy — a beat is still a miss against accuracy, even if revenue is up.
Most orgs track several forecast snapshots:
- AE commit — what each rep commits to in their weekly 1:1
- Manager roll-up — what the front-line manager believes will close
- Forecast call — what the CRO commits to the board
Each level has its own accuracy. The interesting one is the manager roll-up vs actual — that’s where most of the inaccuracy lives.
Why it matters
Boards and investors compound off forecast. A team that misses by 15% twice in a row triggers re-planning, headcount freezes, and credibility loss with the CFO. A team that’s consistently within 5% earns the right to invest ahead of revenue.
How to improve it
- Stage definitions matter. If “Stage 4 — Verbal Yes” means different things to different reps, the forecast is noise. Document stage-entry criteria; train AEs on them; audit deals weekly.
- Use multiple forecast methods. Rep commit + AI-driven forecast (Gong, Clari) + historical conversion model. When all three agree, ship the forecast. When they diverge, investigate.
- Track conversion rates by stage by segment. A 50% Stage-4-to-Closed-Won rate company-wide hides a 25% rate in one segment and 75% in another. Segment-level conversion is the input to a credible forecast.
- Catch slipped deals quickly. A deal that misses its expected close date by 30 days is rarely going to close on the new date. Surface slips weekly; reset expectations.
- Use Gong (or Clari, or BoostUp) for AI-driven sanity checks. These tools score deal health from conversation + email + activity signals. Their forecast is rarely the right answer alone, but it’s a useful complement to rep commit.
Common pitfalls
- Sandbagging. Reps consistently commit low to over-deliver. Reads as accuracy but kills capacity planning. Compare commit to historical to detect.
- Hero ball. A rep commits one giant deal that always slips a quarter. Look at deal aging and individual rep accuracy patterns.
- Stale opportunities. Deals stuck in Stage 2 for 90+ days inflate pipeline and degrade forecast math. Auto-close stalled deals.
- Single-method forecasting. Rep commit alone is too noisy. AI-only is too disconnected from negotiation reality. Use both.
Related
- Pipeline velocity — the upstream metric that determines whether forecast can be hit
- Gong — the conversation intelligence layer for AI-driven forecast
- Salesforce — where forecast lives for $50M+ ARR teams