Pricing Experiments That Don't Break Your Forecast
B2B pricing is the highest-leverage lever in the GTM stack and the one most companies are most afraid to touch. The fear is rational: pricing changes ripple into pipeline, sales rep comp, customer renewals, win/loss talk tracks and finance's forecast simultaneously. Done badly — as a hard cutover with a new price book on day one — they can take two full quarters off plan. But the opposite extreme, never adjusting pricing, is more expensive over a five-year window than almost any other strategic mistake. The middle path is structured pricing experiments that produce real evidence without putting the forecast at risk.
Why most pricing changes break forecasts
The classic mistake is shipping a single all-up pricing change — new tiers, new packaging, new discount discipline, new floor — on a single date. Three things go wrong simultaneously. Reps freeze on every open opp because they don't yet know how to position the new structure. Marketing's lead-to-MQL conversion drops because the public pricing page no longer matches the inbound copy. Finance's forecast model breaks because the historical conversion rates are now invalid.
The fix isn't to change less; it's to change in a sequence that lets you measure each variable independently and keep the forecast honest while you do.
The four-phase pricing experiment cycle
Phase 1: Hold the price, change the pitch
Before any number on the price book moves, run a four-week experiment where reps test new value framing on inbound deals while quoting the existing price. You're measuring two things: whether the new framing changes win rate and whether it changes the price the buyer was anchoring on (look at discount asks). Most pricing problems are actually framing problems — buyers anchor on a low price because the value story doesn't justify a higher one. You'll learn whether the next phase needs to touch price, packaging or both.
Phase 2: Test packaging changes on a holdout cohort
Once framing is steady, introduce new packaging (tier structure, what's in/out of each tier) to a subset of new deals only — typically inbound mid-market, where deal size variance is lowest and you have the most weekly volume. Existing pipeline and existing customers stay on the old packaging. Run for 6-8 weeks minimum. Measure deal size, win rate, sales cycle and feature-tier mix against a control cohort on the old packaging. The fastest way to break this experiment is to allow reps to "mix and match" — the cohorts have to stay clean.
Phase 3: Test price changes on a separate holdout
Only after packaging is validated do you test list-price changes, again on a holdout cohort. The reason to keep these phases separate is forecast hygiene: if you change packaging and price simultaneously and win rate moves, you can't attribute the change. Most pricing experiments go wrong here because the team is impatient to ship the full change. The cost of patience is one quarter of slower learning. The cost of impatience is two quarters of broken forecasts and a team that loses confidence in the model.
Phase 4: Roll out, with a forecast bridge
Once both packaging and price are validated, the rollout to 100% of new business needs an explicit forecast bridge. RevOps rebuilds the forecast model in two columns: the old structure (so historical conversion rates apply) and the new structure (with conversion rates from the experiment cohorts). For one full quarter, both columns get reported. This is the only thing that prevents the "pricing change broke the forecast" narrative — the forecast was never relying on stale rates in the first place.
Where to draw the experiment boundary
Three rules keep pricing experiments from contaminating the wider business:
- New business only, never existing customers during the experiment phase. Touching renewal pricing in an experiment is a fast way to spike churn and learn nothing.
- Single segment per experiment. Run the experiment in one segment (mid-market inbound is usually the cleanest) and hold every other segment constant. Cross-segment learnings come later.
- Comp neutrality. Reps in the experiment cohort should be paid on the same OTE structure as the control cohort. Comp changes during a pricing experiment confound the data.
The metrics that tell you the experiment is working
Win rate is the headline metric, but it lags. Three leading indicators tell you faster whether the new pricing is landing:
- Discount-ask rate. The percentage of deals where the buyer asks for a discount. A drop here means the new pricing/packaging is closer to the buyer's value perception.
- Time-to-quote. Reps quote faster when the packaging fits the buyer; long quote latency usually means reps are mentally translating between the new structure and what the buyer actually wants.
- Tier mix. If 80% of the cohort lands on the lowest tier, the packaging hasn't created enough differentiation between tiers — almost always a packaging problem, not a price problem.
What pricing experiments rarely fix
Pricing isn't a substitute for ICP clarity or product value. Two patterns we see repeatedly: a company runs a pricing experiment hoping to fix slow growth, and the experiment succeeds in lifting deal size by 12% but gross retention drops 4 points 12 months later because the higher price attracted buyers further from the real ICP. Or a company raises prices to fund a CS investment, and NRR doesn't move because the underlying expansion motion was the constraint, not the budget. Pricing amplifies whatever GTM motion you already have. It doesn't fix a broken one (related: the real cost of an undefined ICP).
How long pricing experiments need to run
The most common reason pricing experiments produce ambiguous results is that they're called too early. Pricing decisions ripple slowly: the buyers who would have churned at the new price don't churn for 6-12 months, the deals that would have upgraded don't upgrade for 9-15 months. The minimum honest runtime for a pricing experiment is one full sales cycle plus 90 days of post-sale observation. For most B2B motions that's 4-6 months. Teams that call experiments at 60 days are usually measuring the noise of the rollout, not the signal of the change. The discipline is patience — and that patience is much easier to hold when the experiment is structured cleanly enough that interim data is genuinely informative rather than misleading.
The role of finance in pricing experiments
Pricing experiments fail without finance at the table from day one. Finance owns the model that turns pricing changes into forecast changes, and a pricing experiment that lands without finance's involvement is one that breaks the forecast the next quarter. The right pattern: RevOps designs the experiment, the GTM leader owns execution, and finance owns the bridge between the experiment cohort's results and the company forecast. When all three functions are aligned on the experiment design before week one, the rollout in phase four is essentially mechanical. When they're not, every interim result becomes an argument about methodology rather than a signal about the change.
Where to start this week
Pull the last 200 closed-won deals and look at two numbers: the percentage that received a discount over 10%, and the median discount size. If more than 40% of deals are discounted over 10%, your pricing isn't matching buyer value perception — and the right first experiment is framing (Phase 1), not a price change. Start there before touching the price book.
Pricing & Packaging is one of eight pillars in the GTM Diagnostic. The full methodology covers how we score pricing discipline, packaging clarity and discount governance — three of the most reliable predictors of margin durability we measure.