Attribution models assign credit to touchpoints based on position, timing, or statistical patterns, but presence in a journey does not prove causation. A retargeting ad shown to someone who was already about to convert gets full credit under last-touch attribution even if the conversion would have happened without it.
Incrementality testing measures causal contribution: if this channel were turned off, how many conversions would be lost? For channels where models disagree significantly, incrementality testing provides the closest available ground truth.
Holdout tests
The most direct form of incrementality testing. Randomly split your audience into two groups:
- Exposed group (90%): Sees ads on the channel being tested.
- Holdout group (10%): Does not see ads on the channel. They may see a public service announcement (PSA) or nothing.
Compare conversion rates between the two groups. The difference is the channel’s incremental contribution.
Exposed group conversion rate: 4.2%Holdout group conversion rate: 3.1%Incremental lift: 1.1 percentage points (26% relative lift)In this example, 3.1% of conversions would have happened anyway. The channel’s true incremental contribution is 1.1 percentage points — roughly a quarter of the conversions it would claim under last-touch attribution.
Design considerations
Sample size matters. The holdout group needs to be large enough to detect the expected lift with statistical significance. For a channel driving 5% conversion rate with expected 20% lift, you need roughly 5,000-10,000 users per group. Under-powered tests produce noisy results that don’t resolve the question.
Duration matters. Run the test for at least one full purchase cycle. If your average time from first touch to conversion is 14 days, a 7-day test will miss conversions that the holdout design delayed rather than prevented. Two to four weeks is typical for most B2C products; B2B may need 8-12 weeks.
Contamination risk. Users in the holdout group may still be exposed to the channel through shared devices, cross-device behavior, or organic encounters with the brand. This “leakage” biases results toward zero, making the channel look less incremental than it actually is. Acknowledge this limitation when reporting results.
Opportunity cost. Suppressing ads to 10% of your audience means losing potential conversions during the test period. Size the holdout to balance statistical power against revenue risk. For high-spend channels, even a small holdout can be expensive, which is why you target tests at channels where the disagreement score is highest — those are the channels where the information value justifies the cost.
Geo tests
For channels that can’t be user-targeted — TV, radio, billboards, podcast sponsorships — or when user-level holdouts aren’t technically feasible, run the test geographically.
Setup. Select matched market pairs: cities or regions with similar demographics, purchase patterns, and historical performance. Turn off (or scale up) the channel in the test markets. Keep the control markets unchanged.
Test markets (channel off): Austin, Portland, NashvilleControl markets (no change): Denver, Raleigh, Salt Lake CityDuration: 4 weeks on, 2 weeks post-periodCompare conversion rates, revenue, or whatever KPI you’re measuring between test and control markets. The difference, adjusted for any pre-existing trends, estimates the channel’s incremental impact.
Matching markets correctly
Poor market matching is the most common failure mode. Markets should be similar on:
- Baseline conversion rate (within 10% historical variance)
- Seasonality patterns (both markets peak and trough at the same times)
- Population demographics relevant to your product
- Competitive landscape (a competitor launching in one market but not the other confounds results)
Use at least 4 weeks of pre-test data to validate that your matched markets track closely. If they diverge significantly before the test starts, the pairing is wrong.
Geo test limitations
Geographic tests are blunter instruments than user-level holdouts. They can’t control for spillover effects (someone in a test market sees a TV ad in a control market), and the sample sizes are inherently smaller (markets, not users). They work best for answering big-picture questions: “Does TV drive meaningful incremental lift?” rather than “What’s the precise incremental CPA of our podcast sponsorship?”
Platform lift studies
Meta (Conversion Lift), Google (Conversion Lift, Brand Lift), and TikTok all offer built-in experimentation tools that handle the holdout mechanics for you. The platform randomly suppresses ads to a control group and measures the difference.
Advantages:
- No engineering work to set up holdout groups
- Platform handles randomization and measurement
- Results include confidence intervals and statistical significance
- Some platforms offer cross-device tracking within their ecosystem
Tradeoffs:
- You’re trusting the platform to measure its own channel’s effectiveness — the same incentive problem that makes platform attribution biased applies here, albeit to a lesser degree
- The platform can only measure lift within its own ecosystem — it can’t tell you about cross-channel interactions
- Lift study availability and quality vary by platform; smaller platforms may not offer them
- Results are aggregate, not user-level, limiting the depth of post-analysis
Platform lift studies are most useful as directional validation. If Meta’s Conversion Lift says their ads drive 15% incremental lift and your own holdout test shows 12%, you have reasonable convergence. If Meta says 40% and your test shows 5%, investigate the discrepancy.
Using incrementality to calibrate attribution
Incrementality results don’t replace attribution models — they calibrate them. The workflow:
- Run attribution models in parallel using the comparison pattern.
- Identify channels with high disagreement scores — these are where incrementality testing has the highest information value.
- Run incrementality tests on the high-disagreement channels.
- Compare the incremental results to each attribution model’s estimate.
- Adjust interpretation (not the model itself) based on what you learn.
For example: Markov chain attribution says email drives 15% of conversions. A holdout test shows 8% incremental lift. This doesn’t mean the Markov model is wrong — it means email is present in 15% of converting journeys but only causally responsible for 8% of them. The remaining 7% would have converted anyway; email was part of the journey but not the catalyst.
This calibration is qualitative, not mechanical. The Markov model’s weights are not adjusted by a correction factor. Instead, teams develop channel-by-channel intuition: “Our attribution models tend to over-credit email by roughly 2x compared to incrementality. When the model says email drives $100K, the incremental value is probably closer to $50K.”
This intuition makes attribution outputs more useful over time, even without continuous incrementality tests. The goal is to know which channels a given model over- and under-credits, and apply that knowledge to budget decisions.
Building an incrementality testing program
Most teams can’t afford to test every channel continuously. A practical testing cadence:
Quarterly. Test the 2-3 channels with the highest disagreement scores or the largest budget allocations. Rotate through your channel mix over the course of a year.
Event-driven. Retest a channel when:
- You’re about to make a major budget change (scaling a channel 2x or cutting it significantly)
- Your attribution models show a significant shift in a channel’s contribution
- A platform changes its targeting or measurement methodology (iOS privacy changes, cookie deprecation phases)
Annual. Re-run your highest-spend channel’s test even if disagreement is low. Market conditions change, and a channel that was genuinely incremental last year may have saturated.
What incrementality can’t tell you
Incrementality testing has its own blind spots:
It measures short-term lift, not long-term brand building. A 4-week holdout test won’t capture the cumulative effect of brand advertising over months or years. Channels that build awareness slowly (content marketing, podcast sponsorships, community engagement) will under-perform in short incrementality tests even if they’re genuinely valuable over longer time horizons.
It’s expensive. Every holdout user is a potential lost conversion. High-spend channels can cost thousands of dollars per test in suppressed conversions. This is why targeting tests at high-disagreement channels matters — you want the highest information value per dollar of opportunity cost.
It assumes the rest of your marketing mix stays constant. If you’re running a holdout test on display ads while simultaneously scaling paid search, the interaction effects make it harder to isolate display’s true contribution. Minimize concurrent channel changes during test periods.
It doesn’t scale to all channels simultaneously. Testing channel A while everything else runs normally tells you about channel A in the context of your current mix. It doesn’t tell you about optimal allocation across all channels. For that, you need media mix modeling (MMM), which is a different discipline entirely.
Despite these limitations, incrementality testing remains the closest thing to ground truth in marketing measurement. Attribution models measure presence in a converting journey. Incrementality testing measures causal contribution.