Interpreting Lift in A/B Testing: Absolute vs Relative Changes and Business Impact

Capítulo 5

Estimated reading time: 7 minutes

+ Exercise

1) Absolute lift vs relative lift (and how each can mislead)

Lift is the change in a metric between treatment (B) and control (A). You will commonly see it expressed in two ways:

  • Absolute lift (difference): Δ = p_B − p_A (for rates) or Δ = x̄_B − x̄_A (for averages).
  • Relative lift (percent change): r = (p_B − p_A) / p_A (or (x̄_B − x̄_A)/x̄_A).

How absolute lift can mislead

Absolute lift can look “small” even when it is valuable at scale. Example: conversion rate goes from 10.00% to 10.30%. Absolute lift is only +0.30 percentage points, which may sound minor, but on millions of users it can be large.

How relative lift can mislead

Relative lift can look “huge” when the baseline is tiny. Example: conversion rate goes from 0.10% to 0.12%. Relative lift is (0.12−0.10)/0.10 = +20%, which sounds impressive, but the absolute lift is only +0.02 percentage points (2 extra conversions per 10,000 users).

When to prefer each

  • Use absolute lift when translating to counts, capacity, cost, or revenue (because business impact is usually linear in absolute changes).
  • Use relative lift when comparing improvements across metrics with different units or across markets with different baselines (but always show the baseline).
  • Best practice: report both, side-by-side, to prevent misinterpretation.

2) Baseline dependence: why small relative lifts can be large at scale

Lift is not interpretable without the baseline. Two experiments can have the same absolute lift but very different relative lift, and vice versa. Baseline matters because it determines:

  • Incremental count: incremental_conversions = traffic × (p_B − p_A)
  • Relative narrative: relative_lift = (p_B − p_A)/p_A

Step-by-step: from lift to incremental outcomes

Suppose your primary metric is a rate (e.g., sign-up rate).

Continue in our app.
  • Listen to the audio with the screen off.
  • Earn a certificate upon completion.
  • Over 5000 courses for you to explore!
Or continue reading below...
Download App

Download the app

  1. Write down the baseline: p_A

  2. Compute absolute lift: Δ = p_B − p_A

  3. Compute incremental outcomes for a time window: Δ_count = N × Δ, where N is the number of eligible users in that window.

  4. Sanity-check magnitude: express as “per 1,000 users” or “per day” to make it tangible.

Example: N = 2,000,000 monthly visitors, p_A = 8.0%, p_B = 8.2%. Absolute lift Δ = 0.2% (0.002). Incremental sign-ups: 2,000,000 × 0.002 = 4,000 additional sign-ups per month. The relative lift is only 0.002/0.08 = 2.5%, but the business impact can be substantial.

3) Converting metric changes into expected incremental value (funnel-based translation)

Product and marketing metrics often sit upstream of revenue. To translate lift into business outcomes, connect the experiment metric to downstream value using a simple funnel model. The goal is not perfect forecasting; it is a consistent, auditable mapping from metric movement to expected value.

A practical funnel translation template

Assume your experiment moves an upstream rate (e.g., sign-up rate). You can estimate incremental revenue as:

incremental_signups = N × (p_B − p_A)  # absolute lift on sign-up rate × traffic eligible for sign-up
incremental_payers = incremental_signups × activation_rate × pay_conversion_rate
incremental_revenue = incremental_payers × ARPPU  # or × LTV if appropriate

Where:

  • N = eligible traffic in the decision period (e.g., per month)
  • activation_rate = fraction of sign-ups who activate (e.g., complete onboarding)
  • pay_conversion_rate = fraction of activated users who pay
  • ARPPU = average revenue per paying user (or use contribution margin if you want profit impact)

Step-by-step example: sign-up lift to revenue

Inputs (choose a consistent time horizon, e.g., 30 days):

  • Monthly eligible visitors: N = 1,500,000
  • Control sign-up rate: p_A = 6.0%
  • Treatment sign-up rate: p_B = 6.3%
  • Activation rate: 40%
  • Pay conversion among activated: 10%
  • ARPPU (30-day): $50

Compute:

  • Absolute lift: Δ = 0.063 − 0.060 = 0.003 (0.3 percentage points)
  • Incremental sign-ups: 1,500,000 × 0.003 = 4,500
  • Incremental activated: 4,500 × 0.40 = 1,800
  • Incremental payers: 1,800 × 0.10 = 180
  • Incremental revenue: 180 × $50 = $9,000 per month (for this horizon)

Common pitfalls in value translation

  • Double counting: if your primary metric is already revenue, don’t also multiply by downstream rates.
  • Mixing horizons: don’t combine a 7-day conversion rate with a 12-month LTV without aligning assumptions.
  • Ignoring costs: if the variant increases incentives, support load, or infrastructure usage, translate those too (net impact).
  • Assuming funnel rates are constant: if the change affects user quality, activation and pay rates may shift; consider sensitivity ranges.

4) Heterogeneous effects: why average lift may hide wins/losses across segments

The overall (average) lift can mask meaningful differences across user groups. A variant can help one segment and harm another, producing a near-zero average—or a positive average that still creates unacceptable losses in a key segment.

How heterogeneity shows up

  • Different baselines: new users vs returning users often have different base rates, so the same absolute change implies different relative changes.
  • Different sensitivities: price-sensitive users may react differently than high-intent users.
  • Different constraints: mobile vs desktop can have different friction points; a UI change may help one and hurt the other.

Practical step-by-step: segment-aware interpretation

  1. Predefine key segments tied to product strategy (e.g., device, geo, acquisition channel, new vs returning). Avoid slicing into dozens of segments after the fact.

  2. For each segment, report: baseline (p_A), absolute lift (Δ), relative lift (r), and uncertainty (e.g., interval).

  3. Translate to business impact per segment: segment_incremental = N_segment × Δ_segment × value_per_outcome.

  4. Check for “offsetting effects”: a positive overall result may be driven by one large segment while another strategically important segment declines.

  5. Decide with guardrails: define non-negotiable constraints (e.g., do not reduce checkout conversion on mobile by more than X).

Illustrative example: same overall lift, different segment story

SegmentTraffic sharep_Ap_BAbsolute liftRelative lift
Desktop40%12.0%12.6%+0.6 pp+5.0%
Mobile60%6.0%5.7%−0.3 pp−5.0%

Depending on volumes and value per conversion, the net may be slightly positive or negative. Even if net is positive, a mobile decline might be unacceptable if mobile is the growth channel.

5) Practical guidelines for reporting lift (baseline, effect size, uncertainty, decision threshold)

A decision-ready experiment readout should make it hard to misinterpret lift. A useful reporting format includes:

  • Metric definition: what exactly is counted and over what window (e.g., “purchase within 7 days”).
  • Baseline: p_A (or x̄_A) with sample size.
  • Effect size: absolute lift Δ and relative lift r.
  • Uncertainty: an interval for Δ (and optionally for r) so readers see plausible ranges.
  • Business translation: incremental outcomes and incremental value for a standard period (e.g., per week/month), including key assumptions.
  • Decision threshold: the minimum effect worth shipping (e.g., “ship if expected incremental profit > $X/month and guardrails pass”).
  • Guardrails: secondary metrics that must not degrade beyond a threshold (e.g., refund rate, latency, unsubscribe rate).

A compact reporting table template

ItemControl (A)Treatment (B)LiftUncertaintyBusiness impact (assumptions)
Primary metric (rate)p_Ap_BΔ and rInterval for ΔN × Δ × $/outcome
Guardrail 1............Must be ≥ threshold

When uncertainty is wide, include a range for business impact too (e.g., “$3k to $15k/month”), computed by applying the interval endpoints for Δ to the translation model.

6) Examples: identical relative lifts, different business value due to base rates

Relative lift alone does not determine value. Base rate and volume determine incremental outcomes.

Example A: +10% relative lift on a high base rate

Checkout conversion improves from 20% to 22%:

  • Relative lift: (0.22−0.20)/0.20 = +10%
  • Absolute lift: +2 percentage points (0.02)
  • If N = 500,000 checkout sessions/month, incremental purchases: 500,000 × 0.02 = 10,000

Example B: +10% relative lift on a low base rate

Email click-through improves from 1.0% to 1.1%:

  • Relative lift: (0.011−0.010)/0.010 = +10%
  • Absolute lift: +0.1 percentage points (0.001)
  • If N = 500,000 emails/month, incremental clicks: 500,000 × 0.001 = 500

Both are “+10% lift,” but the absolute change (and therefore the count impact) differs by 20×.

Example C: same relative lift, different value per outcome

Two funnels each show +5% relative lift in their primary metric:

ScenarioMetricp_ARelative liftAbsolute liftValue per outcomeMonthly eligible NExpected incremental value
1Trial starts10%+5%+0.5 pp$8 per trial start (expected)2,000,0002,000,000×0.005×$8 = $80,000
2Support article views40%+5%+2.0 pp$0.05 per view (expected deflection)2,000,0002,000,000×0.02×$0.05 = $2,000

Even when relative lift is identical, the business value depends on (a) baseline, (b) volume, and (c) value per incremental outcome. This is why decision-making should anchor on absolute lift and translated value, while still reporting relative lift for context.

Now answer the exercise about the content:

Why is it recommended to report both absolute lift and relative lift in an A/B test readout?

You are right! Congratulations, now go to the next page

You missed! Try again.

Absolute lift maps directly to incremental outcomes (e.g., N × Δ) and value, while relative lift can help comparisons across different baselines. Reporting both side-by-side helps prevent misleading interpretations from either measure alone.

Next chapter

Confidence Intervals for A/B Testing: Quantifying Uncertainty in Differences

Arrow Right Icon
Free Ebook cover A/B Testing Essentials: Statistics for Product and Marketing
38%

A/B Testing Essentials: Statistics for Product and Marketing

New course

13 pages

Download the app to earn free Certification and listen to the courses in the background, even with the screen off.