From Numbers to Insights: The “So What?” Test
Interpreting results means explaining why a metric moved, whether it matters for the business, and what you will do next. A useful habit is to translate every chart into a sentence that includes: metric + direction + magnitude + context + impact + next step. If you cannot name the impact (revenue, profit, qualified leads, retention), you may be looking at a vanity metric or an incomplete story.
Leading vs. Lagging Indicators (and How to Use Both)
What they are
- Leading indicators move earlier and can predict outcomes. They are useful for steering decisions quickly (e.g., qualified lead volume, add-to-cart rate, trial-to-activated rate, demo show rate, repeat purchase intent signals).
- Lagging indicators confirm outcomes after they happen. They are the scorecard (e.g., revenue, gross profit, net new customers, retention, LTV realized).
Vanity metrics often masquerade as leading indicators (e.g., impressions, followers). They can be “leading” only if you can show a stable relationship to business outcomes in your context.
Practical workflow: connect a leading indicator to a lagging outcome
- Pick one lagging outcome you care about this period (e.g., gross profit).
- List candidate leading indicators that plausibly drive it (e.g., qualified leads, product page engagement, demo bookings).
- Check the chain: can you trace a path from the leading metric to the outcome with intermediate steps? Example chain:
Qualified leads → demos held → opportunities → closed-won → revenue → gross profit. - Validate with data: compare periods where the leading metric changed and see if downstream steps moved in the same direction (allowing for time lag).
- Set interpretation rules: “If qualified leads rise but close rate falls, investigate lead quality and sales capacity before scaling spend.”
Example: impressions up, revenue flat
| Metric | Last week | This week | Change |
|---|---|---|---|
| Impressions | 500,000 | 900,000 | +80% |
| Clicks | 10,000 | 14,000 | +40% |
| Qualified leads | 420 | 380 | -10% |
| Revenue | $52,000 | $51,500 | -1% |
Bad interpretation (vanity-led): “Awareness is booming; the campaign is performing great.”
Better interpretation (impact-led): “Reach increased, but qualified leads fell 10% and revenue stayed flat. The additional traffic is likely lower-intent or mismatched to the offer. Next: break down by audience/placement and landing page to identify where lead quality dropped, then reallocate budget to segments with higher qualified-lead rate.”
Contextualizing Changes: Mix, Seasonality, and Promotions
Most “surprises” in reports come from missing context. Before you attribute a change to creative, targeting, or a new campaign, check three common drivers: traffic mix, seasonality, and promotions.
Continue in our app.
You can listen to the audiobook with the screen off, receive a free certificate for this course, and also have access to 5,000 other free online courses.
Or continue reading below...Download the app
1) Traffic mix: the hidden reason averages change
Overall conversion rate, CPA, or revenue per visitor can change simply because the share of traffic from each source changed—even if each source performed the same as before.
Step-by-step: diagnose a mix shift
- Compare channel shares for the two periods (e.g., % of sessions or clicks by channel).
- Compare channel performance (e.g., conversion rate, qualified-lead rate, revenue per session) for each channel.
- Decompose the change: ask “Did performance change within channels, or did the channel mix change?”
- Write the insight as two parts: (a) mix effect, (b) within-channel effect.
Example: conversion rate down, but nothing is “broken”
| Channel | Share (Week A) | Conv. rate (Week A) | Share (Week B) | Conv. rate (Week B) |
|---|---|---|---|---|
| Brand search | 30% | 6.0% | 20% | 6.1% |
| Non-brand search | 40% | 2.5% | 35% | 2.4% |
| Paid social | 20% | 1.2% | 35% | 1.1% |
| 10% | 5.5% | 10% | 5.6% |
Interpretation: Channel conversion rates are stable, but Week B has more paid social (lower intent) and less brand search (higher intent). The overall conversion rate fell mainly due to mix shift. Action: evaluate paid social using qualified leads and downstream conversion, and decide whether the mix shift is intentional (top-of-funnel push) or accidental (budget drift).
2) Seasonality: compare to the right baseline
Many businesses have predictable cycles: weekdays vs weekends, month-end effects, holidays, back-to-school, tax season, or industry events. If you compare the wrong periods, you can “discover” a problem that is actually normal.
Step-by-step: seasonality-aware comparisons
- Use like-for-like comparisons: week-over-week for weekly cycles, year-over-year for annual cycles, or compare to the same holiday window.
- Check calendar effects: number of days in period, paydays, shipping cutoffs, holidays.
- Use a rolling baseline (e.g., trailing 4-week average) to reduce noise.
- Annotate known events (product launches, outages, policy changes) so they are not misattributed.
Example: “Revenue dropped” vs “Revenue normalized”
If last week included a payday weekend and this week did not, a week-over-week revenue drop may be expected. A better statement is: “Revenue is down 8% WoW but is within 1% of the trailing 4-week average; the prior week was elevated due to payday timing.”
3) Promotions: separate price effects from demand effects
Promotions can inflate volume while hurting profit. Interpreting results requires separating: (a) more units sold, (b) lower price per unit, (c) changes in product mix, and (d) incremental vs pulled-forward demand.
Step-by-step: interpret a promo week
- Track profit-aware metrics: gross margin dollars, contribution margin, or profit per order (not just revenue).
- Compare AOV and margin rate: did revenue rise because more orders happened, or because order value changed?
- Check new vs returning customers: promos that mostly discount existing buyers may not create incremental growth.
- Look for pull-forward: does the week after the promo dip below baseline?
Example: revenue up, profit down
| Metric | Non-promo week | Promo week | Change |
|---|---|---|---|
| Orders | 1,000 | 1,400 | +40% |
| AOV | $80 | $70 | -12.5% |
| Revenue | $80,000 | $98,000 | +22.5% |
| Gross margin % | 45% | 30% | -15 pts |
| Gross margin $ | $36,000 | $29,400 | -18.3% |
Rewrite the analysis: “Promo increased orders and revenue, but gross margin dollars fell 18%. The discount likely shifted demand from full-price purchases and/or attracted lower-margin products. Next: segment by new vs returning customers and product category to see where margin erosion occurred, and test a smaller discount or threshold-based offer to protect margin.”
Cohort Thinking: See Retention and LTV Clearly
Aggregated averages can hide what is happening to different groups of customers. Cohorts are groups that share a starting point, such as first purchase month or acquisition channel. Cohort analysis helps you answer: “Are we acquiring customers who stick around and become profitable?”
Common cohort definitions
- First purchase month cohort: customers grouped by the month they first bought. Useful for retention and repeat revenue patterns over time.
- Acquisition channel cohort: customers grouped by the channel that acquired them (e.g., paid search vs paid social). Useful for comparing downstream quality.
- Campaign or offer cohort: customers acquired during a specific promo or with a specific discount.
Step-by-step: build a first purchase month cohort view
- Identify each customer’s first purchase date and assign a cohort month (e.g., 2025-10).
- Compute “months since first purchase” for each subsequent order (Month 0, Month 1, Month 2…).
- Choose a cohort metric: retention rate (customers who buy again), revenue per customer, gross margin per customer, or cumulative LTV.
- Create a cohort table where rows are cohorts and columns are months since first purchase.
- Interpret patterns: improving early retention often matters more than small changes in acquisition volume.
Example cohort table: retention rate by first purchase month
| First purchase cohort | Month 0 | Month 1 | Month 2 | Month 3 |
|---|---|---|---|---|
| Oct | 100% | 28% | 18% | 14% |
| Nov | 100% | 24% | 15% | — |
| Dec | 100% | 18% | — | — |
Interpretation: Newer cohorts show weaker Month 1 retention (Dec is 18% vs Oct 28%). If Dec acquisition relied more on heavy discounts or a new channel, you may be buying lower-quality customers. Next: split Dec by acquisition channel and offer type to find the driver.
Step-by-step: compare LTV by acquisition channel cohort
- Assign each customer an acquisition channel (the channel that first brought them in).
- Compute cumulative value per customer over time (e.g., 30/60/90-day revenue or margin).
- Compare curves: some channels start strong but flatten; others start slow but compound.
- Use a consistent time window for fairness (e.g., compare 90-day LTV for cohorts that have at least 90 days of data).
Example: paid social looks great upfront, but underperforms later
| Acquisition channel | Customers | 30-day revenue/customer | 90-day revenue/customer | 90-day gross margin/customer |
|---|---|---|---|---|
| Paid social | 1,200 | $42 | $55 | $14 |
| Paid search | 700 | $38 | $72 | $26 |
| Email capture (owned) | 500 | $35 | $80 | $32 |
Interpretation: Paid social acquires many customers and looks strong at 30 days, but its 90-day margin per customer is much lower. If you optimize using only early revenue or clicks, you may scale a channel that does not produce profitable repeat behavior. Next: adjust optimization to qualified leads or margin-based LTV, and test creative/landing pages that set better expectations to improve retention.
Vanity Metrics vs Business Impact: Rewrite the Analysis
Vanity metrics are not “useless”—they are incomplete. They become harmful when they replace impact metrics in decision-making. The fix is to rewrite your analysis so that it (1) ties activity to outcomes, (2) uses context and benchmarks, and (3) highlights what changed and what you will do.
Example 1: followers up, pipeline flat (B2B)
Vanity-led update: “We gained 2,500 followers this month (+25%). Engagement rate improved.”
Impact-led rewrite: “Followers increased 25%, but qualified leads from social stayed flat (310 vs 312) and sales-accepted leads fell (120 to 98). The follower growth came primarily from a broad-interest giveaway post that drove low-intent traffic. Next: shift content toward problem-aware topics with a clear CTA to demo/webinar, and measure success by qualified lead rate and demo show rate, not follower growth.”
Example 2: clicks up, revenue stagnant (ecommerce)
Vanity-led update: “CTR improved from 1.1% to 1.6% and clicks rose 45%.”
Impact-led rewrite: “Clicks rose 45% but revenue was flat because the additional traffic came from lower-intent placements; product page conversion fell from 3.2% to 2.4%. Next: exclude low-intent placements, align ad promise with landing page, and evaluate by gross margin per click and new-customer margin, not CTR.”
Example 3: impressions up during a promo, profit down
Vanity-led update: “The promo campaign reached 1.2M people and generated 18,000 clicks.”
Impact-led rewrite: “The promo increased reach and orders, but gross margin dollars declined 18% due to discount depth and a shift toward lower-margin products. Next: test a threshold discount (e.g., spend $X get $Y off) and measure incremental margin dollars and new-customer share.”
Benchmarks: What “Good” Looks Like (Without Fooling Yourself)
Benchmarks help you interpret whether a number is strong or weak. Use benchmarks carefully: a benchmark is only valid if it matches your context (industry, product, channel, audience, and time period).
Types of benchmarks you can use
- Internal historical: last week/month, trailing average, same period last year. Often the most reliable.
- Segment benchmarks: compare within the same channel, device, geo, or customer type.
- Target benchmarks: planned goals (e.g., “qualified lead to opportunity rate ≥ 25%”).
Step-by-step: benchmark a metric movement
- State the comparison: “Week B vs Week A” or “This month vs same month last year.”
- Check sample size: small denominators can create fake swings (e.g., conversion rate on 200 visits).
- Segment before judging: overall changes can hide opposite movements in key segments.
- Decide if it’s meaningful: is the change outside normal variation (use rolling averages or control charts if available)?
- Translate to impact: “A 0.3 pt conversion drop equals ~45 fewer orders at current traffic.”
Checklist: Interpreting Any Report
- 1) What decision does this report support? Name the decision (reallocate budget, change offer, fix funnel step).
- 2) What is the primary impact metric? Revenue, gross margin/profit, qualified leads, retention, or LTV (pick one).
- 3) What are the supporting leading indicators? Identify 1–3 that plausibly drive the impact metric.
- 4) What changed, exactly? Direction, magnitude, and time window; translate % into absolute counts or dollars.
- 5) Did traffic mix change? Check channel/device/geo/audience shares and within-segment performance.
- 6) Is seasonality or a calendar effect involved? Compare like-for-like periods; annotate known events.
- 7) Were promotions or pricing changes involved? Separate volume, price, product mix, and profit effects; check pull-forward.
- 8) What do cohorts say? Review first purchase month and acquisition channel cohorts for retention and LTV quality.
- 9) Are you relying on vanity metrics? If yes, rewrite the insight to connect activity to qualified outcomes and profit-aware results.
- 10) What is the most likely cause and the next test? State a hypothesis and one concrete action with a success metric.