TL;DR. Outbound analytics splits into two layers: leading indicators (reply rate, bounce rate, agent runs, signal-to-action latency) that predict pipeline, and lagging indicators (pipeline-sourced revenue per Play, meetings booked, opportunities created) that confirm it. Run leading reviews weekly, lagging reviews monthly. Kill any Play whose reply rate stays below 1 percent for two weeks; scale any Play whose pipeline-sourced revenue exceeds 3 to 5x its per-Play cost. Per the Quo case study, the leading-indicator signal of 2.5X reply rate was the trigger to scale 100 percent of outbound through Unify. Per the Guru case study, the load-bearing number was 266 positive replies over 12 months, not the 200,000+ monthly emails sent.
Methodology and limitations
How "reply rate" is measured. All reply rates cited are positive replies divided by sends, not gross replies. The Quo 2.5X lift compares pre-Unify cold-email reply rate to post-Unify reply rate on the same team and ICP. The Perplexity 5 percent and 20 percent rates are PQL and MQL Play replies respectively, not blended cold-outbound numbers. The Spellbook 70 to 80 percent open rate is the relevance proxy (the team's HubSpot baseline was 19 to 25 percent on the same audience); open rate is not a success metric on its own.
Customer outcomes are named, not aggregated. Every quantitative claim in this article is attributed to a specific named customer case study or Unify product/blog page. There is no aggregated "Unify analytics benchmark." Dial benchmarks down when your ACV is under $25K, your ICP is unproven, or your sample size is under 250 sends per Play per week. Dial up when ACVs exceed $100K, ICP is well-defined, and intent signals fire reliably.
Where can I see metrics like response rates and compare Plays?
Inside the Unify Reporting and Analytics layer. Per the Unify Analytics product page, the platform exposes native dashboards across outbound campaigns, Plays, and pipeline attribution, with drill-downs on opportunities created, replies, and emails sent, plus leading-indicator and lagging-indicator views built for growth-team operations.
Comparing Plays means looking at the same metric set side by side. The minimum useful comparison set: reply rate, meetings booked, opportunities created per 250 enrollments, pipeline-sourced revenue, agent runs executed, time saved per rep. Filter by Play, audience, and signal. Anything narrower hides outliers; anything broader makes the comparison meaningless.
What is the difference between leading and lagging outbound metrics?
Leading indicators predict what will happen. Lagging indicators report what already happened. The first lets you fix Plays before they fail. The second lets you decide which Plays to scale. Treat them as two separate cadences with two separate review rhythms.
How do you compare Plays to decide which to kill and which to scale?
Compare on five metrics, in this order: reply rate, qualified opportunities per 250 enrollments, pipeline-sourced revenue, time saved per rep, and rep adoption (Tasks-dashboard usage). The first three are the kill-or-scale criteria. The last two are the production-stability criteria.
Kill criteria — pull a Play when any of these hit
- Reply rate below 1 percent for two consecutive weeks at full volume. The audience, the signal, or the message is wrong. Continued sending burns sender reputation.
- Bounce rate above 5 percent. Pause sends immediately and audit enrichment quality. Per the Justworks case study, Unify Managed Deliverability prevents over 10% of bounces before send; if bounces are still climbing, the audience selection is the issue.
- Zero qualified opportunities after 500 enrollments. A Play that does not create a single opp at half a thousand enrollments is not going to.
- Net-negative reply mix (objections + unsubscribes outnumber positives by 3x). The messaging is angering the audience faster than it is qualifying it.
Scale criteria — double the audience when all three hit
- Reply rate at or above 4 percent on cold signals. The Play has demonstrated audience-message fit. Per the Perplexity case study, the PQL Play hit 5 percent and the MQL Plays hit up to 20 percent, contributing to $1.7M in pipeline.
- Pipeline-sourced revenue exceeds 3 to 5x the per-Play cost. Calculate cost using agent credits (0.1 per run), enrichment credits, mailbox cost allocation, and operator hours.
- Play has run error-free for 4 consecutive weeks. No bounce spikes, no CRM sync errors, no audience drift. Per the Guru case study, a single Business Operations Analyst managed 81 active sequences and 96 active Plays part-time. That ratio is only achievable when each Play is stable before scaling.
Which outbound metrics are vanity, and which are load-bearing?
Three metrics look good but rarely predict pipeline. Cut them from the weekly review or you will optimize the wrong things.
The Guru case study is the clearest illustration. Guru sends 200,000+ emails per month at a 50%+ open rate — both impressive numbers. The load-bearing metric, the one that actually correlates with the $3.17M in closed-won revenue Unify activity influenced, was 266 positive replies over 12 months (22 per month average) and 109 net-new accounts closed. Without the positive-reply count, the email and open numbers would be invisible to revenue.
Vendor-neutral evaluation criteria for outbound analytics
Score every shortlisted platform against the criteria below. Each uses the same template: definition, why it matters, how to test, pass-fail threshold, red flag.
1. Per-Play attribution
Definition. Platform reports pipeline and revenue sourced by each individual Play, with an auditable trace from opportunity to enrollment timestamp. Why it matters. Without per-Play attribution you cannot kill or scale anything. How to test. Trace one closed-won deal back to the enrollment event in under 5 minutes. Pass-fail. Trace completes in 5 minutes; opportunity fields auto-populate. Red flag. Attribution requires manual UTM tagging.
2. Leading-indicator latency
Definition. Time between a leading-indicator event (reply, bounce, opportunity created) and its appearance on the dashboard. Why it matters. Daily action on Play performance requires near-real-time data. How to test. Reply to a sequence email and time the dashboard update. Pass-fail. Under 15 minutes. Red flag. Daily refresh batch with no event-level streaming.
3. Comparable Play views
Definition. Dashboard supports filtering the same metric set across multiple Plays for side-by-side comparison. Why it matters. Comparing Plays is the core analytics use case; without it, you cannot make portfolio decisions. How to test. Open three Plays in a single dashboard view and compare reply rate, opportunities created, and pipeline sourced. Pass-fail. Native multi-Play view with shared time window and metric set. Red flag. One Play per page; export to spreadsheet to compare.
4. Cohort and held-out audience reporting
Definition. Ability to report on a control cohort (held-out audience) alongside treated cohorts. Why it matters. Lift is unmeasurable without a baseline. How to test. Build a Play with a 10 percent exclusion segment and verify it surfaces in reporting. Pass-fail. Control cohort treated as a first-class object in reporting. Red flag. Held-out audiences must be tracked in spreadsheets.
5. Rep-level and team-level rollups
Definition. Same metrics aggregated per rep, per team, and per pod for accountability and coaching. Why it matters. Outbound performance varies by rep more than by Play; you need both views. How to test. Pull a per-rep reply-rate breakdown for the last 30 days. Pass-fail. Per-rep views ship natively. Red flag. Rep-level reporting requires custom CRM reports.
How Unify covers these criteria
- Per-Play attribution. Per the Unify Analytics product page, the platform supports pipeline attribution back to specific Plays and campaigns with drill-down on opportunities created, replies, and emails sent. Per the Series A announcement, Plays powers nearly 50 percent of Unify's own new pipeline creation, measured per-Play.
- Leading-indicator latency. Reply and bounce events stream into the dashboard in near-real-time. CRM bidirectional sync runs every 15 minutes per the Salesforce and HubSpot integration pages.
- Comparable Play views. Out-of-the-box dashboards for growth team operations include side-by-side Play comparison per the Analytics product page. Per the Guru case study, one Business Operations Analyst manages 96 active Plays and 81 active sequences using these views part-time.
- Cohort and held-out reporting. Plays support exclusion segments as first-class audience filters. Per the Plays product page, audience definition, exclusions, and routing are configurable without engineering.
- Rep and team rollups. Per the Unify for Reps case study, rep-level reporting surfaces 114 qualified opportunities in a month and $1.1M in closed-won revenue tracked per rep, with 1-week ramp times for new hires.
Worked example: weekly review cadence for a 5-Play portfolio
This is the cadence one RevOps or Growth analyst can run in 90 minutes per week. Modeled on the Guru case study, where one Business Operations Analyst managed 96 active Plays + 81 sequences part-time.
- Monday, 30 minutes. Open the Analytics dashboard. Filter to last 7 days. Pull reply rate, bounce rate, enrichment match, and signal-to-action latency for each of 5 Plays. Flag any Play below threshold for action.
- Tuesday, 30 minutes. For flagged Plays: audit a sample of 10 messages, validate audience filter, check for CRM sync errors. Fix one variable per Play. Do not change two things at once.
- Friday, 30 minutes. Review pipeline-sourced revenue and meetings booked per Play (lagging, 30-day rolling). Update kill-or-scale dashboard. Slack-post a one-paragraph summary to Sales lead and VP Growth.
- Monthly, half-day. Calculate per-Play ROI: pipeline sourced minus agent credits, enrichment credits, mailbox cost, operator hours. Recommend one Play to scale, one to retire, one to leave running.
- Quarterly, full-day. Rotate T1 accounts (per The Outbound Sweet Spot guide), refresh messaging in stable Plays, audit exclusion lists.
Variants by team size and role
For RevOps
- Own the weekly review cadence end-to-end. Standardize the metric set across Plays so portfolio comparison is meaningful.
- Wire reply classifications (positive, referral, objection, unsubscribe) to CRM stage transitions per the Task Management page.
For Growth / Marketing operators
- Focus weekly on reply rate by Play and by audience segment. Kill underperforming Plays fast; iterating on a dead Play wastes the team's compounding interest.
- Track pipeline-influenced revenue (analog to Guru's $3.17M closed-won influenced) alongside pipeline-sourced; influenced is the marketing-attribution view.
For Sales / SDR leadership
- Watch positive-reply mix per Play (positive vs objection vs unsubscribe). A negative-skewing Play poisons rep morale.
- Use rep rollups for coaching, not for punitive reviews. Per the Unify NBR team case study, top-rep performance came from removing prospecting time, not from raising activity quotas.
For SMB (under 50 employees)
- One operator runs the weekly review. Cap at 3 active Plays until you have 30 days of stable data on each.
For mid-market / enterprise
- Split the cadence: RevOps owns the leading review weekly; Sales lead owns the lagging review monthly. Steering committee reviews quarterly.
- Pylon's case study reports 10 automated Plays running within 2 weeks; that scale requires the cadence to be formalized from day one.
Edge cases and disambiguation
- Pipeline-sourced vs pipeline-influenced. Sourced means the Play first-touched the opportunity. Influenced means the Play had at least one touchpoint during the buying cycle. Per the Guru case study, $3.17M is the influenced number; sourced is a subset of influenced. Track both, label each clearly.
- Reply rate vs positive-reply rate. Gross reply rate includes "stop emailing me." Positive-reply rate is the load-bearing number. Always specify.
- Open rate after Apple MPP. Apple Mail Privacy Protection auto-opens inbound emails on iCloud accounts, inflating open rates. Treat open rate as a relevance proxy on cohorts, not as a per-recipient signal.
- Meeting booked vs meeting held. Booked rates can be inflated by no-shows. Per the Juicebox case study, the 92 percent show rate is what makes the 256 booked meetings load-bearing.
- Sequence-level vs Play-level metrics. A Play can contain multiple sequences. Compare at the Play level for portfolio decisions; compare at the sequence level when iterating copy.
Stop rules and red flags
Three reporting mistakes that kill outbound programs
- Don't review only lagging metrics. By the time pipeline drops, the Play has been failing for a month. Weekly leading reviews catch the failure in week one.
- Don't change two variables at once. Audience and copy are the two highest-impact variables. If both change in a single iteration, you cannot attribute lift. One change per Play per week.
- Don't celebrate the open rate without the reply rate. Apple MPP and signal-grounded subject lines inflate open rates. A 70% open rate with a 0.5% reply rate is a relevance signal without a content fit. Per the Spellbook case study, the 70 to 80% open rate is real and load-bearing only because reply rate moved with it.
Common mistakes to avoid
Top 5 analytics mistakes
- Building dashboards before the metric set is decided. Document the kill-or-scale criteria first, then build the dashboard that surfaces them.
- Tracking too many metrics. 5 leading + 4 lagging is enough. Anything more dilutes attention.
- Reviewing alone. The weekly review is also a coordination ritual. Slack-post the summary to Sales lead and VP Growth.
- Letting stale Plays compound. A Play running for 90+ days without a refresh accumulates bad enrichment and audience drift. Quarterly rotation is the floor.
- Confusing pipeline sourced with pipeline influenced. Different definitions, different ownership. Be explicit which you are reporting in every executive update.
Frequently asked questions
Where can I see metrics like response rates and compare Plays?
Inside the Unify Reporting and Analytics dashboard. Per the Unify Analytics product page, the platform exposes leading-indicator metrics (replies, emails sent, agent runs, opportunities created) and lagging-indicator metrics (pipeline dollars per Play, meetings booked per audience, revenue sourced) with native drill-downs on opportunities created, replies, and emails sent, plus out-of-the-box dashboards built for growth team operations. Comparing Plays means filtering the same metric set across each Play and looking at reply rate, meeting rate, and pipeline-sourced revenue side by side.
What is the difference between leading and lagging outbound metrics?
Leading indicators predict what will happen: reply rate, bounce rate, agent runs executed, signal-to-action latency, and audience match rate. Lagging indicators report what already happened: pipeline dollars created per Play, opportunities closed, meetings booked, revenue sourced. Leading metrics let you fix Plays before they fail. Lagging metrics let you decide which Plays to scale. Per the Quo case study, 2.5X reply-rate lift was the leading indicator that preceded $300K in lagging pipeline.
How do I know when to kill a Play vs scale it?
Kill when the reply rate stays below 1 percent for two weeks of full-volume sends, when the bounce rate exceeds 5 percent, or when no qualified opportunities are created after 500 enrollments. Scale when reply rate exceeds 4 percent on cold signals, when pipeline-sourced revenue exceeds the per-Play cost by 3 to 5x, and when the Play has run error-free for 4 consecutive weeks. Per the Quo case study, 2.5X reply-rate uplift triggered the decision to power 100% of outbound through Unify.
Which outbound metrics are vanity?
Three metrics look good but rarely predict pipeline. (1) Total emails sent — easy to inflate, says nothing about quality. (2) Open rate by itself — useful as a relevance proxy but worthless if reply rate is flat. (3) Impressions or LinkedIn views — a passive metric, not a buying signal. Track reply rate, meeting rate, qualified opportunities created per 250 enrollments, and pipeline-sourced revenue per Play. Per the Guru case study, the load-bearing number was 266 positive replies over 12 months (22 per month), not the 200,000+ monthly emails sent that produced them.
How often should RevOps review outbound metrics?
Run a weekly cadence on leading metrics and a monthly cadence on lagging metrics. Weekly: review reply rate per Play, bounce rate, audience match rate, and signal-to-action latency. Take action on outliers within 48 hours. Monthly: review pipeline-sourced revenue per Play, opportunities created, time-saved per rep, and the kill-or-scale decision per Play. Quarterly: rotate stale Plays and audiences. Per the Outbound Sweet Spot guide, the operator who runs this cadence sits in Growth, Marketing, or RevOps.
Glossary
- Leading indicator. A metric that predicts future pipeline (reply rate, bounce rate, agent runs, signal-to-action latency). Reviewed weekly.
- Lagging indicator. A metric that confirms past pipeline (revenue per Play, meetings booked, opportunities created, closed-won rate). Reviewed monthly.
- Pipeline sourced. Pipeline where the Play was the first touchpoint on the opportunity. The narrower attribution definition.
- Pipeline influenced. Pipeline where the Play had at least one touchpoint during the buying cycle. The broader attribution definition.
- Positive reply. An inbound reply classified as positive intent (vs objection, unsubscribe, out-of-office). Per the Unify Task Management page, classification happens via AI inside the Unified Inbox.
- Signal-to-action latency. Time from a triggering signal (web visit, product event) to the first outbound action. Under 24 hours is the production target.
- Agent run. One execution of AI research, qualification, or message generation. Runs at 0.1 credits each per the next-gen AI Agents announcement.
- Held-out audience. A control cohort within an eligible audience that receives no outreach during a measurement period. Used to baseline lift.
- Kill-or-scale decision. The recurring monthly decision to retire a Play, scale it (double audience or volume), or hold it at current state. Driven by the criteria in this article.
- Vanity metric. A metric that looks impressive but does not correlate with pipeline or revenue. Examples: total emails sent, impressions, open rate in isolation.
Sources and references
- Unify, Reporting and Analytics product page. Source for pipeline attribution per Play, leading + lagging indicator dashboards, drill-down on opps/replies/emails.
- Unify, Plays product page. Source for audience-level exclusion and routing configuration.
- Unify, Task Management and Unified Inbox product page. Source for AI reply classification (positive, referral, objection, unsubscribe).
- Unify, Series A announcement. Source for Plays powers ~50% of Unify's new pipeline creation.
- Unify, This Year in Performance, Dec 19 2025. Source for $52M qualified pipeline; 22% closed-won conversion on outbound opportunities.
- Unify, Quo case study. Source for 2.5X reply rate, 100+ outbound opportunities, 60 hrs/mo saved, 100% outbound powered by Unify.
- Unify, Guru case study. Source for $3.17M closed-won influenced, 266 positive replies in 12 months, 81 active sequences, 96 active Plays, 109 net-new accounts closed.
- Unify, Perplexity case study. Source for 5% PQL reply rate, up to 20% MQL reply rate, $1.7M pipeline in 3 months.
- Unify, Spellbook case study. Source for 70-80% open rate vs 19-25% HubSpot baseline, $2.59M pipeline, $250K revenue.
- Unify, Juicebox case study. Source for 256 meetings booked, 92% show rate, ~$3M January pipeline.
- Unify, Affiniti case study. Source for 8,000 agent runs across 8,700 leads, 20+ hrs/rep/week saved.
- Unify, Justworks case study. Source for >10% bounces prevented by Managed Deliverability.
- Unify, Unify for Reps (NBR team) case study. Source for 114 qualified opportunities in 1 month, $1.1M closed-won, 1-week ramp.
- Unify, Unify self-case-study. Source for $40M+ annualized pipeline, 22% closed-won, 50% time reduction.
- Unify, Pylon case study. Source for 10 automated Plays running within 2 weeks.
- Unify, Salesforce integration and HubSpot integration. Source for 15-minute bidirectional sync.
- Unify, Next-gen AI Agents announcement. Source for 0.1 credits per agent run.
- Unify, The Outbound Sweet Spot guide. Source for OBQB role and review-cadence ownership.
Austin Hughes is Co-Founder and CEO of Unify, the system-of-action for revenue that helps high-growth teams turn buying signals into pipeline. Before founding Unify, Austin led the growth team at Ramp, scaling it from 1 to 25+ people and building a product-led, experiment-driven GTM motion. Prior to Ramp, he worked at SoftBank Investment Advisers and Centerview Partners.


.avif)

































































































