Onboarding Metrics

Surfer โ€” Standard Plan

v1.1 โ€” May 2026
Krystian Prorok
One metric per step โ€” the signal that tells us if it's working. We measure first, then optimize. The targets below are what good looks like. The last section describes what those numbers unlock next.

Primary metrics

> 75%
Completion rate
Are users finishing the full flow?
< 4 min
Time to complete
Is the flow fast enough to not lose people?
Idea
Quality / Confidence
Valuable but hard to measure without post-onboarding correlation data. Candidate for next iteration.

Per-step metrics

1
Brand Setup

Primary metric: GSC connect rate

MetricWhat it tells youTarget
GSC connect rateAre users comfortable connecting Google accounts?> 60%
Edit rate per AI-generated fieldWhich fields does the AI get wrong? High rate on key topics degrades prompt quality downstream.20โ€“40% healthy
Time on stepIs reviewing AI suggestions causing friction?< 90s
Drop-off rateDid users abandon the flow at this step?< 15%
2
Pages / Posts

Primary metric: Average changes to AI pre-selection

MetricWhat it tells youTarget
Average changes (adds + removes)How much effort does selection take? Under 5 changes means the AI pre-selection is accurate.< 5 changes
Time on stepIs the selection task efficient?< 60s
Drop-off rateDid users abandon the flow at this step?< 15%
3
Prompts

Primary metric: Which categories are deselected most

MetricWhat it tells youTarget
Deselection rate per categoryWhich AI-generated categories are irrelevant for most users?Baseline only
% who changed default selectionAre users engaging with prompt review at all?Baseline only
Time on stepIs category review fast enough?< 45s
Drop-off rateDid users abandon the flow at this step?< 15%
3.5
Team Invite optional ยท skippable

Primary metric: Team invite rate

MetricWhat it tells youTarget
Team invite rateTeams that onboard together retain better.Baseline only
Skip rateIs this step causing drop-off?If > 80% skip โ†’ move post-onboarding
4
Completion

Primary metric: Return visit within 24 hours

MetricWhat it tells youTarget
Return visit < 24hDid the user come back to see first ChatGPT data?> 50%
Time to first dataHow fast does the product deliver value?< 24h

Time-to-return โ€” drop-off recovery

SegmentDefinitionSignalAction
Same-sessionReturns < 30 minInterrupted, still motivatedNone needed
Same-dayReturns < 24hStrong intentNone needed
PromptedReturns after nudgeNudge is workingOptimise nudge timing
Organic (2โ€“7 days)Returns without nudgeRecoverableSend nudge before day 7
ChurnedNo return after 7 daysLost โ€” note the stepDiagnose, redesign
Key lens: Split all metrics by connection method (GSC vs URL). URL users are expected to show lower pre-selection accuracy, longer time on step 2, and higher drop-off. That split is the primary diagnostic for what to improve first.

What's next

Once baseline data is in, these are the levers.

  1. 1
    Segment by company type โ€” Compare completion rates and deviation across segments once brand classification is reliable. Same UI, different benchmarks.
  2. 2
    Validate the page impact score โ€” Correlate Step 2 scores with actual ChatGPT citation data. If high-scored pages appear in citations more often, the ranking model is working.
  3. 3
    Instrument field edits โ€” Track which AI-generated values get replaced, and with what. This is the training signal for improving brand extraction.
  4. 4
    Correlate onboarding with 30-day retention โ€” Do GSC users retain better? Do users who edit more prompts churn faster? These become leading product health indicators.
  5. 5
    A/B test Concept A vs Concept B โ€” Use the completion rate baseline from this version to measure which concept performs better.