Chapter 5: Metrics & Success Criteria
North Star Metric, Retention, NPS, and actionable analytics.
What Gets Measured Gets Managed
Startups drown in data. Every analytics tool offers hundreds of metrics -- downloads, page views, sessions, sign-ups. The problem isn't lack of data; it's knowing which data actually matters.
The wrong metrics create a dangerous illusion of progress. The right metrics force honest confrontation with reality.
This distinction isn't academic -- it has life-or-death consequences for your startup. Consider two scenarios: In Scenario A, your team celebrates hitting 10,000 total sign-ups and writes a blog post about "10K users!" In Scenario B, your team discovers that only 3% of those 10,000 users returned after Day 7 and immediately pivots to fix the onboarding flow. Both teams had the same underlying data, but Scenario A looked at vanity metrics (cumulative sign-ups) while Scenario B looked at actionable metrics (retention rate). Scenario A wasted weeks celebrating while the business slowly died. Scenario B identified the problem and started fixing it.
The metrics you choose to track aren't just measurements -- they're decisions about what your team pays attention to. When you put "total sign-ups" on a dashboard, your team optimizes for sign-ups. When you put "Day 7 retention" on the dashboard, your team optimizes for retention. The metric shapes the behavior, which shapes the outcome. Choose wisely.
The Core Insight
There's only one metric that proves product-market fit: retention. If users don't come back, nothing else matters. Build your entire measurement strategy around understanding and improving retention.
Bug #1: Celebrating Vanity Metrics
Vanity metrics are the most dangerous trap in startup analytics. They always go up over time, making the team feel good -- while the business slowly dies.
The Bug
"We hit 100,000 downloads!"
Cumulative metrics only go up. A product can have 100K downloads and 0 active users. These metrics create false confidence while your actual business hemorrhages.
The Fix
Focus exclusively on actionable metrics.
Track cohort retention, activation rate, and conversion rate -- metrics that can go down, that tell you something is broken, and that you can directly influence.
Vanity vs. Actionable Metrics
Vanity Metrics
These lie. They only go up. They feel good but teach nothing.
| Cumulative downloads |
| Total registered users |
| Total page views |
| Social media followers |
| "Time on site" without context |
Actionable Metrics
These tell the truth. They can go down. They drive action.
| Cohort retention (Week 1, Week 4, Week 8) |
| Activation rate (% who reach "aha moment") |
| Conversion rate (Visitor to Trial to Paid) |
| Net Promoter Score (NPS) |
| Revenue per user (RPU) |
The Vanity Test
Ask yourself: "Can this metric go down?" If a metric can only increase over time (like cumulative downloads), it's vanity. If it can decrease and signal something is broken (like Day 30 retention), it's actionable.
There's a subtler form of vanity metrics that catches even experienced founders: rate metrics without denominator transparency. "Our conversion rate is 15%!" sounds impressive until you learn it's based on 20 visitors. "We have 40% Day 1 retention!" sounds great until you learn it's from a cohort of 5 users. Always pair rate metrics with the sample size. Small samples produce wildly unreliable rates that can swing from 0% to 100% based on a handful of individual decisions. As a rule of thumb, you need at least 100 users in a cohort before retention rates become statistically meaningful.
The Pirate Metrics Framework (AARRR)
Dave McClure's Pirate Metrics framework gives you a structured way to think about your entire funnel. Each stage represents a different aspect of user behavior, and each has its own key metric:
The AARRR Funnel
| Stage | Question | Key Metric | Example |
|---|---|---|---|
| Acquisition | How do users find you? | Traffic by channel | 1,000 visitors/month from Google |
| Activation | Do they have a great first experience? | % reaching "aha moment" | 40% complete onboarding |
| Retention | Do they come back? | Cohort retention curve | 25% active at Day 30 |
| Revenue | Do they pay? | Conversion to paid | 5% of activated users convert |
| Referral | Do they tell others? | Viral coefficient | Each user invites 0.3 others |
The AARRR framework is powerful because it forces you to think about your business as a funnel where each stage depends on the previous one. Investing in Acquisition when your Activation rate is 5% is like pouring water into a bucket with a hole in the bottom. Fix the Activation hole first, then scale Acquisition. The framework makes these dependencies explicit.
For MVP-stage startups, the most common mistake is over-indexing on Acquisition (top of funnel) while ignoring Activation and Retention (middle of funnel). LeanPivot's Pirate Metrics (AARRR) tool helps you define and track metrics across all five stages, identifying your biggest funnel bottleneck so you focus improvement efforts where they'll have the most impact.
The North Star Metric (NSM)
Your North Star Metric is the single number that best captures the core value your product delivers. It aligns the entire organization -- product, marketing, sales, support -- around one goal.
Criteria for a Good North Star Metric
| Value-Based | Reflects value delivered to customers, not just revenue extracted |
| Leading Indicator | Predicts future success (revenue is a lagging indicator) |
| Actionable | The team can directly influence it through product decisions |
| Understandable | Every team member can explain what it means |
The distinction between leading and lagging indicators is critical. Revenue is the ultimate lagging indicator -- by the time revenue drops, the underlying problem has been festering for weeks or months. A good North Star Metric predicts revenue changes before they happen. If Spotify's "Time Spent Listening" drops this month, churned subscriptions will follow next month. The NSM gives you an early warning system that lets you act before the damage is done.
North Star Examples by Category
Consumption Products
Spotify: Time Spent Listening
Netflix: Watch Time
YouTube: Watch Minutes
Transaction Products
Airbnb: Nights Booked
Uber: Rides per Week
Amazon: Purchases per Month
Productivity Products
Slack: Messages Sent
Zoom: Hosted Meetings
Notion: Blocks Created
Finding Your North Star
Answer this: "What single action, when done repeatedly, indicates the customer is getting value?" That action is your North Star. For Airbnb, it's booking nights. For Spotify, it's listening to music. For Slack, it's sending messages.
A common mistake is choosing revenue as your North Star Metric. Revenue measures value captured, not value delivered. A company can increase revenue in the short term by raising prices or reducing service quality -- both of which destroy long-term value. A value-based North Star forces you to focus on what matters: delivering value to customers, which is the only sustainable path to revenue growth.
The Activation Metric: Your First Critical Gate
Before users can be retained, they need to be activated. Activation is the moment when a new user first experiences the core value of your product -- the "aha moment."
Finding Your Activation Moment
| Product | Aha Moment | Activation Metric |
|---|---|---|
| Slack | First team message exchange | Team sends 2,000+ messages |
| Dropbox | First file synced across devices | User saves 1 file to Dropbox folder |
| Twitter (X) | First engaging follow | User follows 30+ accounts |
| First friend connections | User adds 7 friends in 10 days |
Identifying your activation metric requires looking at behavioral data: what do retained users do in their first session that churned users don't? Segment your users into two groups -- those who are still active after 30 days and those who aren't -- then compare their first-session behavior. The behavioral differences reveal what drives retention. Facebook famously discovered that users who added 7 friends within 10 days were dramatically more likely to become long-term users, and they restructured their entire onboarding flow to drive that specific behavior.
For MVP-stage products, you may not have enough data for this kind of statistical analysis. In that case, use your concierge and Wizard of Oz experiments to observe the activation moment qualitatively: watch users interact with your product and note the moment their expression changes from confusion to understanding. That's your aha moment. Design your onboarding to accelerate that moment.
Retention: The Only Metric That Matters
For an early-stage MVP, retention is the only metric that truly proves product-market fit. If users don't come back, you haven't built something people want.
The Bug
"Let's focus on acquisition. We need more users."
Acquisition without retention is burning cash into a leaky bucket. You're paying to acquire users who immediately leave. Fix retention before scaling acquisition.
The Fix
Build a retention curve first.
Track cohort retention over time. Your curve should flatten (users stick around), not plummet to zero. Don't scale until the curve flattens.
Retention Benchmarks by Business Model
What "Good" Looks Like
| Business Model | Metric | Weak | Good | Great |
|---|---|---|---|---|
| B2B SaaS | Net Revenue Retention | <90% | 100-110% | >120% |
| Monthly Logo Churn | >5% | 2-3% | <2% | |
| B2C Mobile | Day 1 Retention | <25% | 30-40% | >50% |
| Day 30 Retention | <10% | 15-20% | >25% | |
| Consumer Web | Week 1 Retention | <15% | 20-30% | >40% |
| Month 3 Retention | <10% | 15-25% | >30% |
The Retention Curve
Reading Your Retention Curve
Plot the percentage of users still active over time. What you're looking for:
| Curve goes to zero | You don't have product-market fit. Users try and leave. Fix the product. |
| Curve flattens low | You have a niche audience. Consider if the segment is big enough. |
| Curve flattens high | Product-market fit achieved. You can now scale acquisition. |
The shape of your retention curve tells you more than almost any other piece of data. A curve that flattens means you've found a group of users who find lasting value in your product -- these are your true customers. The level at which it flattens tells you how broadly your product resonates. If it flattens at 5%, you have a narrow niche. If it flattens at 40%, you have broad appeal. Both can be viable businesses, but they require very different growth strategies.
One nuance that many founders miss: retention should be measured by cohort, not in aggregate. Aggregate retention can look stable even while each individual cohort's retention is declining -- because new users mask the churn of old users. Cohort analysis separates each group of users who signed up in the same period and tracks their retention independently. This reveals whether your product is actually improving over time (later cohorts retain better) or just treading water.
Net Promoter Score (NPS)
NPS measures customer loyalty and predicts viral growth. It's simple, standardized, and benchmarkable.
How NPS Works
The Question: "On a scale of 0-10, how likely are you to recommend [Product] to a friend or colleague?"
Promoters (9-10)
Active advocates. Will refer others.
Passives (7-8)
Satisfied but unenthusiastic. Won't refer.
Detractors (0-6)
Unhappy. May actively discourage others.
Formula: NPS = % Promoters - % Detractors (ranges from -100 to +100)
NPS Benchmarks
Below 0
Danger zone. More detractors than promoters. Fix immediately.
0-30
Good. More lovers than haters. Room to improve.
30-50
Great. Typical for successful SaaS companies.
50+
World-class. Apple, Tesla territory. Strong viral potential.
The Real Value of NPS
The number matters less than the follow-up question: "Why did you give that score?" The qualitative feedback reveals what's broken (from detractors) and what's magical (from promoters). Use LeanPivot's Usability Testing tool to design NPS surveys with effective follow-up questions that generate actionable insights.
The Sean Ellis Test: Measuring Product-Market Fit
NPS measures recommendation intent. The Sean Ellis test measures something more fundamental: dependency. If users would be "very disappointed" without your product, you've achieved product-market fit.
The Sean Ellis Survey
The Question: "How would you feel if you could no longer use [Product]?"
| Very disappointed | Target: 40%+ of respondents | Product-market fit signal |
| Somewhat disappointed | These users like you but don't need you | Room to improve |
| Not disappointed | These users are at risk of churning | Product issue |
The benchmark: If 40%+ of users say "very disappointed," you have product-market fit. Below 40%, keep iterating on the product before scaling.
The 40% threshold comes from empirical analysis across hundreds of startups. Companies that crossed the 40% "very disappointed" threshold consistently achieved sustainable growth, while those below it struggled regardless of how much they invested in marketing and sales. The test is particularly useful for MVP-stage products because it requires only 40-50 survey responses to be meaningful -- far fewer than retention analysis, which requires hundreds or thousands of users.
Your Metrics Dashboard Checklist
The MVP Metrics Stack
| North Star defined: One metric that captures core value delivery | |
| Retention tracked: Cohort retention by week (D1, D7, D30 minimum) | |
| Activation defined: Clear "aha moment" and % of users reaching it | |
| NPS measured: Score + qualitative feedback collection | |
| Vanity eliminated: No cumulative metrics in weekly reviews | |
| AARRR funnel mapped: Conversion rates between each stage identified |
Key Takeaways
Remember These Truths
- Vanity metrics lie. If a metric can only go up, it's hiding the truth.
- Define your North Star. One metric that captures the core value you deliver.
- Retention is everything. If users don't come back, nothing else matters.
- Flatten the curve first. Don't scale acquisition until retention stabilizes.
- NPS reveals why. The qualitative follow-up is more valuable than the score.
Now that you know what to measure, let's explore how to prepare for launch and manage beta testing effectively.
Define Your Success Metrics with AI
Use our unit economics and metrics tools to define KPIs that actually matter for your business model.
Save Your Progress
Create a free account to save your reading progress, bookmark chapters, and unlock Playbooks 04-08 (MVP, Launch, Growth & Funding).
Ready to Design Your MVP?
LeanPivot.ai provides 80+ AI-powered tools to design, build, and launch your MVP.
Start Free TodayRelated Guides
Lean Startup Guide
Master the build-measure-learn loop and the foundations of validated learning to build products people actually want.
From Layoff to Launch
A step-by-step guide to turning industry expertise into a thriving professional practice after a layoff.
Fintech Playbook
Master regulatory moats, ledger architecture, and BaaS partnerships to build successful fintech products.
Works Cited & Recommended Reading
RAT vs MVP Philosophy
- 1. Ries, E. (2011). The Lean Startup. Crown Business.
- 2. "Why RAT (Riskiest Assumption Test) beats MVP every time." LinkedIn
- 3. "Pretotyping: The Art of Innovation." Pretotyping.org
- 6. "Continuous Discovery: Product Trio." Product Talk
- 7. "MVP Fidelity Spectrum Guide." SVPG
Minimum Lovable Product
- 8. Olsen, D. (2015). The Lean Product Playbook. Wiley.
- 9. "From MVP to MLP: Why 'Viable' Is No Longer Enough." First Round Review
- 10. "Minimum Lovable Product framework." Amplitude Blog
Hypothesis-Driven Development
- 11. Gothelf, J. & Seiden, J. (2021). Lean UX. O'Reilly Media.
- 12. "Hypothesis-Driven Development in Practice." ThoughtWorks Insights
- 13. "Experiment Tracking Best Practices." Optimizely
- 14. "Build-Measure-Learn: The Scientific Method for Startups." Harvard Business Review
Assumption Mapping
- 15. Bland, D. & Osterwalder, A. (2019). Testing Business Ideas. Wiley.
- 16. "Risk vs. Knowledge Matrix." Miro Templates
- 17. "Identifying Riskiest Assumptions." Intercom Blog
User Story & Impact Mapping
- 20. Patton, J. (2014). User Story Mapping. O'Reilly Media.
- 21. Adzic, G. (2012). Impact Mapping. Provoking Thoughts.
- 22. "Jobs-to-Be-Done Story Framework." JTBD.info
- 23. "The INVEST Criteria for User Stories." Agile Alliance
- 24. "North Star Metric Framework." Amplitude
- 25. "Opportunity Solution Trees." Product Talk
- 26. Torres, T. (2021). Continuous Discovery Habits. Product Talk LLC.
Pretotyping Techniques
- 27. Savoia, A. (2019). The Right It. HarperOne.
- 28. "Fake Door Testing Guide." UserTesting
- 29. "Wizard of Oz Testing Method." Nielsen Norman Group
- 30. "Concierge MVP Explained." Grasshopper
Prioritization Frameworks
- 31. "ICE Scoring Model." ProductPlan
- 32. "RICE Prioritization Framework." Intercom
- 33. "Kano Model for Feature Analysis." Folding Burritos
- 34. "MoSCoW Method Guide." ProductPlan
Build vs Buy & No-Code
- 35. "No-Code MVP Tools Landscape." Makerpad
- 37. "Technical Debt in Early Startups." a16z
- 38. "Prototype Fidelity Selection." Interaction Design Foundation
- 39. "API-First Development Strategy." Swagger
- 40. "Rapid Prototyping with Bubble & Webflow." Bubble Blog
Metrics & Analytics
- 41. Croll, A. & Yoskovitz, B. (2013). Lean Analytics. O'Reilly.
- 42. "One Metric That Matters (OMTM)." Lean Analytics
- 43. McClure, D. "Pirate Metrics (AARRR)." 500 Startups
- 44. "Vanity Metrics vs. Actionable Metrics." Mixpanel
- 45. "Cohort Analysis Deep Dive." Amplitude
- 46. "A/B Testing Statistical Significance." Optimizely
- 47. "Product Analytics Instrumentation." Segment Academy
- 48. "Activation Metrics Framework." Reforge
- 49. "Leading vs Lagging Indicators." Productboard
- 50. "Retention Curve Analysis." Sequoia Capital
- 51. "Feature Adoption Tracking." Pendo
- 52. "Experimentation Velocity Metrics." ExP Platform
Launch Operations & Analysis
- 53. "Soft Launch Strategy." Mind the Product
- 54. "Feature Flag Best Practices." LaunchDarkly
- 55. "Beta Testing Program Design." BetaList
- 56. "Customer Feedback Loop Systems." Canny
- 57. "Rollback Strategy Planning." Atlassian
- 58. "Why Startups Fail: Post-Mortems." CB Insights
- 59. "Pivot vs Persevere Decisions." Steve Blank
- 60. "Learning from Failed Experiments." HBR Innovation
This playbook synthesizes methodologies from Lean Startup, Design Thinking, Jobs-to-Be-Done, Pretotyping, and modern product management practices. References are provided for deeper exploration of each topic.