In December 2003, Fred Reichheld published an article in the Harvard Business Review titled "The One Number You Need to Grow." The article made a remarkable claim: after years of research across hundreds of companies, Reichheld had identified a single survey question that correlated more strongly with business growth than any other customer feedback metric. That question was: "How likely is it that you would recommend our company to a friend or colleague?"
The resulting metric — the Net Promoter Score — became one of the most widely adopted management tools in business history. Within a decade, two-thirds of Fortune 1000 companies were measuring it. Bain and Company built a consulting practice around it. Business books praised it as the holy grail of customer loyalty measurement. Executives received bonuses tied to it.
It also generated some of the most pointed methodological criticism in the history of applied market research. Understanding both what NPS measures and what it does not is essential for anyone who encounters it at work — which is, increasingly, almost everyone.
How Net Promoter Score Works
The mechanics of NPS are straightforward. A customer is asked one question:
"On a scale of 0 to 10, how likely is it that you would recommend [Company/Product] to a friend or colleague?"
Responses are categorized into three groups:
| Score | Category | Description |
|---|---|---|
| 9-10 | Promoters | Enthusiastic customers likely to fuel growth through referrals and repeat purchases |
| 7-8 | Passives | Satisfied but unenthusiastic customers who are vulnerable to competitive offerings |
| 0-6 | Detractors | Unhappy customers who may actively discourage others from using the product |
The NPS is calculated by subtracting the percentage of Detractors from the percentage of Promoters:
NPS = % Promoters - % Detractors
Passives are excluded from the calculation entirely. The resulting score ranges from -100 (all respondents are Detractors) to +100 (all respondents are Promoters).
Reichheld published this framework in a 2003 HBR article and expanded it in the 2006 book The Ultimate Question: Driving Good Profits and True Growth, co-authored with Rob Markey. Bain and Company, where Reichheld is a fellow, commercialized the NPS system, creating certification programs, industry benchmarks, and consulting services built around its implementation.
The Origin Story: Good Profits vs. Bad Profits
The theoretical foundation of NPS rests on a distinction Reichheld drew between two types of profit: bad profits, earned by exploiting or trapping customers in ways that damage loyalty, and good profits, earned by delivering genuine value that customers want to pay for and recommend.
Reichheld's 2003 analysis of customer loyalty data from multiple industries — including financial services, telecommunications, and technology — claimed to find that the likelihood-to-recommend question outperformed longer satisfaction surveys in predicting business growth. His methodology compared loyalty metrics across roughly 400 companies, tracking revenue growth over a two-year period. The study was not published in a peer-reviewed academic journal; it appeared as a management article aimed at practitioners.
The commercial context matters. Bain and Company, where Reichheld is a Fellow, subsequently built significant consulting revenue around NPS implementation and benchmarking. This commercial interest does not invalidate the framework, but it explains both the scale of its adoption and the relative lack of independent scrutiny in its early years.
What NPS Attempts to Measure
Reichheld argued that customer recommendation behavior — captured by the likelihood-to-recommend question — is the behavioral proxy that best distinguishes companies that earn good profits from those that earn bad ones.
"The percentage of customers enthusiastic enough to refer a friend or colleague — perhaps the ultimate act of loyalty — turns out to be the best predictor of growth." — Fred Reichheld, Harvard Business Review, December 2003
The appeal of NPS beyond its predictive claims was its simplicity. A single question requires less customer effort to answer, produces higher response rates, and is easier to benchmark against industry peers than a 30-question satisfaction survey. In an era when organizations were drowning in customer feedback data they lacked the capacity to act on, the promise of one number was genuinely attractive.
Reichheld's framework also offered a philosophical reframe of customer service: the goal was not merely satisfaction but generating enough enthusiasm that customers would stake their own reputation on recommending the company. That is a meaningfully higher bar than simply avoiding dissatisfaction, and it gave organizations a conceptually compelling north star.
How NPS Differs From Customer Satisfaction Surveys
Traditional customer satisfaction (CSAT) surveys ask respondents to rate their satisfaction with a product, service, or specific interaction — typically on a 1-5 or 1-10 scale. They capture a moment-in-time emotional reaction to a specific experience.
NPS attempts to measure something different: the overall relationship between the customer and the brand. The framing of a recommendation implies the customer is being asked to evaluate not just their own experience but whether they would vouch for that experience to someone they care about. This is theoretically a more demanding and more relationship-level question than satisfaction.
In practice, the distinction is less clean. Many customers interpret both questions in similar ways, and the correlation between NPS and CSAT scores is often high in the same customer population. But the conceptual distinction is real and explains why many organizations use both metrics for different purposes.
The Criticisms: What the Research Actually Found
The most rigorous academic challenge to Reichheld's claims came in a 2007 meta-analysis by Timothy Keiningham and colleagues published in the Journal of Marketing. The study analyzed data from 21 longitudinal studies covering multiple industries and countries and asked a simple question: does NPS predict business growth better than other loyalty metrics?
The answer was no.
Keiningham et al. found that customer satisfaction and other loyalty metrics predicted revenue growth as well as or better than NPS in most of their analyses, and that NPS did not hold a statistically consistent advantage across industries or contexts. The specific claim that likelihood to recommend is the single best predictor of growth was not supported by the data.
"Our findings indicate that NPS is not the 'single most reliable indicator of a company's ability to grow,' as has been claimed. Furthermore, our study finds that customer satisfaction measures provide a better predictor of revenue growth than NPS in most contexts." — Keiningham, Cooil, Andreassen, and Aksoy, Journal of Marketing, 2007
This finding has been replicated and extended in subsequent research. A 2020 paper by Schneider and colleagues examining customer data from B2B subscription businesses found that actual renewal behavior — not NPS — was the strongest predictor of retention, with NPS adding minimal explanatory power above simpler behavioral signals.
A 2016 meta-analysis by Morgan and Rego, published in the Journal of Marketing Research, examined the predictive validity of NPS across industries and found that other loyalty metrics, including customer satisfaction and repurchase intention, consistently matched or exceeded NPS in predicting business outcomes. The "single best predictor" claim found no empirical support across independent analyses.
Specific Methodological Criticisms
Beyond the predictive validity question, NPS has attracted a range of methodological criticisms from survey researchers:
The 11-point scale has questionable interval properties. The difference between a 6 and a 7 (which crosses the Detractor/Passive boundary) is treated as qualitatively different from the difference between a 5 and a 6, even though respondents presumably treat adjacent numbers as roughly equivalent. The categorization discards information from the raw scale. Statistically, using the raw continuous score as a predictor outperforms the categorized NPS calculation in predictive models.
Response set effects vary by culture. Research on cross-cultural survey methodology documents that respondents in different countries use rating scales differently: American respondents tend toward higher scores, European respondents toward more moderate ones, and East Asian respondents show different patterns still. This makes international NPS comparisons unreliable without cultural calibration. A Scandinavian company with an NPS of 25 may have healthier customer relationships than an American company with an NPS of 45.
The metric is highly sensitive to survey methodology. NPS measured by phone differs from NPS measured by email, which differs from NPS measured by in-app pop-up. The timing of the survey relative to the customer experience has large effects — asking immediately after a successful support interaction produces higher scores than asking a week later. These context effects make it difficult to compare scores across channels or over time if the methodology has changed.
Tying compensation to NPS encourages gaming. When bonuses, performance reviews, or team rankings depend on NPS, organizations reliably see scores improve — through survey cherry-picking, coaching customers before the survey, or selecting the survey sample to exclude likely detractors. The metric that was supposed to reveal the truth about customer experience becomes a managed number rather than a genuine signal.
What NPS Does and Does Not Predict
Understanding NPS requires distinguishing between what the metric has been shown to correlate with and what it has not.
| NPS May Predict | NPS May Not Predict |
|---|---|
| Directional trends in customer loyalty over time | Specific revenue growth better than alternatives |
| Likelihood of customer churn when scores are very low | Customer lifetime value at an individual level |
| General relationship health between company and customer | Future purchase behavior with consistent reliability |
| Employee and customer perception of brand | Actual word-of-mouth referral behavior |
| Competitive position relative to industry peers | Individual-level propensity to refer |
The last row is particularly significant. Reichheld's original hypothesis was about actual recommendation behavior — customers physically recommending the product to others. But NPS measures stated likelihood to recommend, not actual recommendation behavior. Research by Morgan and Rego (2006) found that actual referral behavior is a poor proxy for NPS scores, and that satisfied customers who give high NPS scores do not necessarily generate more referrals than moderately satisfied customers.
The gap between stated intention and actual behavior is one of the oldest findings in social psychology, sometimes called the intention-behavior gap. Customers say they would recommend a product in a survey context but do not do so in daily life — either because the opportunity does not arise, because they do not think about it, or because they have a general positive inclination without specific enthusiasm.
The Detractor-Churn Relationship
Where NPS does show more consistent predictive validity is at the low end: very low NPS scores, particularly responses of 0-3, are meaningfully associated with elevated churn rates and negative word-of-mouth. The relationship at the bottom of the scale is more reliable than the relationship at the top.
This asymmetry suggests a practically useful reframe: NPS may be more valuable as an early warning system for serious relationship damage than as a positive indicator of growth potential. A score of 9 from a customer may not reliably predict referrals; a score of 2 is a reliable signal of customer distress that warrants attention.
NPS in Practice: How Organizations Use It
Transactional vs. Relational NPS
Organizations typically implement NPS in two forms:
Transactional NPS surveys customers immediately after a specific interaction — a purchase, a support call, a service visit. It captures feedback on that specific experience. It is higher-response-rate, more diagnostic, and more actionable at the team level.
Relational NPS surveys a sample of customers periodically — typically quarterly or annually — regardless of recent interactions. It attempts to measure the overall health of the relationship rather than reaction to a specific event. It is better for tracking trends and competitive benchmarking.
The two types of NPS measure different things and should not be conflated. A high transactional NPS average can coexist with a declining relational NPS if individual interactions are handled well while structural issues are eroding the broader relationship.
Closing the Loop
One of the highest-return investments an NPS program can make is direct outreach to customers who gave scores of 0-6. Understanding their specific experience, acknowledging the problem, and attempting to resolve it has been shown to increase retention rates among at-risk customers and provides a direct source of improvement priorities. This practice — called closed-loop feedback — transforms NPS from a measurement exercise into a recovery program.
Bain and Company research suggests that successfully recovering a detractor can increase their likelihood of repurchase by up to 15 percentage points relative to detractors who are not contacted. The commercial value of recovery justifies the operational investment.
Alternatives and Complementary Metrics
The criticisms of NPS have produced a small industry of alternative metrics. The most widely adopted are:
Customer Satisfaction Score (CSAT)
CSAT measures satisfaction with a specific interaction or experience, typically using a 1-5 or 1-10 scale immediately after the interaction. It answers the question: "How satisfied were you with this interaction?"
Strengths: More specific than NPS; directly tied to particular touchpoints; widely understood; actionable at the team level; easy to segment by interaction type.
Weaknesses: Measures satisfaction in the moment, not relationship quality; satisfaction and loyalty are not the same thing; satisfaction with individual transactions can be high while overall loyalty is low; does not predict growth or referral behavior.
Customer Effort Score (CES)
Customer Effort Score was developed by the Corporate Executive Board (now Gartner) and published in a 2010 Harvard Business Review article titled "Stop Trying to Delight Your Customers." CES measures how much effort a customer had to exert to resolve an issue or complete a task, typically using a scale from "very low effort" to "very high effort."
The CEB's research found that reducing customer effort predicted customer retention better than increasing customer delight in service contexts — a counterintuitive finding that challenged the prevailing emphasis on exceeding expectations. Making it easy to cancel, return products, or resolve billing issues reduced churn more effectively than trying to provide exceptional experiences on the way in.
"The biggest driver of disloyalty is not low satisfaction. It is high effort. When customers have to work hard to get their problems solved, they switch. This is the finding that fundamentally challenged the delight hypothesis." — Dixon, Freeman, and Toman, Harvard Business Review, 2010
Strengths: Strong predictor of retention in service contexts; actionable (the question of "where did you have to work hard?" directly identifies process failures); removes subjective emotional framing.
Weaknesses: Less useful for relationship measurement; primarily applicable to service and support interactions; does not capture satisfaction with the overall product or value proposition.
First Contact Resolution (FCR) and Behavioral Metrics
Some organizations have moved toward behavioral metrics that do not require surveys at all. First contact resolution rate (whether a customer's issue was resolved on the first contact) is a strong predictor of service satisfaction. Churn rate and repeat purchase rate capture actual behavior rather than stated intent.
These behavioral metrics are harder to game than survey-based metrics, capture what customers actually do rather than what they say they might do, and require no survey infrastructure. Their limitation is that they do not explain the reasons behind the behavior — a rising churn rate tells you something is wrong but not what.
Combining Multiple Metrics
Most measurement experts now recommend using multiple customer metrics rather than any single number:
| Metric | What It Measures | Best Used For |
|---|---|---|
| NPS | Overall relationship sentiment; likelihood to recommend | Trend tracking; competitive benchmarking |
| CSAT | Satisfaction with specific interactions | Transaction-level quality feedback |
| CES | Effort required to resolve issues | Identifying service friction |
| Churn rate | Actual customer departure | Behavioral baseline for loyalty |
| Repeat purchase rate | Actual repurchase behavior | E-commerce and subscription loyalty |
| Average resolution time | Operational efficiency in support | Support team performance |
The combination approach acknowledges that no single metric captures the full complexity of customer experience, while avoiding the paralysis of tracking too many numbers simultaneously. The art is selecting the subset that best maps to how your specific business creates and destroys customer value.
NPS Industry Benchmarks
NPS scores vary dramatically by industry, making absolute score comparisons across sectors misleading. A score of 30 might be excellent in one industry and mediocre in another. Industry benchmarks also shift over time as competitive dynamics change customer expectations.
| Industry | Typical NPS Range | Notes |
|---|---|---|
| Software / SaaS | 30-60 | Wide variance by product type and segment |
| Consumer electronics | 30-50 | Apple, Samsung drive high end |
| Financial services (banking) | 20-40 | Highly regulated; trust is a primary driver |
| Airlines | 0-40 | High variance; loyalty programs inflate scores |
| Internet / cable providers | -10-20 | Structural low satisfaction; captive markets |
| Healthcare | 20-40 | Complex; varies dramatically by specialty |
| Retail | 30-50 | E-commerce typically higher than in-store |
| Streaming services | 45-65 | High due to opt-in nature of subscriptions |
| Restaurants (casual dining) | 20-50 | High variance; brand differentiation strong |
Bain and Company publishes annual industry benchmark reports that provide more granular segmentation. Comparing your NPS to industry peers is more meaningful than comparing it to an absolute scale, but even industry comparisons require controlling for survey methodology, customer segment, and geographic market. A company serving the enterprise segment of a market will typically score differently from one serving SMBs in the same space.
What "Good" Actually Looks Like
Industry context also determines what a given score means for competitive health. Satmetrix and Bain benchmarking data from 2022-2023 indicates that within the SaaS sector, scores above 50 are associated with strong competitive positioning, while scores below 20 signal meaningful loyalty risk relative to sector peers. Within cable and internet provision, however, a score of 0 might represent industry-average performance simply because the structural dynamics of that sector produce widespread customer frustration.
Best Practices for Using NPS Responsibly
Given the criticisms and the evidence, how should organizations use NPS?
Use it as a signal, not a verdict. NPS is most valuable as a directional indicator that prompts investigation, not as a precise performance measurement. A declining NPS score tells you that something has changed in the customer relationship; it does not tell you what or what to do about it.
Always follow up with the "why." The single-question format is NPS's practical advantage but its analytical weakness. Without open-ended follow-up questions asking respondents to explain their score, you know your NPS but not the reasons behind it. The follow-up question ("What is the primary reason for your score?") is where the actionable insight lives. Qualitative analysis of open-ended responses consistently provides more strategic value than the numeric score itself.
Segment your results. An overall NPS of 35 may conceal a score of 55 among enterprise customers and 15 among SMB customers — two completely different businesses in the same number. Segment by customer type, product line, channel, customer tenure, and geography to find where the relationship is strong and where it is breaking down.
Track trends, not absolutes. The comparison that matters most is your score this quarter versus last quarter, or this year versus last year. Absolute NPS values are too sensitive to methodology and cultural context to be reliable across companies; trends within a consistent methodology are more informative.
Never tie individual employee compensation directly to NPS. The research on this is unambiguous: financial incentives tied to satisfaction scores reliably inflate the scores without improving the underlying experience. A 2019 study examining NPS programs across multiple industries found that score gaming was present in approximately 40% of programs where individual compensation was tied to the metric. Use NPS at the team or division level, and pair it with behavioral metrics that are harder to game.
Maintain methodological consistency. The single most important factor for reliable NPS trends is consistency in how the survey is administered. Changing from email to in-app delivery, altering the timing of survey sends, or modifying the follow-up questions will produce score changes that reflect methodology, not customer experience. If you must change methodology, run a parallel period with both approaches to understand the impact.
Close the loop with detractors. One of the highest-return investments an NPS program can make is direct outreach to customers who gave scores of 0-6. Understanding their specific experience, acknowledging the problem, and attempting to resolve it has been shown to increase retention rates among at-risk customers and provides a direct source of improvement priorities.
The Philosophical Dimension: Why Simplicity Is Seductive
The enduring appeal of NPS is worth examining beyond its technical merits. Organizations face enormous pressure to make complex phenomena legible and comparable: how is the customer relationship doing? Is it better or worse than last year? Better or worse than competitors?
The appeal of reducing this complexity to a single number is not simply laziness. It is a genuine organizational need. Executives making capital allocation decisions, product leaders prioritizing development work, customer success teams managing accounts at scale — all of them need compressed signals they can act on. The question is not whether compression is desirable (it is) but whether a single number can compress the right information without losing what matters.
The evidence suggests NPS captures something real but imprecise. It correlates with customer relationship health in the aggregate and provides a consistent benchmark over time when measured consistently. It does not do what its most evangelical proponents claimed: it is not a reliable predictor of company-level revenue growth, it is not the best single predictor of loyalty, and it does not reliably capture actual referral behavior.
Net Promoter Score is not the one number you need to grow. It is one useful signal among several, limited by the same survey constraints as any other metric, and best used as a prompt for deeper investigation rather than a destination in itself. Used with those limitations in mind, it can be a valuable tool. Used as a substitute for genuine customer understanding, it becomes another number that people optimize rather than an experience that people improve.
References
- Reichheld, F. F. (2003). "The one number you need to grow." Harvard Business Review, 81(12), 46-54.
- Keiningham, T. L., Cooil, B., Andreassen, T. W., & Aksoy, L. (2007). "A longitudinal examination of net promoter and firm revenue growth." Journal of Marketing, 71(3), 39-51.
- Morgan, N. A., & Rego, L. L. (2006). "The value of different customer satisfaction and loyalty metrics in predicting business performance." Marketing Science, 25(5), 426-439.
- Dixon, M., Freeman, K., & Toman, N. (2010). "Stop trying to delight your customers." Harvard Business Review, 88(7/8), 116-122.
- Reichheld, F. F., & Markey, R. (2011). The Ultimate Question 2.0: How Net Promoter Companies Thrive in a Customer-Driven World. Harvard Business Review Press.
- Bain & Company. (2023). Net Promoter Score Benchmarks by Industry. Bain.com.
- Satmetrix. (2022). Net Promoter Industry Benchmarks. Satmetrix Systems.
- Schneider, M. J., et al. (2020). "Customer satisfaction and firm performance in B2B services." Journal of Business Research, 118, 451-462.
Frequently Asked Questions
What is Net Promoter Score (NPS)?
Net Promoter Score is a customer loyalty metric based on a single question: 'How likely are you to recommend this company to a friend or colleague?' Respondents answer on a 0-10 scale. Those scoring 9-10 are Promoters, 7-8 are Passives, and 0-6 are Detractors. NPS is calculated by subtracting the percentage of Detractors from the percentage of Promoters, producing a score from -100 to +100. Fred Reichheld introduced the concept in a 2003 Harvard Business Review article titled 'The One Number You Need to Grow.'
What is a good NPS score?
NPS benchmarks vary significantly by industry, making cross-company comparisons meaningful only within sectors. Generally, any positive NPS is considered acceptable, above 50 is considered excellent, and above 70 is world-class. Consumer technology companies and subscription services tend to post higher scores than utilities or healthcare. Bain and Company, which commercialized NPS, publishes annual industry benchmarks that show median scores ranging from below 20 in some industries to above 60 in others.
What are the main criticisms of NPS?
The most rigorous critique comes from a 2007 meta-analysis by Timothy Keiningham and colleagues published in the Journal of Marketing, which analyzed 21 longitudinal studies and found that NPS was not consistently superior to other satisfaction and loyalty metrics in predicting revenue growth — contradicting Reichheld's central claim. Additional criticisms include: the 11-point scale has weak interval properties; the Promoter/Passive/Detractor categorization discards information; scores are highly sensitive to survey method, timing, and market context; and the metric encourages gaming.
What are alternatives to NPS?
Customer Satisfaction Score (CSAT) measures satisfaction with a specific interaction using a simple rating scale and is better suited for transactional feedback. Customer Effort Score (CES), developed by the Corporate Executive Board in 2010, measures how much effort a customer had to exert to resolve an issue and has shown strong correlation with customer retention in service contexts. Some researchers advocate for combining multiple metrics — NPS for relationship tracking, CSAT for transaction quality, and CES for service efficiency — rather than relying on any single number.
How should companies use NPS responsibly?
NPS is most useful as a directional signal and conversation starter rather than a precise performance metric. Best practices include following up with open-ended questions to understand the 'why' behind scores, segmenting results by customer type and journey stage, tracking trends over time rather than absolute values, closing the loop with detractors through direct outreach, and never using NPS as the sole input for major strategic decisions. Tying employee compensation to NPS scores tends to produce gaming and score inflation rather than genuine loyalty improvement.