Quick Definition
eNPS, or Employee Net Promoter Score, is a single-question metric that asks employees how likely they are to recommend their employer as a place to work, on a 0–10 scale. It's calculated as the percentage of promoters minus the percentage of detractors and is widely used as a fast pulse on engagement.
What Is eNPS?
eNPS — Employee Net Promoter Score — is a single-question metric adapted from the customer Net Promoter Score (NPS) introduced by Fred Reichheld and Bain & Company. The question is always the same: "On a scale of 0 to 10, how likely are you to recommend [company] as a place to work?"
Responses sort into three buckets: promoters (9–10), passives (7–8), and detractors (0–6). The score is the percentage of promoters minus the percentage of detractors, expressed as a number between −100 and +100. Most organizations use eNPS as the anchor metric inside a regular pulse survey because it's fast to collect, easy to compare across periods, and surfaces a usable signal even with small response volumes.
How to Calculate eNPS
The formula is simple:
eNPS = % Promoters (9–10) − % Detractors (0–6)
Worked example: 100 employees respond. 50 score 9 or 10 (promoters), 30 score 7 or 8 (passives), and 20 score 0 through 6 (detractors). Promoters are 50%; detractors are 20%; passives are excluded. eNPS = 50 − 20 = +30.
The scale runs from −100 (every respondent is a detractor) to +100 (every respondent is a promoter). Track the trend at the company level, then segment by department, manager, tenure, and location. A company-wide score of +25 can hide a −10 inside a single struggling team — and that's where the most actionable insight usually lives.
Why eNPS Matters
eNPS gives leadership a short, repeatable read on how employees feel about the organization. It correlates loosely with retention and referral activity, and the trend line over time is more useful than any single snapshot. A score that's been climbing for four quarters tells a different story than a flat or declining one, even at the same absolute level.
Because eNPS is a single number, it's also useful for executive dashboards, board reporting, and goal-setting. But it works best as the anchor metric in a broader engagement program — paired with driver questions on recognition, manager quality, and growth, plus open-ended responses that explain why the score is moving. eNPS tells you what; the rest of the survey tells you why.
How to Improve Your eNPS
- Start with the detractors. Moving a 5 to a 7 doesn't help your score, but moving a 5 to an 8 does. Detractors are usually concentrated by team and manager, which makes them addressable.
- Build a consistent recognition practice. Lack of recognition is one of the most common detractor drivers and one of the cheapest to fix.
- Invest in manager quality. Manager relationships explain more variance in eNPS than almost any other variable.
- Close the feedback loop. When employees see action on the last survey, they engage more honestly with the next one — and detractor counts drop.
- Focus on employee experience moments that matter. Onboarding, work anniversaries, and milestone moments disproportionately move eNPS.
- Cross-reference with employee morale data. Falling morale shows up in eNPS, but earlier in qualitative pulse responses.
Benefits and Limitations of eNPS
eNPS is popular for good reasons, but it's not a complete engagement metric on its own.
- Benefit: speed. One question, two-minute survey, fast trend lines.
- Benefit: comparability. Same metric across teams, departments, and quarters.
- Benefit: simplicity. Easy to explain to executives and easy to track in dashboards.
- Limitation: shallow. The score doesn't tell you why employees feel the way they do.
- Limitation: cultural variability. Some cultures rate everything lower; cross-country comparisons can mislead without normalization.
- Limitation: gaming. If managers are evaluated on eNPS, they have an incentive to influence the survey rather than the underlying experience.
Common Pitfalls (and How to Avoid Them)
- Treating eNPS as the goal. The score is a thermometer, not a thermostat. Don't manage to the number; manage the underlying experience.
- Over-segmenting. Small teams produce noisy scores. Aggregate to a minimum group size of 5–10 to keep results meaningful.
- Anchor changes. Tweaking the question wording breaks the trend line. Lock the anchor for at least 12 months.
- Reporting the score without context. Share the number alongside what's moving, why, and what you'll do about it.
- Ignoring the comments. The numerical score is half the value. Pair every eNPS report with a representative sample of qualitative feedback.
Frequently Asked Questions
What is eNPS in simple terms?
eNPS, or Employee Net Promoter Score, is a single-question survey that asks employees how likely they are to recommend their company as a place to work on a 0–10 scale. The result is a single number that tracks engagement over time and is fast to collect and easy to compare across periods.
How is eNPS calculated?
Subtract the percentage of detractors (employees who answered 0–6) from the percentage of promoters (employees who answered 9–10). Passives (7–8) are excluded from the calculation. Formula: eNPS = % Promoters − % Detractors. The score ranges from −100 to +100.
What is a good eNPS score?
Scores above 0 mean more promoters than detractors. Anything above +20 is generally considered good, +50 is excellent, and +70+ is world-class. Negative scores signal a workforce more likely to discourage candidates than recommend the company. Benchmarks vary by industry, so trend matters more than the absolute number.
What is the difference between eNPS and engagement?
Engagement is the broader concept — how connected, motivated, and committed employees are to their work and employer. eNPS is one specific metric used to track engagement over time. eNPS is fast and simple but doesn't tell you why the score is what it is; that's what driver questions in a pulse survey are for.
How often should you measure eNPS?
Most organizations measure eNPS monthly or quarterly inside a pulse survey. Annual measurement is too slow to spot emerging issues. Weekly is usually overkill and drives response rate decay. The cadence should match how often you can act on the result.