WPM vs KPH vs CPM Typing Metrics Explained Clearly
Confused by WPM, KPH, CPM, or ten-key scores? Learn what each metric means, how to convert them safely, and how to set pass thresholds you can defend.

Typing test results can feel oddly inconsistent. One vendor reports 52 WPM, another shows 15,600 KPH, and a third hands you 260 CPM. Then a hiring manager asks, “Is this passing?” and someone else adds, “What about ten key?”
If you’ve ever had to defend a typing threshold to candidates, recruiters, or compliance, you already know the real problem: the metric is only meaningful when you know exactly how it was calculated.
This guide is vendor-neutral. You’ll learn what WPM, KPH, and CPM actually mean, how to convert between them (without fooling yourself), how ten-key scoring works, and how to set defensible pass thresholds that match the job.
A good typing standard isn’t “high.” It’s relevant, consistent, and explainable.
What WPM, KPH, and CPM Actually Measure
Let’s start by making the metrics “plain English” again.
WPM (Words Per Minute)
WPM estimates typing speed by converting characters into “words.” In most typing tests, one word is standardized to five characters, often including spaces.
A common formula looks like this:
Gross WPM = (Total characters typed ÷ 5) ÷ minutes
Net WPM = Gross WPM minus penalties for errors (or adjusted by accuracy)
Here’s the catch: vendors handle mistakes differently.
Some tests:
Count mistakes only if they remain uncorrected at the end
Penalize every incorrect character even if you backspace and fix it
Reduce WPM by multiplying by accuracy (example: 60 WPM gross × 95% accuracy = 57 net)
Takeaway: WPM is useful for comparing candidates only if you also know the accuracy rule and whether the score is gross or net.
CPM (Characters Per Minute)
CPM is raw character throughput per minute.
CPM is often simpler because it doesn’t rely on the “five characters per word” assumption.
CPM can still be gross or net depending on how errors are handled.
If you type 260 characters in one minute, that’s 260 CPM.
Takeaway: CPM is great when you care about pure throughput, but you still need error handling defined.
KPH (Keystrokes Per Hour)
KPH scales up typing activity to an hourly rate.
But here’s the big source of confusion:
Some vendors define KPH as characters per hour (every character counts once).
Others define KPH as keystrokes per hour, where corrections like backspace can count as extra keystrokes.
Those are not the same number.
Takeaway: Ask whether KPH means characters per hour or literal key presses per hour, including corrections.
Accuracy (The Metric That Changes Everything)
Speed without accuracy is like a fast cashier who rings up the wrong items.
Accuracy is usually:
Accuracy % = (Correct characters ÷ total characters) × 100
But again, “correct characters” depends on whether corrected mistakes are forgiven.
For most real jobs, you want a standard that reflects outcomes:
If an employee can catch and correct errors quickly, that’s valuable.
If the job is high-volume with little time to proofread, uncorrected errors matter more.
Takeaway: Decide whether your job rewards careful correction, or requires near-perfect first-pass entry.
Ten-Key (Numeric Keypad) Metrics
Ten-key tests measure speed and accuracy entering numeric data using the keypad.
You’ll usually see:
KPH (keystrokes per hour)
SPH (strokes per hour)
NPH (numbers per hour) or entries per hour
Errors (wrong digits, transpositions, missing decimal points)
Ten-key performance is highly job-dependent. Data entry roles can be limited by:
How often you switch between keyboard and mouse
How frequently you hit decimal points, commas, or tab
Whether you’re copying from paper, PDF, or a system screen
Takeaway: Ten-key scoring is only meaningful if the test reflects the same data patterns the job uses.
Conversions That Work and Conversions That Mislead
Conversions are helpful, but only when you match assumptions. The safest conversions are those that treat “a keystroke” as a “character.” If a vendor counts backspaces and corrections as extra keystrokes, conversions become estimates.
The basic conversions (character-based)
Use these when KPH is truly “characters per hour.”
CPM to WPM
WPM = CPM ÷ 5
Example:
275 CPM ÷ 5 = 55 WPM
WPM to CPM
CPM = WPM × 5
Example:
60 WPM × 5 = 300 CPM
CPM to KPH
KPH = CPM × 60
Example:
280 CPM × 60 = 16,800 KPH
WPM to KPH
Because WPM × 5 = CPM, you can combine steps:
KPH = WPM × 5 × 60
KPH = WPM × 300
Examples:
45 WPM → 45 × 300 = 13,500 KPH
60 WPM → 60 × 300 = 18,000 KPH
The conversion table (quick reference)
WPM | CPM | KPH |
35 | 175 | 10,500 |
40 | 200 | 12,000 |
45 | 225 | 13,500 |
50 | 250 | 15,000 |
55 | 275 | 16,500 |
60 | 300 | 18,000 |
65 | 325 | 19,500 |
70 | 350 | 21,000 |
Takeaway: If you only remember one thing, remember KPH = WPM × 300 (when using character-based KPH).
Where conversions go wrong
Conversions mislead when any of these change across vendors:
Five characters per word assumption
Most tests use five characters as the “word,” but some incorporate spaces differently.
Error handling
If Vendor A forgives corrected errors and Vendor B penalizes them, two candidates can “convert” to the same WPM but represent different real performance.
What counts as a keystroke
If backspace counts as a keystroke, a careful proofreader can show a much higher KPH than someone who types cleanly. That can be unfair unless the job values on-the-fly correction.
Text complexity
Typing “the quick brown fox” is not the same as typing medication names, legal clauses, or customer IDs with hyphens.
Takeaway: Convert only after you confirm: input type, length, allowed corrections, and whether the speed is gross or net.
A defensible way to normalize across vendors
If you must compare results from different tools, standardize your interpretation:
Step 1: Decide your “core metric.”
For general typing roles: Net WPM with accuracy reported
For heavy numeric roles: Ten-key KPH/SPH with accuracy and error count
Step 2: Convert everything into that core metric using the character-based formulas.
Step 3: Add a rule that prevents “fast but sloppy” from passing.
Example: “Pass requires speed threshold and accuracy threshold.”
For a deeper framework on making scores comparable, see Typing Test Score Normalization for Consistent Hiring Decisions.
Takeaway: Normalization is not just math, it’s policy. Write down the assumptions so you can defend them.
What Is a Good Typing Speed for Different Jobs
“Good” typing speed depends on what the job punishes.
Some roles punish slow throughput (queues, backlogs, service levels).
Some punish errors (compliance, billing, medical records).
Some punish inconsistency (high variability means you can’t forecast productivity).
Below are practical ranges you can use as a starting point, then validate with job reality.
General admin and office support
Typical tasks: emails, documents, scheduling notes, basic data entry.
A reasonable standard often looks like:
40 to 55 WPM
95%+ accuracy
Scenario: A coordinator types meeting notes and client emails. Speed matters, but clarity matters more.
A defensible pass rule:
Pass if net WPM ≥ 45 and accuracy ≥ 95%
Takeaway: For general office roles, a moderate WPM with strong accuracy is usually enough.
Customer support and chat-based roles
Typical tasks: live chat responses, ticket updates, CRM notes.
What changes:
Candidates may rely on templates, but they still need fast free-typing.
Mistakes can frustrate customers or cause miscommunication.
Starting ranges:
45 to 65 WPM
95%+ accuracy
Scenario: A chat agent handles two concurrent conversations. The slow typist creates dead air.
A defensible pass rule:
Pass if net WPM ≥ 55 and accuracy ≥ 95%
Takeaway: For chat-heavy work, push speed higher while keeping accuracy stable.
Data entry (alphanumeric)
Typical tasks: IDs, addresses, inventory counts, form fields.
Important nuance:
Many data entry tasks are constrained by reading and system navigation, not only typing.
Tests that are pure paragraph typing can overestimate job performance.
Starting ranges:
45 to 70 WPM depending on complexity
97%+ accuracy for high-stakes records
Scenario: A warehouse admin enters item codes and quantities all day. One digit wrong creates returns and rework.
A defensible pass rule:
Pass if net WPM ≥ 55 and accuracy ≥ 97%
Takeaway: In data entry, accuracy often saves more time than raw speed because rework is expensive.
Medical and legal typing
Typical tasks: terminology, long-form notes, structured documentation, dictated edits.
What changes:
Terminology increases cognitive load.
Compliance risk makes errors costly.
Starting ranges:
55 to 80 WPM for experienced roles
98%+ accuracy for critical documentation
Scenario: A medical scribe types names and dosage details. One small error can cause a major incident.
A defensible pass rule:
Pass if net WPM ≥ 60 and accuracy ≥ 98%
Takeaway: In regulated work, set accuracy high and keep the test content realistic.
Ten-key heavy roles (AP, AR, payroll, claims)
Typical tasks: invoices, check numbers, payment amounts, dates, codes.
Ten-key standards vary a lot by test style, but common pass bands are:
8,000 to 12,000+ KPH with low errors for entry-level numeric work
12,000 to 15,000+ KPH for high-volume roles
Scenario: An accounts payable clerk keys invoice totals all day. A single wrong digit can trigger payment errors.
A defensible pass rule:
Pass if KPH ≥ 10,000 and errors ≤ 1 per test (or accuracy ≥ 98%)
Takeaway: For ten-key, combine a speed floor with an error cap. Small error rates can still be unacceptable.
A note on disability, language, and fairness
Typing tests can unintentionally screen out strong workers who:
Use assistive technology
Have a different primary language
Are excellent with job software shortcuts but slower at free typing
If you’re using tests in hiring, document the job need and provide reasonable accommodations.
For practical guidance on aligning standards to the role, read Set Fair Role Based Typing Test Pass Scores.
Takeaway: Fair standards focus on job outcomes, not “typing bragging rights.”
How to Set Defensible Pass Thresholds Step by Step
A pass threshold should survive three conversations:
With a hiring manager who wants “the best.”
With a candidate who asks “why did I fail?”
With a compliance reviewer who asks “is this job-related?”
Here’s a practical process you can follow.
Step 1: Define the typing job, not the title
Start with the day-to-day output.
Ask:
What percentage of the day involves typing?
Is the output mostly paragraphs, short entries, or numeric fields?
Which errors are unacceptable (names, amounts, IDs)?
Do employees have time to proofread?
Write a one-paragraph “typing profile.” Example:
“Role requires frequent CRM notes and live chat responses. Errors cause miscommunication. Speed affects response time. Candidate must type quickly with high accuracy.”
Takeaway: A threshold is defensible when it maps to the actual workflow.
Step 2: Choose one primary metric and one guardrail
Don’t overcomplicate the scorecard.
Good pairs:
Net WPM + Accuracy % (most roles)
Ten-key KPH + Error cap (numeric roles)
Avoid:
Only speed, no accuracy rule
Multiple speed metrics at once (WPM + CPM + KPH) unless you’re translating across vendors
Takeaway: Pick a main metric, then add one rule that blocks sloppy passing.
Step 3: Run a short internal benchmark
You don’t need hundreds of data points. You need a reality check.
Run the same test on:
3 to 10 solid performers (people you would rehire)
1 to 3 new hires (if available)
Capture:
Speed
Accuracy
Notes about test realism (too easy, too hard, weird vocabulary)
Then set your initial pass threshold using a simple approach:
Set the pass line slightly below the average of solid performers, as long as it still protects quality.
Example:
Solid performers average: 58 net WPM at 97% accuracy
Pass threshold: 52 net WPM and 96% accuracy
Takeaway: Benchmarking turns “opinions” into a standard you can explain.
Step 4: Build a two-tier policy (recommended)
Instead of one harsh cutoff, use two tiers.
Example for a support role:
Fail: < 45 net WPM or < 94% accuracy
Pass: ≥ 45 net WPM and ≥ 94% accuracy
Strong pass: ≥ 60 net WPM and ≥ 97% accuracy
This helps you:
Keep hiring moving when the market is tight
Still highlight top candidates
Avoid rejecting people who can succeed with training
Takeaway: Two tiers reduce bad rejections and give managers more signal.
Step 5: Make the test hard to game, not hard to take
Candidates will try to paste, switch tabs, or rehearse the exact prompt.
A defensible testing process includes:
Time limits that reflect real work pace
Multiple attempts if the role allows learning and improvement
Monitoring that flags suspicious behavior
If you’re running typing tests at scale, TypeFlow includes a real-time engine with accuracy tracking, plus security monitoring like tab switch and paste detection. That makes results easier to trust and easier to defend. If you want to standardize your pass rules and share tests with candidates, start with Set Fair Role Based Typing Test Pass Scores.
Takeaway: The best threshold is pointless if candidates can cheat their way past it.
Step 6: Document your standard in plain language
Write a short policy you can reuse in job postings and candidate emails.
Template:
“This role includes frequent typing. We assess (metric) and require (threshold). We also require (accuracy/error rule). This helps us predict performance in the day-to-day work.”
Example:
“We require at least 55 net WPM with 95% accuracy. This role involves live customer chat where speed and accuracy both affect customer experience.”
Takeaway: If you can’t explain the threshold simply, it’s harder to defend.
Step 7: Review outcomes and adjust carefully
After hiring, look for patterns:
Are people who barely passed struggling?
Are people who failed still being hired and succeeding?
Are certain groups disproportionately failing due to test mismatch?
Adjust only when you have a clear reason:
The test content doesn’t match the job
The threshold is too high or too low relative to performance
The scoring method changed
Takeaway: Thresholds should evolve with evidence, not pressure.
If you’re tired of comparing apples to oranges across typing vendors, simplify your approach: choose a primary metric, define how errors count, convert only with clear assumptions, and benchmark against real performers. That’s how you set pass thresholds you can stand behind.
If you want a clean way to create role-specific typing tests, share a single link with candidates, and get results you can actually defend, use the guidance above to write your standard, then apply it consistently.
Recommended Reading
How to Prepare for an Employment Typing Test and Pass
Got an employment typing test coming up? Learn exactly what WPM and accuracy benchmarks employers expect, how to practice effectively, and proven test day strategies to pass with confidence.
How Recruiters Can Detect AI Cheating in Remote Typing Tests
Voice assistants, second screens, and scripted bots are undermining remote typing assessments. Learn how keystroke analysis and behavioral monitoring catch AI-assisted cheating before it corrupts your hiring pipeline.
Skills-Based Hiring Playbook Using Typing Tests as Proof of Ability
Replace degree requirements with objective typing assessments. This playbook shows business leaders how to implement skills-based hiring with measurable benchmarks and data-driven decisions.
Copy Paste Detection for Secure Remote Typing Tests
Copy/paste is only one clue. Learn a practical integrity model for remote typing tests using tab switches, focus loss, typing rhythm, and clear thresholds.