TypeFlow
Career Advice

How to Set Defensible Typing Test Pass/Fail Thresholds for Hiring

Setting typing test cutoffs shouldn't be guesswork. Learn the data-driven process for building defensible pass/fail thresholds that hold up to legal scrutiny and actually predict job performance.

Fred Johnson
11 min
How to Set Defensible Typing Test Pass/Fail Thresholds for Hiring

A candidate finishes your typing test at 38 WPM with 94% accuracy. Another clocks in at 52 WPM but only hits 87% accuracy. Which one passes? Which one gets rejected? And more importantly, can you defend that decision if someone challenges it?

Setting pass/fail thresholds for typing tests sounds straightforward, but get it wrong and you're looking at one of two problems. Set the bar too high, and you'll filter out qualified candidates who could do the job perfectly well. Set it too low, and you'll onboard people who slow down your entire team. Either way, you're leaving money on the table and potentially exposing your organization to legal risk.

The good news? There's a structured, data-driven approach to setting typing test cutoffs that holds up to scrutiny. It doesn't require a PhD in psychometrics, and it doesn't require guesswork. What it does require is understanding your actual job requirements, knowing the data behind your decisions, and documenting your process clearly.

This guide walks you through exactly how to build defensible pass/fail thresholds, from conducting a proper job analysis to monitoring for adverse impact after your tests go live. Whether you're hiring data entry clerks or customer service reps, the principles are the same. And when you're ready to put them into practice, TypeFlow's configurable pass criteria let you set custom WPM and accuracy thresholds for every test you create.

Why Arbitrary Cutoff Scores Create Real Problems

Let's start with the mistake most recruiters make. They Google "average typing speed," find a number like 40 WPM, round it up to something that feels professional, and plug it into their hiring process. Maybe they pick 45 WPM because it sounds better. Maybe they bump it to 50 because "we want good typists." The number feels reasonable, so they move on.

This approach has three serious flaws.

First, generic benchmarks don't reflect your specific job. A medical transcriptionist and a front desk receptionist both type, but the volume, complexity, and speed requirements are wildly different. The Bureau of Labor Statistics categorizes dozens of distinct office and administrative support roles, each with different performance expectations. A one-size-fits-all number ignores this variation entirely.

Second, arbitrary cutoffs are legally vulnerable. Under Title VII of the Civil Rights Act, any selection procedure that has an adverse impact on a protected group must be shown to be job-related and consistent with business necessity. If your cutoff disproportionately screens out candidates of a particular race, sex, age, or national origin, you need documentation showing why that specific number is necessary for the job. "It seemed about right" won't hold up. The EEOC's guidance on employment tests makes this crystal clear: employers must be prepared to demonstrate the validity and job-relatedness of any test used in hiring decisions.

Third, poorly set thresholds waste your own resources. Imagine you set a 55 WPM cutoff for a role where 40 WPM is genuinely sufficient. You just eliminated a chunk of your applicant pool for no business reason. In a tight labor market, that's candidates you'll never see again, candidates who would have performed the job just fine.

So what does a defensible process actually look like? It starts with understanding the job itself, not the test.

The Real Cost of Getting It Wrong

Consider a scenario many HR teams have faced. A company sets a 50 WPM minimum for a customer service role. Six months later, they realize their pass rate among candidates over 50 years old is significantly lower than among younger candidates. Without documentation showing why 50 WPM is necessary (not just preferred), they're exposed to an age discrimination claim. The fix isn't to lower standards. The fix is to set them correctly in the first place, anchored to actual job demands.

Another common scenario: a recruiter sets accuracy requirements at 98% because "accuracy matters." But when you observe people actually doing the job, they use autocorrect, spell-check, and templates. The real accuracy requirement for raw typing is much lower. That 98% threshold screened out candidates who would have been perfectly productive.

The pattern is the same every time. Without job analysis data backing your numbers, you're guessing. And guessing creates risk.

Building Your Threshold From Job Analysis Data

The foundation of any defensible cutoff is a job analysis. This doesn't have to be a six-month research project. It can be a focused, practical process that takes a few days and produces clear, documented results.

Here's how to do it step by step.

Step 1: Identify the Typing-Intensive Tasks

Start by listing every task in the role that requires typing. Be specific. Don't write "types documents." Instead, break it down:

  • Enters customer information into CRM during live phone calls

  • Composes 30-50 email responses per shift using templates

  • Transcribes handwritten notes into the electronic health record system

  • Logs incident reports with detailed narrative descriptions

For each task, note how much of the workday it occupies. A role where typing accounts for 80% of the day justifies higher standards than one where it's 20%. This distinction matters for legal defensibility, because the threshold needs to be proportional to the task's importance.

Step 2: Measure Incumbent Performance

This is the most powerful step most recruiters skip. Have your current top performers, average performers, and acceptable performers take the same typing test you plan to give candidates. Record their WPM and accuracy scores.

What you're looking for is the performance floor, the minimum score among employees who are doing the job successfully. Not the average. Not the top score. The minimum among acceptable performers.

Let's say you test 15 current employees in a data entry role. Their scores might look like this:

Performance Level

Number of Employees

WPM Range

Accuracy Range

Top Performers

4

58-72

96-99%

Solid Performers

7

44-57

93-97%

Acceptable Performers

4

38-46

91-95%

The lowest acceptable performer types at 38 WPM with 91% accuracy. That gives you a data-driven starting point. Setting your cutoff at 40 WPM with 92% accuracy would capture all your acceptable performers while providing a small margin. Setting it at 55 WPM would exclude people who are already doing the job well.

Step 3: Factor in Job Conditions

Raw typing speed in a quiet test environment isn't the same as typing while handling a live phone call, navigating between software systems, or working under time pressure. If the job involves multitasking that reduces effective typing speed, your test conditions should reflect that, or your cutoff should account for the difference.

Some teams add a 5-10% buffer above the incumbent floor to account for the learning curve new hires face. That's reasonable, but document why you chose that specific buffer. "We added 5 WPM because new hires spend the first month learning our software, which temporarily reduces their effective speed" is a defensible rationale.

Step 4: Document Everything

Your documentation should include the job tasks analyzed, the incumbent data collected, how you derived the cutoff number, and who was involved in the decision. Keep this in a file you can access if your threshold is ever questioned. This paper trail is what transforms a number from "something we picked" into "a validated selection criterion."

Industry Benchmarks and How to Use Them Responsibly

Job analysis should always be your primary method. But industry benchmarks can serve as a useful sanity check, especially when you don't have enough incumbents to test or when you're creating a brand-new role.

Here's a practical reference table based on common role categories:

Role Category

Typical WPM Range

Typical Accuracy Range

Key Consideration

General Administrative

40-50

90-95%

Broad task variety, moderate speed needs

Customer Service (Live Chat)

45-60

92-96%

Speed under pressure, concurrent conversations

Data Entry

50-65

95-98%

High volume, accuracy-critical

Medical Transcription

55-70

97-99%

Specialized terminology, patient safety

Legal/Court Reporting

60-80+

95-99%

Verbatim accuracy, fast dictation

General Office/Reception

35-45

88-93%

Typing is secondary to other duties

These numbers are useful starting points, but they come with an important caveat: they represent ranges observed across many organizations, not universal standards. Your specific role, workflow, and tools may justify numbers above or below these ranges.

For instance, if your customer service team uses extensive text macros and templates, actual keystroke-by-keystroke typing speed matters less. You might justifiably set a lower WPM threshold and weight accuracy more heavily. On the other hand, if your data entry team works with handwritten source documents and a legacy system with no autocomplete, raw speed and accuracy both matter more.

The key principle is that benchmarks inform your decision; they don't make it. Your cutoff should always be traceable back to the actual job requirements at your organization. If you're exploring how different industries and government agencies handle these requirements, this overview of typing test requirements for government and civil service jobs provides a useful comparison point.

Setting Dual Thresholds: WPM and Accuracy Together

One of the biggest mistakes in typing assessment is treating speed and accuracy as independent pass/fail criteria. A candidate who types 65 WPM with 82% accuracy is making roughly 12 errors per minute. In a data entry role, that error rate could be catastrophic. Meanwhile, a candidate at 45 WPM with 98% accuracy produces far more usable work.

The strongest approach is a combined threshold:

  • Minimum WPM (based on job analysis): The speed floor below which a candidate can't keep up with work volume

  • Minimum Accuracy (based on error tolerance): The accuracy floor below which errors create unacceptable rework or risk

  • Weighted Score (optional): A formula that balances both, such as Adjusted WPM = WPM × (Accuracy / 100)

Using the weighted approach, that 65 WPM / 82% candidate scores 53.3 adjusted WPM, while the 45 WPM / 98% candidate scores 44.1 adjusted WPM. Depending on whether the role values throughput or precision, either candidate could be the better fit. The point is that your scoring model should reflect what the job actually demands.

TypeFlow's test creation tools let you configure both WPM and accuracy thresholds independently, so you can implement exactly the criteria your job analysis supports.

Monitoring for Adverse Impact and Keeping Thresholds Current

Setting your threshold is not a one-time event. It's the beginning of an ongoing process. Once your tests are live and candidates are being assessed, you need to monitor the results for two things: effectiveness and fairness.

Running the Four-Fifths Rule Analysis

The EEOC uses the "four-fifths rule" as a rule of thumb for identifying adverse impact. Here's how it works:

  1. Calculate the pass rate for each demographic group (by race, sex, age, etc.)

  2. Identify the group with the highest pass rate

  3. Divide each group's pass rate by the highest group's pass rate

  4. If any group's ratio falls below 80% (four-fifths), adverse impact may exist

For example:

Group

Candidates Tested

Passed

Pass Rate

Ratio to Highest

Group A

120

96

80.0%

100% (highest)

Group B

85

59

69.4%

86.8%

Group C

60

36

60.0%

75.0% ⚠️

In this example, Group C's ratio falls below 80%, which triggers a potential adverse impact flag. This doesn't automatically mean your test is discriminatory, but it does mean you need to be prepared to demonstrate that the cutoff is job-related and consistent with business necessity.

If adverse impact exists and you can't demonstrate validity, you have several options: lower the cutoff to a point that reduces adverse impact while still meeting business needs, explore alternative assessments, or refine the job analysis to see if the threshold can be adjusted.

Reviewing and Updating Your Thresholds

Jobs change. Software gets upgraded. Workflows get streamlined. The typing demands that justified a 50 WPM cutoff three years ago might only require 40 WPM now because your team adopted new tools with predictive text. Or maybe your volume has increased and the speed requirement has gone up.

Build a review cycle into your process. A practical cadence might include:

  • Reviewing pass rates and adverse impact data quarterly

  • Re-testing a sample of incumbents annually to verify the threshold still aligns with actual job performance

  • Updating documentation whenever job duties, tools, or workflows change significantly

  • Checking that test content (the passages candidates type) remains relevant and free of cultural bias

This isn't busywork. It's how you maintain a defensible process over time. A threshold that was perfectly justified when it was set can become indefensible if the underlying job changes and you never update the standard.

Practical Tips for Small HR Teams

If you're a smaller organization without a dedicated I/O psychologist on staff, don't panic. You can still build a defensible process by following these principles:

  • Start with the job, not the test. Write down what the person actually types, how much, and how fast. This is your job analysis.

  • Test your current team. Even five incumbent scores give you better data than zero. Record them.

  • Pick a cutoff you can explain in plain English. "Our lowest acceptable performer types at 42 WPM, so we set the cutoff at 42 WPM" is powerful in its simplicity.

  • Track your pass rates by group. If you notice disparities, investigate before they become complaints.

  • Keep your documentation in one place. A single document with your job analysis, incumbent data, cutoff rationale, and review dates is all you need.

Building this kind of structured, evidence-based hiring process is exactly what platforms like TypeFlow are designed to support. From configurable test durations and pass criteria to detailed candidate performance reports with accuracy tracking and keystroke analysis, the platform gives you the tools to implement the methodology described in this guide without stitching together spreadsheets and stopwatches.


Setting typing test pass/fail thresholds isn't about finding the "right" number from a chart. It's about building a defensible process that connects your cutoff to the actual demands of the job. Start with a focused job analysis, measure your incumbents, set dual thresholds for speed and accuracy, document your rationale, and monitor for adverse impact over time.

Do this, and your typing test becomes more than a screening tool. It becomes a competitive advantage that attracts the right candidates, protects your organization, and makes better hires.

Ready to implement data-driven typing assessments with configurable pass criteria? Explore TypeFlow's plans and start building defensible tests that match your exact hiring requirements.

Try TypeFlow Free