False Positives Explained
A false positive occurs when a test incorrectly indicates a condition is present when it is actually absent. In medical testing, this means a healthy person receives a positive (disease-detected) result. The rate of false positives depends on the test's specificity and the prevalence of the condition in the tested population.
False positives have real consequences: unnecessary medical procedures, wasted resources, psychological distress, and in judicial contexts, wrongful convictions. Understanding and minimizing false positive rates is crucial in diagnostics, security screening, and spam filtering.
Key Formulas
Metrics Table
| Metric | Formula | Meaning |
|---|---|---|
| PPV | TP/(TP+FP) | If positive, probability of truly having condition |
| NPV | TN/(TN+FN) | If negative, probability of truly not having it |
| FPR | FP/(FP+TN) | Rate of false alarms among negatives |
| FDR | FP/(FP+TP) | Proportion of positives that are false |
The Base Rate Effect
Even with 95% sensitivity and 95% specificity, if prevalence is only 1%, the PPV is just 16.1%. Most positive results would be false! This counterintuitive result shows why rare-condition screening produces many false positives, and why confirmatory testing is essential.
Frequently Asked Questions
How can a 95% accurate test be wrong so often?
When a condition is rare, the 5% false positive rate applied to the large healthy population generates more false positives than the 95% true positive rate applied to the small affected population. This is the base rate fallacy.
How to reduce false positives?
Increase specificity, test only high-risk populations (higher prevalence), or use sequential testing where a second independent test confirms initial positives.
False positive vs false discovery rate?
FPR is the rate of false alarms among truly negative cases. FDR is the proportion of all positive results that are actually false. FDR is often more relevant to the patient or user who receives a positive result.