A diagnostic test is only useful if it discriminates reliably between patients with and without the condition. Before recommending a test for screening, staging, or monitoring, you need to know its sensitivity and specificity, how those change across decision thresholds, and whether it outperforms the alternatives. A test that looks adequate on AUC alone may fail at the threshold that matters clinically.
The full quantitative and qualitative evaluation in one analysis. ROC curves with DeLong AUC, head-to-head comparison of up to 10 tests, decision plots across every possible threshold, and cost-based optimisation that accounts for the clinical consequences of false positives and false negatives. For binary tests, PPA/NPA and kappa.
See diagnostic accuracy results in detail — ROC curves, AUC comparison, decision plots, and qualitative test evaluation — using example datasets you can download and follow along with.
Diagnostic accuracy evaluation is one part of the Medical edition, alongside Bland–Altman agreement, reference intervals, and survival analysis — plus the full Standard edition for hypothesis testing, regression, and descriptive statistics.
For formal method validation with Passing–Bablok, Deming regression, and CLSI protocol support, see the Method Validation edition.