Academic Calculations Versus Clinical Judgments: Practicing Physicians' Use of Quantitative Measures of Test Accuracy
Overview
Authors
Affiliations
Purpose: To determine how often practicing physicians use the customarily recommended quantitative methods that include sensitivity, specificity, and likelihood ratio indexes; receiver operator characteristic (ROC) curves; and Bayesian diagnostic calculations.
Participants And Methods: A random sample of 300 practicing physicians (stratified by specialty to include family physicians, general internists, general surgeons, pediatricians, obstetrician/gynecologists, and internal medicine subspecialists) were briefly interviewed in a telephone survey. They were asked about the frequency with which they used the formal methods, the reasons for non-use, and if they employed alternative strategies when appraising tests' diagnostic accuracy.
Results: Of the 300 surveyed physicians, 8 (3%) used the recommended formal Bayesian calculations, 3 used ROC curves, and 2 used likelihood ratios. The main reasons cited for non-use included impracticality of the Bayesian method (74%), and nonfamiliarity with ROC curves and likelihood ratios (97%). Of the 174 physicians who said they used sensitivity and specificity indexes, 165 (95%) did not do so in the recommended formal manner. Instead, the physicians directly estimated tests' diagnostic accuracy by determining how often the test results were correct in groups of patients later found to have, or to be free of, the selected disease.
Conclusions: The results indicate that most practicing physicians do not use the formal recommended quantitative methods to appraise tests' diagnostic accuracy, and instead report using an alternative direct approach. Although additional training might make physicians use the formal methods more often, the physicians' direct method merits further evaluation as a potentially pragmatic tool for the determination of tests' diagnostic accuracy in clinical practice.
The accuracy of prehospital triage decisions in English trauma networks - a case-cohort study.
Fuller G, Baird J, Keating S, Miller J, Pilbery R, Kean N Scand J Trauma Resusc Emerg Med. 2024; 32(1):47.
PMID: 38773613 PMC: 11110388. DOI: 10.1186/s13049-024-01219-9.
The leaf plot: a novel way of presenting the value of tests.
Coulthard M, Coulthard T Br J Gen Pract. 2019; 69(681):205-206.
PMID: 30858331 PMC: 6428474. DOI: 10.3399/bjgp19X702113.
Artificial neural networks and risk stratification in emergency departments.
Falavigna G, Costantino G, Furlan R, Quinn J, Ungar A, Ippoliti R Intern Emerg Med. 2018; 14(2):291-299.
PMID: 30353271 DOI: 10.1007/s11739-018-1971-2.
What Evidence is There for Biochemical Testing?.
Horvath A EJIFCC. 2018; 14(3):141-151.
PMID: 30275815 PMC: 6156968.
Interactive visualisation for interpreting diagnostic test accuracy study results.
Fanshawe T, Power M, Graziadio S, Ordonez-Mena J, Simpson J, Allen J BMJ Evid Based Med. 2018; 23(1):13-16.
PMID: 29367318 PMC: 5867448. DOI: 10.1136/ebmed-2017-110862.