Designing studies to ensure that estimates of test accuracy are transferableBMJ 2002; 324 doi: http://dx.doi.org/10.1136/bmj.324.7338.669 (Published 16 March 2002) Cite this as: BMJ 2002;324:669
- Les Irwig, professor (firstname.lastname@example.org)a,
- Patrick Bossuyt, professor of clinical epidemiologyb,
- Paul Glasziou, professor of evidence based practicec,
- Constantine Gatsonis, professord,
- Jeroen Lijmer, clinical researcherb
- a Screening and Test Evaluation Program, Department of Public Health and Community Medicine, University of Sydney, NSW 2006, Australia
- b Department of Clinical Epidemiology and Biostatistics, Academic Medical Centre, PO Box 22700, 1100 DE Amsterdam, Netherlands
- c School of Population Health, University of Queensland Medical School, Herston, QLD 4006, Australia
- d Center for Statistical Sciences, Brown University, Providence, RI 02192, USA
This is the third in a series of five articles
Measures of test accuracy are often thought of as fixed characteristics determinable by research and then applicable in practice. Yet even when tests are evaluated in a study of adequate quality—one including such features as consecutive patients, a good reference standard, and independent, blinded assessments of tests and the reference standard1—performance of a diagnostic test in one setting may vary significantly from the results reported elsewhere.2–8 In this paper, we explore the reasons for this variability and its implications for the design of studies of diagnostic tests.
Test accuracy may vary considerably from one setting to another
This may be due to the target condition, the clinical problem, what other tests have been done, or how the test is carried out
Larger studies than those usually done for diagnostic tests will be needed to assess transferability of results
These studies should explore the extent to which variation in test accuracy between populations can be explained by patient and test features
True variability in test accuracy
To interpret a test's results in different setting requires an understanding of whether and why the test's accuracy varies. Broadly speaking, measures of accuracy fall into two broad categories: measures of discrimination between people who are and who are not diseased, and measures of prediction used to estimate post-test probability of disease.
Measures of discrimination
Global measures of test accuracy assess only the ability of the test to discriminate between people with and without a disease. Common examples are the area under the receiver operating characteristic curve (ROC), and the odds ratio (OR), sometimes also referred to as the diagnostic odds ratio. Such results may suffice for some broad health policy decisions—for example, to decide whether a new test is in general better than an existing test for …
Log in using your username and password
Log in through your institution
Register for a free trial to thebmj.com to receive unlimited access to all content on thebmj.com for 14 days.
Sign up for a free trial