Assessment |
Biopsychology |
Comparative |
Cognitive |
Developmental |
Language |
Individual differences |
Personality |
Philosophy |
Social |
Methods |
Statistics |
Clinical |
Educational |
Industrial |
Professional items |
World psychology |
Statistics: Scientific method · Research methods · Experimental design · Undergraduate statistics courses · Statistical tests · Game theory · Decision theory
Initially identified in 1978,[1] spectrum bias refers to the phenomenon that the performance of a diagnostic test may change between different clinical settings owing to changes in the patient case-mix thereby affecting the transferability of study results in clinical practice.[2] In the statistical sense it is not truly a bias and has led some authors to refer to the effect(s) as 'spectrum effects',[3] whilst others maintain it is bias if the true performance of the test differs from that which is 'expected'.[2] Usually the performance of a diagnostic test is measured in terms of its sensitivity and specificity and it is changes in these that are considered when referring to spectrum bias. However, other performance measures such as the likelihood ratios may also be affected by spectrum bias.[2]
Generally it is considered to have three causes.[2] The first is due to a change in the case-mix of those patients with the target disorder (disease) and this affects the sensitivity. The second is due to a change in the case-mix of those without the target disorder (disease-free) and this affects the specificity. The third is due to a change in the prevalence, and this affects both the sensitivity and specificity.[4] This final cause is not widely appreciated, but there is mounting empirical evidence[4][5] as well as theoretical[6] which suggest that it does indeed affect a test's performance.
Examples where the sensitivity and specificity change between different sub-groups of patients may be found with the carcinoembryonic antigen test[7] and urinary dipstick tests.[8]
Diagnostic test performances reported by some studies may be artificially overestimated if it is a case-control design where a healthy population ('fittest of the fit') is compared with a population with advanced disease ('sickest of the sick') .[9]
If properly analyzed, recognition of heterogeneity of subgroups can lead to insights about the test's performance in varying populations.[3]
See also[]
- Biased sample
- Reporting bias
- Sampling (experimental)
References[]
- ↑ Ransohoff DF, Feinstein AR (1978). Problems of spectrum and bias in evaluating the efficacy of diagnostic tests. N. Engl. J. Med. 299 (17): 926–30.
- ↑ 2.0 2.1 2.2 2.3 Willis BH, (2008). Spectrum bias—why clinicians need to be cautious when applying diagnostic test studies,. Family Practice 25: 390-396. (5): 390–6.
- ↑ 3.0 3.1 Mulherin SA, Miller WC (2002). Spectrum bias or spectrum effect? Subgroup variation in diagnostic test evaluation. Ann. Intern. Med. 137 (7): 598–602. Cite error: Invalid
<ref>
tag; name "pmid12353947" defined multiple times with different content - ↑ 4.0 4.1 Willis BH, (2012). Evidence that disease prevalence may affect the performance of diagnostic tests with an implicit threshold: a cross sectional study,. BMJ Open. 2 (1): e000746.
- ↑ Leeflang MM, Bossuyt PM, Irwig L., Diagnostic test accuracy may vary with prevalence: implications for evidence-based diagnosis, J Clin Epidemiol. 2009 Jan;62(1):5-12.
- ↑ Goehring C, Perrier A, Morabia A (2004). Spectrum bias: a quantitative and graphical analysis of the variability of medical diagnostic test performance. Statistics in medicine 23 (1): 125–35.
- ↑ Fletcher RH (1986). Carcinoembryonic antigen. Ann. Intern. Med. 104 (1): 66–73.
- ↑ Lachs MS, Nachamkin I, Edelstein PH, Goldman J, Feinstein AR, Schwartz JS (1992). Spectrum bias in the evaluation of diagnostic tests: lessons from the rapid dipstick test for urinary tract infection. Ann. Intern. Med. 117 (2): 135–40.
- ↑ Rutjes AWS, Reitsma JB, Vandenbroucke JP, Glas AS, Bossuyt PMM, Case-control and two-gate designs in diagnostic accuracy studies, Clin Chem 2005;51(8):1335-41.