Pathologists should probably forget about kappa. Percent agreement, diagnostic specificity and related metrics provide more clinically applicable measures of interobserver variability
2
Publication type: Journal Article
Publication date: 2020-08-01
scimago Q2
wos Q3
SJR: 0.571
CiteScore: 3.2
Impact factor: 1.4
ISSN: 10929134, 15328198
PubMed ID:
32623312
General Medicine
Pathology and Forensic Medicine
Abstract
Kappa statistics have been widely used in the pathology literature to compare interobserver diagnostic variability (IOV) among different pathologists but there has been limited discussion about the clinical significance of kappa scores. Five representative and recent pathology papers were queried using clinically relevant specific questions to learn how IOV was evaluated and how the clinical applicability of results was interpreted. The papers supported our anecdotal impression that pathologists usually assess IOV using Cohen's or Fleiss' kappa statistics and interpret the results using some variation of the scale proposed by Landis and Koch. The papers did not cite or propose specific guidelines to comment on the clinical applicability of results. The solutions proposed to decrease IOV included the development of better diagnostic criteria and additional educational efforts, but the possibility that the entities themselves represented a continuum of morphologic findings rather than distinct diagnostic categories was not considered in any of the studies. A dataset from a previous study of IOV reported by Thunnissen et al. was recalculated to estimate percent agreement among 19 international lung pathologists for the diagnosis of 74 challenging lung neuroendocrine neoplasms. Kappa scores and diagnostic sensitivity, specificity, positive and negative predictive values were calculated using the majority consensus diagnosis for each case as the gold reference diagnosis for that case. Diagnostic specificity estimates among multiple pathologists were > 90%, although kappa scores were considerably more variable. We explain why kappa scores are of limited clinical applicability in pathology and propose the use of positive and negative percent agreement and diagnostic specificity against a gold reference diagnosis to evaluate IOV among two and multiple raters, respectively.
Found
Nothing found, try to update filter.
Found
Nothing found, try to update filter.
Top-30
Journals
|
1
2
|
|
|
International Journal of Legal Medicine
2 publications, 8.7%
|
|
|
Frontiers in Molecular Biosciences
1 publication, 4.35%
|
|
|
Virchows Archiv
1 publication, 4.35%
|
|
|
BMC Pulmonary Medicine
1 publication, 4.35%
|
|
|
Lung Cancer
1 publication, 4.35%
|
|
|
Saudi Dental Journal
1 publication, 4.35%
|
|
|
Veterinary Immunology and Immunopathology
1 publication, 4.35%
|
|
|
medRxiv : the preprint server for health sciences
1 publication, 4.35%
|
|
|
bioRxiv
1 publication, 4.35%
|
|
|
Journal of Thoracic Oncology
1 publication, 4.35%
|
|
|
Journal of Pathology Informatics
1 publication, 4.35%
|
|
|
GeroScience
1 publication, 4.35%
|
|
|
Modern Pathology
1 publication, 4.35%
|
|
|
Gait and Posture
1 publication, 4.35%
|
|
|
Anaesthesia, critical care & pain medicine
1 publication, 4.35%
|
|
|
Toxicologic Pathology
1 publication, 4.35%
|
|
|
Journal of Clinical Medicine
1 publication, 4.35%
|
|
|
Resuscitation Plus
1 publication, 4.35%
|
|
|
European Journal of Psychology Open
1 publication, 4.35%
|
|
|
Health Information Management Journal
1 publication, 4.35%
|
|
|
Forensic Science International
1 publication, 4.35%
|
|
|
1
2
|
Publishers
|
1
2
3
4
5
6
7
8
9
|
|
|
Elsevier
9 publications, 39.13%
|
|
|
Springer Nature
5 publications, 21.74%
|
|
|
Cold Spring Harbor Laboratory
2 publications, 8.7%
|
|
|
SAGE
2 publications, 8.7%
|
|
|
Frontiers Media S.A.
1 publication, 4.35%
|
|
|
King Saud University
1 publication, 4.35%
|
|
|
MDPI
1 publication, 4.35%
|
|
|
Hogrefe Publishing Group
1 publication, 4.35%
|
|
|
1
2
3
4
5
6
7
8
9
|
- We do not take into account publications without a DOI.
- Statistics recalculated weekly.
Are you a researcher?
Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
23
Total citations:
23
Citations from 2024:
12
(52.17%)
Cite this
GOST |
RIS |
BibTex
Cite this
GOST
Copy
Marchevsky A. M. et al. Pathologists should probably forget about kappa. Percent agreement, diagnostic specificity and related metrics provide more clinically applicable measures of interobserver variability // Annals of Diagnostic Pathology. 2020. Vol. 47. p. 151561.
GOST all authors (up to 50)
Copy
Marchevsky A. M., Walts A. E., Witte B. I., Thunnissen E. Pathologists should probably forget about kappa. Percent agreement, diagnostic specificity and related metrics provide more clinically applicable measures of interobserver variability // Annals of Diagnostic Pathology. 2020. Vol. 47. p. 151561.
Cite this
RIS
Copy
TY - JOUR
DO - 10.1016/j.anndiagpath.2020.151561
UR - https://doi.org/10.1016/j.anndiagpath.2020.151561
TI - Pathologists should probably forget about kappa. Percent agreement, diagnostic specificity and related metrics provide more clinically applicable measures of interobserver variability
T2 - Annals of Diagnostic Pathology
AU - Marchevsky, Alberto M.
AU - Walts, Ann E.
AU - Witte, Birgit I.
AU - Thunnissen, Erik
PY - 2020
DA - 2020/08/01
PB - Elsevier
SP - 151561
VL - 47
PMID - 32623312
SN - 1092-9134
SN - 1532-8198
ER -
Cite this
BibTex (up to 50 authors)
Copy
@article{2020_Marchevsky,
author = {Alberto M. Marchevsky and Ann E. Walts and Birgit I. Witte and Erik Thunnissen},
title = {Pathologists should probably forget about kappa. Percent agreement, diagnostic specificity and related metrics provide more clinically applicable measures of interobserver variability},
journal = {Annals of Diagnostic Pathology},
year = {2020},
volume = {47},
publisher = {Elsevier},
month = {aug},
url = {https://doi.org/10.1016/j.anndiagpath.2020.151561},
pages = {151561},
doi = {10.1016/j.anndiagpath.2020.151561}
}