Responses
Other responses
Jump to comment:
- Published on: 14 January 2021
- Published on: 14 January 2021Approaches for assessing agreement in continuous measurements in a multi-observer setup, comment on “Retrospective comparison of approaches to evaluating inter-observer variability in CT tumour measurements in an academic health centre”
We read with interest the article by Woo and colleagues, evaluating sensitivity of statistical methods for detecting different levels of interobserver variability in CT measurements of cancer lesions [1]. It is increasingly recognized that in order to evaluate the efficacy of medical imaging there is a need to conduct multi-observer studies in which proper statistical analysis is a critical component [2]. Thus, the study by Woo et al. is a welcome addition to the literature, and the authors are commended for providing open access to data.
The authors supplemented an observed dataset based on the diameters of 10 CT lesions measured by 13 observers by generating two additional datasets of increased and decreased measurement variability, respectively.
These three datasets were used to compare three statistical approaches 1) intraclass correlation coefficient (ICC), and what the authors refer to as 2) outlier counts (score) from standard Bland-Altman plotting with limits of agreement, and 3) outlier score from Bland-Altman plotting with fixed limits of agreement.We have a few comments.
We ardently agree with the authors that although the widely used ICC accommodates a multi-observer setup, it is not an ideal method for evaluating interobserver variability; the ICC reveals little about the degree of discrepancy nor supply information to investigate whether the variability may change with the magnitude of measurements (e.g. to reveal that the diameter...
Show MoreConflict of Interest:
None declared.