Statistical inference for agreement between multiple raters on a binary scale

Sophie Vanbelle*

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Agreement studies often involve more than two raters or repeated measurements. In the presence of two raters, the proportion of agreement and of positive agreement are simple and popular agreement measures for binary scales. These measures were generalized to agreement studies involving more than two raters with statistical inference procedures proposed on an empirical basis. We present two alternatives. The first is a Wald confidence interval using standard errors obtained by the delta method. The second involves Bayesian statistical inference not requiring any specific Bayesian software. These new procedures show better statistical behaviour than the confidence intervals initially proposed. In addition, we provide analytical formulas to determine the minimum number of persons needed for a given number of raters when planning an agreement study. All methods are implemented in the R package simpleagree and the Shiny app simpleagree.
Original languageEnglish
Pages (from-to)245-260
Number of pages16
JournalBritish Journal of Mathematical & Statistical Psychology
Volume77
Issue number2
Early online date1 Jan 2024
DOIs
Publication statusE-pub ahead of print - 1 Jan 2024

Keywords

  • confidence interval
  • credibility interval
  • dichotomous
  • raters
  • sample size
  • CONFIDENCE-INTERVALS
  • KAPPA
  • RELIABILITY
  • ASSOCIATION
  • MODEL

Cite this