[HTML][HTML] Measuring inter-rater reliability for nominal data–which coefficients and confidence intervals are appropriate?

A Zapf, S Castell, L Morawietz, A Karch - BMC medical research …, 2016 - Springer
Background Reliability of measurements is a prerequisite of medical research. For nominal
data, Fleiss' kappa (in the following labelled as Fleiss' K) and Krippendorff's alpha provide …

Interrater agreement statistics under the two-rater dichotomous-response case with correlated decisions

Z Tian, VM Chinchilli, C Shen, S Zhou - arXiv preprint arXiv:2402.08069, 2024 - arxiv.org
Measurement of the interrater agreement (IRA) is critical in various disciplines. To correct for
potential confounding chance agreement in IRA, Cohen's kappa and many other methods …

An Exact Bootstrap Confidence Interval for κ in Small Samples

N Klar, SR Lipsitz, M Parzen… - Journal of the Royal …, 2002 - academic.oup.com
Agreement between a pair of raters for binary outcome data is typically assessed by using
the κ-coefficient. When the total sample size is small to moderate, and the proportion of …

Assessing agreement with relative area under the coverage probability curve

HX Barnhart - Statistics in Medicine, 2016 - Wiley Online Library
There has been substantial statistical literature in the last several decades on assessing
agreement, and coverage probability approach was selected as a preferred index for …

Asymptotic variability of (multilevel) multirater kappa coefficients

S Vanbelle - Statistical methods in medical research, 2019 - journals.sagepub.com
Agreement studies are of paramount importance in various scientific domains. When several
observers classify objects on categorical scales, agreement can be quantified through …

Exact one-sided confidence limits for Cohen's kappa as a measurement of agreement

G Shan, W Wang - Statistical methods in medical research, 2017 - journals.sagepub.com
Cohen's kappa coefficient, κ, is a statistical measure of inter-rater agreement or inter-
annotator agreement for qualitative items. In this paper, we focus on interval estimation of κ …

Central tendency and matched difference approaches for assessing interrater agreement.

MJ Burke, A Cohen, E Doveh… - Journal of Applied …, 2018 - psycnet.apa.org
In Study 1 of this two-part investigation, we present a “central tendency approach” and
procedures for assessing overall interrater agreement across multiple groups. We define …

Simultaneous estimation of intrarater and interrater agreement for multiple raters under order restrictions for a binary trait

H Lester Kirchner, JH Lemke - Statistics in medicine, 2002 - Wiley Online Library
It is valuable in many studies to assess both intrarater and interrater agreement. Most
measures of intrarater agreement do not adjust for unequal estimates of prevalence …

[HTML][HTML] Information-based measure of disagreement for more than two observers: a useful tool to compare the degree of observer disagreement

T Henriques, L Antunes, J Bernardes, M Matias… - BMC medical research …, 2013 - Springer
Background Assessment of disagreement among multiple measurements for the same
subject by different observers remains an important problem in medicine. Several measures …

Bivariate modeling of interobserver agreement coefficients

MM Shoukri, A Donner - Statistics in Medicine, 2009 - Wiley Online Library
Many studies of interobserver agreement yield outcome measures on more than one binary
trait. For example, Becker et al.(Int. J. Psychiatry Med. 2002; 32 (3): 271–283) reported on a …