[HTML][HTML] Measuring inter-rater reliability for nominal data–which coefficients and confidence intervals are appropriate?
Background Reliability of measurements is a prerequisite of medical research. For nominal
data, Fleiss' kappa (in the following labelled as Fleiss' K) and Krippendorff's alpha provide …
data, Fleiss' kappa (in the following labelled as Fleiss' K) and Krippendorff's alpha provide …
Interrater agreement statistics under the two-rater dichotomous-response case with correlated decisions
Measurement of the interrater agreement (IRA) is critical in various disciplines. To correct for
potential confounding chance agreement in IRA, Cohen's kappa and many other methods …
potential confounding chance agreement in IRA, Cohen's kappa and many other methods …
An Exact Bootstrap Confidence Interval for κ in Small Samples
N Klar, SR Lipsitz, M Parzen… - Journal of the Royal …, 2002 - academic.oup.com
Agreement between a pair of raters for binary outcome data is typically assessed by using
the κ-coefficient. When the total sample size is small to moderate, and the proportion of …
the κ-coefficient. When the total sample size is small to moderate, and the proportion of …
Assessing agreement with relative area under the coverage probability curve
HX Barnhart - Statistics in Medicine, 2016 - Wiley Online Library
There has been substantial statistical literature in the last several decades on assessing
agreement, and coverage probability approach was selected as a preferred index for …
agreement, and coverage probability approach was selected as a preferred index for …
Asymptotic variability of (multilevel) multirater kappa coefficients
S Vanbelle - Statistical methods in medical research, 2019 - journals.sagepub.com
Agreement studies are of paramount importance in various scientific domains. When several
observers classify objects on categorical scales, agreement can be quantified through …
observers classify objects on categorical scales, agreement can be quantified through …
Exact one-sided confidence limits for Cohen's kappa as a measurement of agreement
G Shan, W Wang - Statistical methods in medical research, 2017 - journals.sagepub.com
Cohen's kappa coefficient, κ, is a statistical measure of inter-rater agreement or inter-
annotator agreement for qualitative items. In this paper, we focus on interval estimation of κ …
annotator agreement for qualitative items. In this paper, we focus on interval estimation of κ …
Central tendency and matched difference approaches for assessing interrater agreement.
In Study 1 of this two-part investigation, we present a “central tendency approach” and
procedures for assessing overall interrater agreement across multiple groups. We define …
procedures for assessing overall interrater agreement across multiple groups. We define …
Simultaneous estimation of intrarater and interrater agreement for multiple raters under order restrictions for a binary trait
H Lester Kirchner, JH Lemke - Statistics in medicine, 2002 - Wiley Online Library
It is valuable in many studies to assess both intrarater and interrater agreement. Most
measures of intrarater agreement do not adjust for unequal estimates of prevalence …
measures of intrarater agreement do not adjust for unequal estimates of prevalence …
[HTML][HTML] Information-based measure of disagreement for more than two observers: a useful tool to compare the degree of observer disagreement
T Henriques, L Antunes, J Bernardes, M Matias… - BMC medical research …, 2013 - Springer
Background Assessment of disagreement among multiple measurements for the same
subject by different observers remains an important problem in medicine. Several measures …
subject by different observers remains an important problem in medicine. Several measures …
Bivariate modeling of interobserver agreement coefficients
MM Shoukri, A Donner - Statistics in Medicine, 2009 - Wiley Online Library
Many studies of interobserver agreement yield outcome measures on more than one binary
trait. For example, Becker et al.(Int. J. Psychiatry Med. 2002; 32 (3): 271–283) reported on a …
trait. For example, Becker et al.(Int. J. Psychiatry Med. 2002; 32 (3): 271–283) reported on a …