Statistical inference of Gwet's AC1 coefficient for multiple raters and binary outcomes

T Ohyama - Communications in Statistics-Theory and Methods, 2021 - Taylor & Francis
Cohen's kappa and intraclass kappa are widely used for assessing the degree of agreement
between two raters with binary outcomes. However, many authors have pointed out its …

Statistical inference of agreement coefficient between two raters with binary outcomes

T Ohyama - Communications in Statistics-Theory and Methods, 2020 - Taylor & Francis
Scott's pi and Cohen's kappa are widely used for assessing the degree of agreement
between two raters with binary outcomes. However, many authors have pointed out its …

[HTML][HTML] Gwet's AC1 is not a substitute for Cohen's kappa–A comparison of basic properties

W Vach, O Gerke - MethodsX, 2023 - Elsevier
Gwet's AC1 has been proposed as an alternative to Cohen's kappa in evaluating the
agreement between two binary ratings. This approach is becoming increasingly popular …

[PDF][PDF] A new measure of agreement to resolve the two paradoxes of Cohen's Kappa

MH Park, YG Park - The Korean Journal of Applied Statistics, 2007 - koreascience.kr
Abstract In a $2\times2 $ table showing binary agreement between two raters, it is known
that Cohen's $\kappa $, a chance-corrected measure of agreement, has two paradoxes …

Agreement between raters and groups of raters

S Vanbelle - 2009 - orbi.uliege.be
Agreement between raters on a categorical scale is not only a subject of scientific research
but also a problem frequently encountered in practice. Whenever a new scale is developed …

[HTML][HTML] Measures of agreement with multiple raters: Fréchet variances and inference

J Moss - Psychometrika, 2024 - Springer
Most measures of agreement are chance-corrected. They differ in three dimensions: their
definition of chance agreement, their choice of disagreement function, and how they handle …

[HTML][HTML] Homogeneity score test of AC1 statistics and estimation of common AC1 in multiple or stratified inter-rater agreement studies

C Honda, T Ohyama - BMC medical research methodology, 2020 - Springer
Background Cohen's κ coefficient is often used as an index to measure the agreement of
inter-rater determinations. However, κ varies greatly depending on the marginal distribution …

An Exact Bootstrap Confidence Interval for κ in Small Samples

N Klar, SR Lipsitz, M Parzen… - Journal of the Royal …, 2002 - academic.oup.com
Agreement between a pair of raters for binary outcome data is typically assessed by using
the κ-coefficient. When the total sample size is small to moderate, and the proportion of …

[HTML][HTML] Estimators of various kappa coefficients based on the unbiased estimator of the expected index of agreements

A Martín Andrés, M Álvarez Hernández - Advances in Data Analysis and …, 2024 - Springer
To measure the degree of agreement between R observers who independently classify n
subjects within K categories, various kappa-type coefficients are often used. When R= 2, it is …

Statistical inference for agreement between multiple raters on a binary scale

S Vanbelle - British Journal of Mathematical and Statistical …, 2024 - Wiley Online Library
Agreement studies often involve more than two raters or repeated measurements. In the
presence of two raters, the proportion of agreement and of positive agreement are simple …