Interrater agreement statistics under the two-rater dichotomous-response case with correlated decisions

Z Tian, VM Chinchilli, C Shen, S Zhou - arXiv preprint arXiv:2402.08069, 2024 - arxiv.org
Measurement of the interrater agreement (IRA) is critical in various disciplines. To correct for
potential confounding chance agreement in IRA, Cohen's kappa and many other methods …

Implementing a general framework for assessing interrater agreement in Stata

D Klein - The Stata Journal, 2018 - journals.sagepub.com
Despite its well-known weaknesses, researchers continuously choose the kappa coefficient
(Cohen, 1960, Educational and Psychological Measurement 20: 37–46; Fleiss, 1971 …

Lambda Coefficient of Rater-Mediated Agreement: Evaluation of an Alternative Chance-Corrected Agreement Coefficient

TS Holcomb - 2022 - search.proquest.com
In this study, the performance of the Lambda Coefficient of Rater-Mediated Agreement was
evaluated with other chance-corrected agreement coefficients. Lambda is grounded in rater …

Weighted inter-rater agreement measures for ordinal outcomes

D Tran, A Dolgun, H Demirhan - Communications in Statistics …, 2020 - Taylor & Francis
Estimation of the degree of agreement between different raters is of crucial importance in
medical and social sciences. There are lots of different approaches proposed in the …

An overview of interrater agreement on Likert scales for researchers and practitioners

TA O'Neill - Frontiers in psychology, 2017 - frontiersin.org
Applications of interrater agreement (IRA) statistics for Likert scales are plentiful in research
and practice. IRA may be implicated in job analysis, performance appraisal, panel …

Bayesian approaches to the weighted kappa-like inter-rater agreement measures

QD Tran, H Demirhan, A Dolgun - Statistical Methods in …, 2021 - journals.sagepub.com
Inter-rater agreement measures are used to estimate the degree of agreement between two
or more assessors. When the agreement table is ordinal, different weight functions that …

A unified model for continuous and categorical data

L Lin, AS Hedayat, W Wu, L Lin, AS Hedayat… - Statistical Tools for …, 2012 - Springer
In this chapter, we generalize agreement assessment for continuous and categorical data to
cover multiple raters (k≥ 2), and each with multiple readings (m≥ 1) from each of the n …

Interrater reliability estimators tested against true interrater reliabilities

X Zhao, GC Feng, SH Ao, PL Liu - BMC medical research methodology, 2022 - Springer
Background Interrater reliability, aka intercoder reliability, is defined as true agreement
between raters, aka coders, without chance agreement. It is used across many disciplines …

An empirical comparative assessment of inter-rater agreement of binary outcomes and multiple raters

M Konstantinidis, LW Le, X Gao - Symmetry, 2022 - mdpi.com
Background: Many methods under the umbrella of inter-rater agreement (IRA) have been
proposed to evaluate how well two or more medical experts agree on a set of outcomes. The …

[PDF][PDF] Impact of rater disagreement on chance-corrected inter-rater agreement indices with equal and unequal marginal proportions

DA Walker - Multiple Linear Regression Viewpoints, 2008 - glmj.org
Methods The data for the subsequent situations tested on each of the inter-rater agreement
indices were derived from an SPSS (Statistical Package for the Social Sciences v. 15.0) …