Ridit and exponential type scores for estimating the kappa statistic

AE Yilmaz, S Aktas - Kuwait Journal of Science, 2018 - journalskuwait.org
Cohen's kappa coefficient is a commonly used method for estimating interrater agreement
for nominal and/or ordinal data; thus agreement is adjusted for that expected by chance. The …

Estimating rater agreement in 2 x 2 tables: Correction for chance and intraclass correlation

NJM Blackman, JJ Koval - Applied Psychological …, 1993 - journals.sagepub.com
Many estimators of the measure of agreement between two dichotomous ratings of a person
have been proposed. The results of Fleiss (1975) are extended, and it is shown that four …

Measuring agreement between two judges on the presence or absence of a trait

JL Fleiss - Biometrics, 1975 - JSTOR
At least a dozen indexes have been proposed for measuring agreement between two
judges on a categorical scale. Using the binary (positive-negative) case as a model, this …

The equivalence of Cohen's Kappa and Pearson's chi-square statistics in the 2× 2 table

M Feingold - Educational and psychological measurement, 1992 - journals.sagepub.com
With two judges and a two-point rating scale, the test statistic for Kappa is the same as
Pearson's chi-square statistic applied to the 2× 2 table of paired observations. This …

Integration and generalization of kappas for multiple raters.

AJ Conger - Psychological bulletin, 1980 - psycnet.apa.org
Abstract JA Cohen's kappa (1960) for measuring agreement between 2 raters, using a
nominal scale, has been extended for use with multiple raters by RJ Light (1971) and JL …

A new coefficient of interrater agreement: The challenge of highly unequal category proportions.

R Van Oest - Psychological Methods, 2019 - psycnet.apa.org
We derive a general structure that encompasses important coefficients of interrater
agreement such as the S-coefficient, Cohen's kappa, Scott's pi, Fleiss' kappa, Krippendorff's …

Measuring agreement among several raters classifying subjects into one-or-more (hierarchical) nominal categories. A generalisation of Fleiss' kappa

F Moons, E Vandervieren - arXiv preprint arXiv:2303.12502, 2023 - arxiv.org
Cohen's and Fleiss' kappa are well-known measures for inter-rater reliability. However, they
only allow a rater to select exactly one category for each subject. This is a severe limitation …

Cohen's linearly weighted kappa is a weighted average of 2× 2 kappas

MJ Warrens - Psychometrika, 2011 - Springer
An agreement table with n∈ ℕ≥ 3 ordered categories can be collapsed into n− 1 distinct 2×
2 tables by combining adjacent categories. Vanbelle and Albert (Stat. Methodol. 6: 157–163 …

The exact variance of weighted kappa with multiple raters

PW Mielke Jr, KJ Berry… - Psychological Reports, 2007 - journals.sagepub.com
Weighted kappa described by Cohen in 1968 is widely used in psychological research to
measure agreement between two independent raters. Everitt then provided the exact …

Measures of concordance for assessing agreement in ratings and rank order data

M Raghavachari - Advances in Ranking and Selection, Multiple …, 2005 - Springer
Consider an array of ratings or rank orders provided by M judges or respondents on n
subjects or products. We propose a new general measure of concordance (agreement) in …