Bias, prevalence and kappa

T Byrt, J Bishop, JB Carlin - Journal of clinical epidemiology, 1993 - Elsevier
Since the introduction of Cohen's kappa as a chance-adjusted measure of agreement
between two observers, several “paradoxes” in its interpretation have been pointed out. The …

[PDF][PDF] Sample size determination and power analysis for modified Cohen's Kappa statistic

P Yimprayoon - Applied Mathematical Sciences, 2013 - m-hikari.com
In this research, the statistical inference of the problem of measuring agreement between
two observers who employ measurements on a 2-point nominal scale is focused. One of the …

Interval estimation under two study designs for kappa with binary classifications

CA Hale, JL Fleiss - Biometrics, 1993 - JSTOR
Cornfield's test-based method of setting a confidence interval on a parameter associated
with a two-by-two contingency table is adapted for use with the measure of agreement …

Interrater reliability: the kappa statistic

ML McHugh - Biochemia medica, 2012 - hrcak.srce.hr
Sažetak The kappa statistic is frequently used to test interrater reliability. The importance of
rater reliability lies in the fact that it represents the extent to which the data collected in the …

Measures of agreement with multiple raters: Fréchet variances and inference

J Moss - Psychometrika, 2024 - Springer
Most measures of agreement are chance-corrected. They differ in three dimensions: their
definition of chance agreement, their choice of disagreement function, and how they handle …

Assessment of observer agreement for matched repeated binary measurements

J Gao, Y Pan, M Haber - Computational Statistics & Data Analysis, 2012 - Elsevier
Agreement is a broad term simultaneously covering evaluations of accuracy and precision of
measurements. Assessment of observer agreement is based on the similarity between …

Overall indices for assessing agreement among multiple raters

JH Jang, AK Manatunga, AT Taylor… - Statistics in …, 2018 - Wiley Online Library
The need to assess agreement exists in various clinical studies where quantifying inter‐rater
reliability is of great importance. Use of unscaled agreement indices, such as total deviation …

Measurement of interobserver disagreement: Correction of Cohen's kappa for negative values

TO Kvålseth - Journal of Probability and Statistics, 2015 - Wiley Online Library
As measures of interobserver agreement for both nominal and ordinal categories, Cohen's
kappa coefficients appear to be the most widely used with simple and meaningful …

HOW reliable are change‐corrected measures of agreement?

I Guggenmoos‐Holzmann - Statistics in Medicine, 1993 - Wiley Online Library
Chance‐corrected measures of agreement are prone to exhibit paradoxical and counter‐
intuitive results when used as measures of reliability. It is demonstrated that these problems …

Multiple‐rater kappas for binary data: Models and interpretation

D Stoyan, A Pommerening, M Hummel… - Biometrical …, 2018 - Wiley Online Library
Interrater agreement on binary measurements with more than two raters is often assessed
using Fleiss' κ, which is known to be difficult to interpret. In situations where the same raters …