Computing inter‐rater reliability and its variance in the presence of high agreement

KL Gwet - British Journal of Mathematical and Statistical …, 2008 - Wiley Online Library
Pi (π) and kappa (κ) statistics are widely used in the areas of psychiatry and psychological
testing to compute the extent of agreement between raters on nominally scaled data. It is a …

Beyond kappa: A review of interrater agreement measures

M Banerjee, M Capozzoli… - Canadian journal of …, 1999 - Wiley Online Library
In 1960, Cohen introduced the kappa coefficient to measure chance‐corrected nominal
scale agreement between two raters. Since then, numerous extensions and generalizations …

Sample size determinations for the two rater kappa statistic

VF Flack, AA Afifi, PA Lachenbruch, HJA Schouten - Psychometrika, 1988 - Springer
This paper gives a method for determining a sample size that will achieve a prespecified
bound on confidence interval width for the interrater agreement measure, κ. The same …

[PDF][PDF] Kappa statistic is not satisfactory for assessing the extent of agreement between raters

K Gwet - Statistical methods for inter-rater reliability assessment, 2002 - agreestat.com
Evaluating the extent of agreement between 2 or between several raters is common in
social, behavioral and medical sciences. The objective of this paper is to provide a detailed …

Learning how to differ: agreement and reliability statistics in psychiatry

LS David - The Canadian Journal of Psychiatry, 1995 - journals.sagepub.com
Whenever two or more raters evaluate a patient or student, it may be necessary to determine
the degree to which they assign the same label or rating to the subject. The major problem in …

Fleiss' kappa statistic without paradoxes

R Falotico, P Quatto - Quality & Quantity, 2015 - Springer
The Fleiss' kappa statistic is a well-known index for assessing the reliability of agreement
between raters. It is used both in the psychological and in the psychiatric field. Unfortunately …

[HTML][HTML] Kappa coefficient: a popular measure of rater agreement

T Wan, HU Jun, H Zhang, WU Pan… - Shanghai archives of …, 2015 - ncbi.nlm.nih.gov
In mental health and psychosocial studies it is often necessary to report on the between-
rater agreement of measures used in the study. This paper discusses the concept of …

Five ways to look at Cohen's kappa

MJ Warrens - Journal of Psychology & Psychotherapy, 2015 - research.rug.nl
The kappa statistic is commonly used for quantifying inter-rater agreement on a nominal
scale. In this review article we discuss five interpretations of this popular coefficient. Kappa is …

A new interpretation of the weighted kappa coefficients

S Vanbelle - Psychometrika, 2016 - Springer
Reliability and agreement studies are of paramount importance. They do contribute to the
quality of studies by providing information about the amount of error inherent to any …

Variance estimation of nominal-scale inter-rater reliability with random selection of raters

KL Gwet - Psychometrika, 2008 - Springer
Most inter-rater reliability studies using nominal scales suggest the existence of two
populations of inference: the population of subjects (collection of objects or persons to be …