Computing inter‐rater reliability and its variance in the presence of high agreement

KL Gwet - British Journal of Mathematical and Statistical …, 2008 - Wiley Online Library
Pi (π) and kappa (κ) statistics are widely used in the areas of psychiatry and psychological
testing to compute the extent of agreement between raters on nominally scaled data. It is a …

Chance-corrected measures of reliability and validity in KK tables

AM Andrés, PF Marzo - Statistical methods in medical …, 2005 - journals.sagepub.com
When studying the degree of overall agreement between the nominal responses of two
raters, it is customary to use the coefficient kappa. A more detailed analysis requires the …

Moments of the statistics kappa and weighted kappa

BS Everitt - British Journal of Mathematical and Statistical …, 1968 - Wiley Online Library
MOMENTS OF THE STATISTICS KAPPA AND WEIGHTED KAPPA Page 1 Vol. Part 1 97-103
The British Journal of Mathematical and Statistical Psychology May 1968 MOMENTS OF THE …

An Evaluation of Interrater Reliability Measures on Binary Tasks Using d-Prime

MJ Grant, CM Button, B Snook - Applied psychological …, 2017 - journals.sagepub.com
Many indices of interrater agreement on binary tasks have been proposed to assess
reliability, but none has escaped criticism. In a series of Monte Carlo simulations, five such …

Nominal scale response agreement and rater uncertainty

R Gillett - British Journal of Mathematical and Statistical …, 1985 - Wiley Online Library
Current methods of assessing nominal scale response agreement between two raters allow
each rater to make only a single response per object. When raters are uncertain in their …

Appropriate statistics for determining chance-removed interpractitioner agreement

M Popplewell, J Reizes, C Zaslawski - The Journal of Alternative …, 2019 - liebertpub.com
Abstract Objectives: Fleiss' Kappa (FK) has been commonly, but incorrectly, employed as the
“standard” for evaluating chance-removed inter-rater agreement with ordinal data. This …

[PDF][PDF] Kappa statistic is not satisfactory for assessing the extent of agreement between raters

K Gwet - Statistical methods for inter-rater reliability assessment, 2002 - agreestat.com
Evaluating the extent of agreement between 2 or between several raters is common in
social, behavioral and medical sciences. The objective of this paper is to provide a detailed …

A generalization of Cohen's kappa agreement measure to interval measurement and multiple raters

KJ Berry, PW Mielke Jr - Educational and Psychological …, 1988 - journals.sagepub.com
Cohen's kappa statistic is frequently used to measure agreement between two observers
employing categorical polytomies. In this paper, Cohen's statistic is shown to be inherently …

[图书][B] Analyzing rater agreement: Manifest variable methods

A Von Eye, EY Mun - 2014 - books.google.com
Agreement among raters is of great importance in many domains. For example, in medicine,
diagnoses are often provided by more than one doctor to make sure the proposed treatment …

PROPERTIES OF THE HOLLEY‐GUILFORD 'G INDEX OF AGREEMENT' IN R AND Q FACTOR ANALYSIS

P Levy - Scandinavian Journal of Psychology, 1966 - Wiley Online Library
The Holley‐Guilford G index for 2× 2 contingency tables has spatial properties which
suggest that, as with other such indices, problems may arise in factor analysis of G matrices …