With two judges and a two-point rating scale, the test statistic for Kappa is the same as Pearson's chi-square statistic applied to the 2 × 2 table of paired observations. This equivalence allows a quick test of the null hypothesis of no agreement, as Pearson's chisquare statistic is much less cumbersome to compute than the Kappa statistic and its variance. A simple formula for the null hypothesis variance is also derived.
Get full access to this article
View all access options for this article.
References
1.
Butler, D. and Stokes, D. (1969). Political change in Britain: Forces shaping electoral choice. New York: St. Martin's Press.
2.
Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20, 37-46.
3.
Fleiss, J. L. , Cohen, J., and Everitt, B. S. (1969). Large sample standard errors of Kappa and weighted Kappa. Psychological Bulletin, 72, 323-327.