Abstract
This paper gives a method for determining a sample size that will achieve a prespecified bound on confidence interval width for the interrater agreement measure,κ. The same results can be used when a prespecified power is desired for testing hypotheses about the value of kappa. An example from the literature is used to illustrate the methods proposed here.
Similar content being viewed by others
References
Cohen, J. (1960). A coefficient of agreement for nominal scales.Educational and Psychological Measurement, 20, 37–46.
Cohen, J. (1968). Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit.Psychological Bulletin, 70, 213–220.
Dixon, W. J., & Massey, F. J., Jr. (1983).Introduction to statistical analysis (4th ed.). New York: McGraw-Hill.
Flack, V. F. (1987). Confidence intervals for the two rater kappa.Communications in Statistics: Theory and Methods, 16, 953–968.
Fleiss, J. L. (1981).Statistical methods for rates and proportions (2nd ed.). New York: Wiley.
Fleiss, J. L., Cohen, J., & Everitt, B. S. (1969). Large sample standard errors of kappa and weighted kappa.Psychological Bulletin, 72, 323–327.
Landis, J. R., & Koch, G. G. (1977). The measurement of interrater agreement for categorical data.Biometrics, 33, 159–174.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Flack, V.F., Afifi, A.A., Lachenbruch, P.A. et al. Sample size determinations for the two rater kappa statistic. Psychometrika 53, 321–325 (1988). https://doi.org/10.1007/BF02294215
Received:
Revised:
Issue Date:
DOI: https://doi.org/10.1007/BF02294215