Hostname: page-component-745bb68f8f-b95js Total loading time: 0 Render date: 2025-01-07T18:55:13.742Z Has data issue: false hasContentIssue false

Sample Size Determinations for the Two Rater Kappa Statistic

Published online by Cambridge University Press:  01 January 2025

V. F. Flack*
Affiliation:
Division of Biostatistics, UCLA School of Public Health
A. A. Afifi
Affiliation:
Division of Biostatistics, UCLA School of Public Health
P. A. Lachenbruch
Affiliation:
Division of Biostatistics, UCLA School of Public Health
H. J. A. Schouten
Affiliation:
Department of Medical Informatics and Statistics, University of Limburg
*
Reprint requests should be sent to Virginia F. Flack, UCLA School of Public Health, Division of Biostatistics, Los Angeles, CA 90024.

Abstract

This paper gives a method for determining a sample size that will achieve a prespecified bound on confidence interval width for the interrater agreement measure, κ. The same results can be used when a prespecified power is desired for testing hypotheses about the value of kappa. An example from the literature is used to illustrate the methods proposed here.

Type
Original Paper
Copyright
Copyright © 1988 The Psychometric Society

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20, 3746.CrossRefGoogle Scholar
Cohen, J. (1968). Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit. Psychological Bulletin, 70, 213220.CrossRefGoogle ScholarPubMed
Dixon, W. J., & Massey, F. J. Jr (1983). Introduction to statistical analysis 4th ed.,, New York: McGraw-Hill.Google Scholar
Flack, V. F. (1987). Confidence intervals for the two rater kappa. Communications in Statistics: Theory and Methods, 16, 953968.CrossRefGoogle Scholar
Fleiss, J. L. (1981). Statistical methods for rates and proportions 2nd ed.,, New York: Wiley.Google Scholar
Fleiss, J. L., Cohen, J., & Everitt, B. S. (1969). Large sample standard errors of kappa and weighted kappa. Psychological Bulletin, 72, 323327.CrossRefGoogle Scholar
Landis, J. R., & Koch, G. G. (1977). The measurement of interrater agreement for categorical data. Biometrics, 33, 159174.CrossRefGoogle Scholar