Residential College | false |
Status | 已發表Published |
A Reliability Index(ai) that Assumes Honest Coders and Variable Randomness | |
Zhao XS(趙心樹) | |
2012-08 | |
Conference Name | the 2012 Annual Conference of Association for Education in Journalism and Mass Communication, Chicago. |
Conference Date | 2012-08 |
Conference Place | Chicago |
Abstract | The performances of six major indices of inter-coder reliability were evaluated against actual judgments of human coders in a behavior-based Monte Carlo (BMC) experiment. The correlations between the indices’ estimated chance agreements (ac) and the observed chance agreements (oac) turned out to be negative for Cohen’s κ, Scott’s π and Krippendorff’s α, and mild although positive for Bennett et al’s S, Perrault and Leigh’s Ir and Gwet’s AC1. While each of the indices was designed to improve on percent agreement, each underperformed percent agreement (ao) when estimating observed true agreement (at) in the BMC experiment. The poor or negative correlations between the calculated estimates and the observed estimands question the validity of the estimators, namely the indices. The findings support the emerging theory that reliability indices available today assume dishonest coders who deliberately maximize chance coding, and they are therefore unsuitable for typical studies where coders perform chance coding involuntarily when the task is too difficult. A new index or indices are needed. This manuscript also reports the effort to develop such a new index, agreement index (ai), which assumes honest coders and involuntary chance coding. Subsequent analysis shows that ai is void of the 23 known paradoxes that plague other indices. In the BMC experiment, the chance agreement estimated by ai was by far the best predictor of the observed chance agreement between coders. Index ai also outperformed percent agreement and all other six indices while predicting true agreements among the coders. Empirical testing of theories and indices should continue, especially by different researchers using different methods, and so should the search for a better index. Until better evidences are available, however, researchers may refrain from using κ, π, and α, and add ai as a reasonable measure of true agreements between two coders on a nominal scale. Online software has been provided at http://reliability.hkbu.edu.hk/ to facilitate calculation. |
Keyword | Reliability Intercoder Reliability Interrater Reliability Agreement Index Estimator Estimate Estimand Maximum Randomness Variable Randomness Behavioral Monte-carlo Experiment Bmc Simulation-augmented Behavior Experiment Sab Kappa Alpha Pi |
Indexed By | 其他 |
Document Type | Conference paper |
Collection | Faculty of Social Sciences DEPARTMENT OF COMMUNICATION |
Affiliation | Hong Kong Baptist University |
Recommended Citation GB/T 7714 | Zhao, X.S. (2012, August). A Reliability Index(ai) that Assumes Honest Coders and Variable Randomness [Presentation Session] the 2012 Annual Conference of Association for Education in Journalism and Mass Communication, Chicago, America. https://repository.um.edu.mo/handle/10692/95485 |
APA | Zhao XS.(2012). A Reliability Index(ai) that Assumes Honest Coders and Variable Randomness. . |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment