Abstract
This study proposes an empirical Bayesian approach with loss functions for testing the fairness of an automated scoring algorithm. The proposed method outperforms traditional approaches by (1) being robust to small samples, (2) incorporating parameter uncertainty, and (3) balancing the loss of keeping versus flagging items. The impact of flagging potentially unfair items on the classification of examinees is investigated. The effectiveness of the proposed method is illustrated through simulations and an application to language assessment data.
Get full access to this article
View all access options for this article.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
