Project Activities
The research team will first develop new models for automated scoring that align with human judgment. The team will write these models into statistical software and then tested on real text datasets. The researchers will then expand the models and software to allow the machine learning algorithm to evaluate potential differences between experimental or quasi-experimental groups in the texts produced by the groups, not just average score differences between the groups. They will continue software testing on additional real datasets to gauge its performance at discerning text pattern differences.
People and institutions involved
IES program contact(s)
Project contributors
Products and publications
When the software is ready, the research team will create a free, web-based, user-friendly version for use by applied researchers. The team will then create instructional materials for use in workshops and short courses for teaching users how to use the software. The team will also produce manuscripts for publication in methodological and applied peer-reviewed journals.
Publications:
Mozer, R., & Miratrix, L. (2023). Decreasing the human coding burden in randomized trials with text-based outcomes via model-assisted impact analysis. arXiv preprint arXiv:2309.13666.
Mozer, R., & Miratrix, L. (2025). More power to you: Using machine learning to augment human coding for more efficient inference in text-based randomized trials. The Annals of Applied Statistics, 19(1), 440-464.
Mozer, R., Miratrix, L., Relyea, J. E., & Kim, J. S. (2023). Combining Human and Automated Scoring Methods in Experimental Assessments of Writing: A Case Study Tutorial. Journal of Educational and Behavioral Statistics, 10769986231207886.
Questions about this project?
To answer additional questions about this project or provide feedback, please contact the program officer.