Abstract
This paper describes the integration of Natural Language Processing (NLP) features within an explanatory item response modelling framework in the context of a reading comprehension assessment item bank. Item properties derived through NLP algorithms were incorporated into a Rasch Latent Regression Linear Logistic Test Model with item error, extending the model described by Wilson and de Boeck (2004) on the item side with a random error term. Specifically, item difficulties were modelled as random variables that could be predicted (with uncertainty) (Janssen, Schepers and Peres, 2004) by NLP item property fixed effects, and person covariates were included to increase the accuracy of estimation of latent ability distributions. The focus of this study was on the extent to which different kinds of NLP features explained variance in item difficulties. We investigated how these results could be used to develop and validate proficiency level descriptors and item bank meta-data.
Original language | English |
---|---|
Publication status | Published - Nov 2021 |
Event | NBME Online Conference - Natural Language Processing in Assessment - Philadelphia, PA Duration: 1 Nov 2021 → … |
Conference
Conference | NBME Online Conference - Natural Language Processing in Assessment |
---|---|
Period | 1/11/21 → … |
Disciplines
- Educational Assessment, Evaluation, and Research