Modeling improvement for educational assessments: comparisons and extensions
Hu, Mingqi
This item is only available for download by members of the University of Illinois community. Students, faculty, and staff at the U of I may log in with your NetID and password to view the item. If you are trying to access an Illinois-restricted dissertation or thesis, you can request a copy through your library's Inter-Library Loan office or purchase a copy directly from ProQuest.
Permalink
https://hdl.handle.net/2142/121337
Description
Title
Modeling improvement for educational assessments: comparisons and extensions
Author(s)
Hu, Mingqi
Issue Date
2023-07-06
Director of Research (if dissertation) or Advisor (if thesis)
Zhang, Jinming
Doctoral Committee Chair(s)
Zhang, Jinming
Committee Member(s)
Anderson, Carolyn J.
Chang, Hua-hua
Xia, Yan
Department of Study
Educational Psychology
Discipline
Educational Psychology
Degree Granting Institution
University of Illinois at Urbana-Champaign
Degree Name
Ph.D.
Degree Level
Dissertation
Keyword(s)
item response theory
cognitive diagnostic models
neural network
Abstract
Item response theory models (IRTMs) and cognitive diagnostic models (CDMs) are two well-known families of theory-driven educational assessment models. IRTMs focus on the relative position of students in an ability continuum, while CDMs focus on the absolute diagnostic classification of students’ latent traits. Multidimensional IRTMs (MIRTMs) share similarities with CDMs. Both can measure the multidimensional abilities of students through dichotomously scored items. And they can be divided into compensatory and noncompensatory branches. The application of CDMs in real assessment practices is less prevalent than IRTMs, and one possible reason is the assessment stability among calibration samples. If CDMs do not perfectly match the data which is common in real practices, the item invariance property is violated. Chapter 2 explores the robustness of CDMs assuming underlying attributes are continuous, studying factors that influence the definition of mastery in MIRTM-based assessments. Results showed that 0 was not the universal cut point of continuous attributes for mastery vs non-mastery. Indeed, the scaling issue occurred when CDMs were used to handle MIRTM-based datasets.
Additionally, because of the development of computing power and data science, researchers adopt machine learning techniques to improve educational assessment models. Specifically, the neural network (NN) is a nonparametric and model-agnostic approach that does not require any assumptions of the parameter distribution, which is powerful in approximating complicated nonlinear functions. Previous neural assessment models constricted hidden layers of the NN in different ways to improve the interpretability of the internal working of the model. A novel neural CDM (NN-CDM) was introduced in Chapter 3, where a small sample of test takers’ responses and attribute mastery states were used for training. Simulation studies were conducted in DINA-based datasets and MIRTM-based datasets with predetermined cut points. Also, this chapter filled the research gap in the application of neural assessment models in attribute hierarchy conditions. The NN-CDM outperformed the existing supervised neural CDM in all conditions.
The other important extension of traditional assessment models is to deal with longitudinal assessments. Longitudinal assessments can track students’ stepwise performance, which fits the need for e-learning platforms. There is a limited discussion on model weaknesses and strengths, as well as practical concerns. Most previous literature focuses on the model performance of conditions with less than 4 time points. In Chapter 4, simulation studies were conducted to compare the performance of four longitudinal assessment models (three L-IRTMs and one L-CDM) delivering continuous latent ability estimates. The strengths and weaknesses of each model were discussed, from both conceptual and practical perspectives. In terms of L-IRTMs, more time points, larger sample sizes, and longer tests slightly improved the accuracy of ability parameter estimation. The L-UIRTM obtained the most accurate estimates in all conditions.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.