Skip to main content

Breadcrumb

Home arrow_forward_ios Information on IES-Funded Research arrow_forward_ios Model-based Multiple Imputation for ...
Home arrow_forward_ios ... arrow_forward_ios Model-based Multiple Imputation for ...
Information on IES-Funded Research
Grant Closed

Model-based Multiple Imputation for Multilevel Data: Methodological Extensions and Software Enhancements

NCER
Program: Statistical and Research Methodology in Education
Program topic(s): Core
Award amount: $868,046
Principal investigator: Craig Enders
Awardee:
University of California, Los Angeles
Year: 2019
Project type:
Methodological Innovation
Award number: R305D190002

Purpose

Missing data are exceedingly common in educational research. Education research studies have missing data because students opt out of achievement testing, skip test items, or move to a different school district, among many other reasons. A previous IES award funded the development of a data analysis application, called Blimp, that addresses this issue by filling in missing values using sophisticated predictive models. The purpose of the current work was twofold: expand Blimp's missing data imputation features to handle a broader range of reasons for missing data and further develop the software into a general-use data analysis program.

Project Activities

The project team increased the capabilities of the Blimp software application by including a set of not missing at random. In addition, the team added residual-based model diagnostics, estimation for count outcomes, dispersion modeling features, and sampling weights.

People and institutions involved

IES program contact(s)

Allen Ruby

Associate Commissioner for Policy and Systems
NCER

Products and publications

ERIC Citations: Find available citations in ERIC for this award here

Project Website: www.appliedmissingdata.com/blimp

Additional Online Resources and Information:

  • Keller, B. T., & Enders, C. K. (2023). Blimp user's guide (Version 3). Available at www.appliedmissingdata.com/blimp.
  • www.appliedmissingdata.com/blimp-papers - analysis scripts and data sets from PI Enders' 2022 missing data book and papers and chapters written by the research team and others
  • www.appliedmissingdata.com/videos - training videos and a software tutorial document with annotated software scripts and output files

Select Publications:

Alacam, E., Enders, C.K., Du, H., & Keller, B.T. (2023). A factored regression model for composite scores with item-level missing data. Psychological Methods, Advanced online publication.

Du, H., Keller, B. T., Alacam, E., & Enders, C. K. (2023). Comparing DIC and WAIC for multilevel models with missing data. Behavior Research Methods. 1-20.

Du, H., Enders, C.K., Keller, B.T., Bradbury, T., & Karney, B. (2022). A Bayesian latent variable selection model for nonignorable missingness. Multivariate Behavioral Research, 57, 478-512.

Keller, B. T. (2022). An introduction to factored regression models with Blimp. Psych, 4(1), 10-37.

Keller, B.T., & Enders, C.K. (2023). An investigation of factored regression missing data methods for multilevel models with cross-level interactions. Multivariate Behavioral Research, 58, 938-963.

Project website:

https://www.appliedmissingdata.com/blimp

Related projects

Multiple Imputation Procedures for Multilevel Data

R305D150056

Model-based Multiple Imputation for Multilevel Data: Methodological Extensions and Software Enhancements

R305D190002

Dealing with Missing Data in Educational Research: Methodological Innovations and Contemporary Recommendations

R305D220001

Supplemental information

Co-Principal Investigators: Du, Han; Keller, Brian

Key Outcomes

  • The researchers developed and released a new version (#3) of the Blimp software, with user guide, for other researchers to use to fit a statistical model of interest to a data set with missing values (www.appliedmissingdata.com/blimp).
  • Technical details about the software's algorithms were released in peer-reviewed publications (Alcam et al., 2023; Du et al., 2022; Du et al., in press; Keller 2022; Keller & Enders 2023).

Statistical/Methodological Product: Blimp is a general-use data analysis application available for MacOS, Windows, and Linux. The software was created for education researchers with incomplete data sets. Blimp's suite of data analytic capabilities additionally makes the software broadly applicable to researchers in the social, behavioral, and medical sciences, among others. The software is available from the project's website (www.appliedmissingdata.com/blimp), and a detailed user guide is available through the graphical interface's Help pull-down menu. The project also produced several peer-reviewed publications that provide technical details about the software's algorithms.

Development/Refinement Process: Developing reliable and accurate data analysis tools requires intensive testing and development and a high level of quality control. The development process involved the following steps:

  • Members of the research team developed a technical appendix that describes the procedure to be implemented, its algorithmic details, and relevant extant literature.
  • Initial coding and testing were conducted in R.
  • The programmer refined the initial code and implemented the new procedure into Blimp's C++ codebase. At this point, new functionality was available in a special version of the computational engine that team members accessed for testing.
  • The research team conducted exhaustive and extensive tests using computer simulation studies and benchmarking against other software packages, when available. The testing also involved applying the new methodology to numerous real data sets.
  • New functionality was made available in a special beta version of the computational engine that members of the public could access by listing a shebang at the top of the script.
  • Additional testing on real-world data sets was done.
  • The beta features were moved to the software's public build.
  • Updates continue to be distributed in real time over the internet when the user launches the software.

User Testing: The research team relied on three sources of user testing. First, team members used Blimp in their graduate-level statistics courses. Second, researchers worldwide. Observing the target audience interacting with the software in these settings has been an important source of user testing and feedback. Third, the team maintains a user support email to assist practicing researchers. Addressing user inquires has allowed the team to observe the software's behavior across a vast collection of real-word data structures. Collectively, these user testing experiences have allowed the research team to continually refine the graphical interface and printed output to enhance and simplify the user experience.

Questions about this project?

To answer additional questions about this project or provide feedback, please contact the program officer.

 

Tags

MathematicsData and Assessments

Share

Icon to link to Facebook social media siteIcon to link to X social media siteIcon to link to LinkedIn social media siteIcon to copy link value

Questions about this project?

To answer additional questions about this project or provide feedback, please contact the program officer.

 

You may also like

Zoomed in IES logo
Workshop/Training

Data Science Methods for Digital Learning Platform...

August 18, 2025
Read More
Zoomed in IES logo
Workshop/Training

Meta-Analysis Training Institute (MATI)

July 28, 2025
Read More
Zoomed in Yellow IES Logo
Workshop/Training

Bayesian Longitudinal Data Modeling in Education S...

July 21, 2025
Read More
icon-dot-govicon-https icon-quote