Skip to main content

Breadcrumb

Home arrow_forward_ios Information on IES-Funded Research arrow_forward_ios A Study of the Missing Data Assumpt ...
Home arrow_forward_ios ... arrow_forward_ios A Study of the Missing Data Assumpt ...
Information on IES-Funded Research
Grant Closed

A Study of the Missing Data Assumptions of the Non-equivalent Groups with Anchor Test Design and Their Implications for Test Equating

NCER
Program: Statistical and Research Methodology in Education
Program topic(s): Core
Award amount: $74,934
Principal investigator: Sandip Sinharay
Awardee:
Educational Testing Service (ETS)
Year: 2007
Project type:
Methodological Innovation
Award number: R305U070009

Purpose

With the introduction of the No Child Left Behind Act of 2001, the number of tests taken by U.S. grade school students grew. Many of these tests administer different test forms on different dates so that students who take the test on different dates will not answer exactly the same test questions. To ensure fair scores on high-stakes exams, individual differences in scores taken on exams on different test dates must reflect differences in achievement levels of the students and not differences in test difficulty. To achieve this goal, researchers employed a technique called equating, a statistical procedure to ensure that scores from different test forms can be used interchangeably. Among the different equating designs, the Non-Equivalent groups with Anchor Test (NEAT) design is widely used for equating tests used to help make high-stakes decisions, such as the SAT, Advanced Placement tests, and standardized state assessments. Three popular equating methods that can be used with a NEAT design are: the poststratification equipercentile equating method (also called the frequency estimation method), the chain equipercentile equating method, and the item response theory observed score equating method. Each of these methods makes distinct assumptions about the missing data in the NEAT design. These assumptions are usually neither explicit nor testable under the usual equating situations. This project team examined the missing data assumptions and their implications on equating, and obtained a deeper understanding of the comparative performance of the three abovementioned equating methods.

Project Activities

In this project, the research team examined the three above-mentioned equating methods. First, the missing data assumptions for the three methods were described explicitly. Next, the research team compared the three methods using datasets from several operational tests, including those given to U.S. grade school students, which use the NEAT design for equating. For any dataset, the team examined the three equating methods performed when the missing data satisfy the assumptions made by only one of these equating methods. The outcomes from this project were intended to assist practitioners who use the NEAT design to choose the most appropriate method for equating; this will in turn ensure reporting of fair scores in tests that employ a NEAT design.

Structured Abstract

Research design and methods

In this project, the research team examined three popular equating methods that can be used with a NEAT design: the poststratification equipercentile equating method (also called the frequency estimation method), the chain equipercentile equating method, and the item response theory observed score equating method. First, the missing data assumptions for the three methods were described explicitly. Next, the research team compared the three methods using datasets from several operational tests, including those given to U.S. grade school students, which use the NEAT design for equating. For any dataset, the team examined how the three equating methods perform in comparison to each other when the missing data satisfy the assumptions made by only one of these equating methods.

Data analytic strategy

For data from each test, the following three-step procedure was used to compare the methods under different missing data assumptions:

People and institutions involved

IES program contact(s)

Allen Ruby

Associate Commissioner for Policy and Systems
NCER

Products and publications

Sinharay, S., & Holland, P. W. (2009). The missing data assumptions of the nonequivalent groups with anchor test (NEAT) design and their implications for test equating. ETS Research Report Series, 2009(1), i-53.

Sinharay, S., & Holland, P. W. (2010). A new approach to comparing several equating methods in the context of the NEAT design. Journal of Educational Measurement, 47(3), 261-285.

Sinharay, S., & Holland, P. W. (2010). The missing data assumptions of the NEAT design and their implications for test equating. Psychometrika, 75, 309-327.

** This project was submitted to and funded as an Unsolicited application in FY 2007.

Supplemental information

  • The "true equating function" under each method is obtained by making the missing data assumptions inherent in the method. For example, a typical missing data assumption made by the poststratification equating method is that of all the examinees who obtained a score of, say, 10, on the anchor test, the proportion that obtained a score of, say, 30, on the test to be equated is the same irrespective of whether the examinees belong to the new form population or to the old form population.
  • The equating function is then obtained for each equating method. These are called the "observed equating functions." These are the standard equating functions that, for example, a testing company employing these methods computes from the data in an operational equating environment.
  • The difference between the two (the true and observed equating functions) at each score point is computed for each pair of true and observed equating functions. The differences are plotted in two-dimensional graphical displays. The differences are summarized using their weighted averages.

The above steps were carried out under two alternative methods (linear equating and kernel equating) for continuizing discrete data and for several tests.

Questions about this project?

To answer additional questions about this project or provide feedback, please contact the program officer.

 

Tags

Data and AssessmentsMathematics

Share

Icon to link to Facebook social media siteIcon to link to X social media siteIcon to link to LinkedIn social media siteIcon to copy link value

Questions about this project?

To answer additional questions about this project or provide feedback, please contact the program officer.

 

You may also like

Zoomed in IES logo
Workshop/Training

Data Science Methods for Digital Learning Platform...

August 18, 2025
Read More
Zoomed in IES logo
Workshop/Training

Meta-Analysis Training Institute (MATI)

July 28, 2025
Read More
Zoomed in Yellow IES Logo
Workshop/Training

Bayesian Longitudinal Data Modeling in Education S...

July 21, 2025
Read More
icon-dot-govicon-https icon-quote