Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 2 |
Descriptor
Source
Journal of Educational… | 16 |
Author
Bridgeman, Brent | 2 |
Wainer, Howard | 2 |
Ankenman, Robert D. | 1 |
Ben-Shakhar, Gershon | 1 |
Chen, Shu-Ying | 1 |
Dorans, Neil J. | 1 |
Freedle, Roy | 1 |
Gerritz, Kalle | 1 |
Gressard, Risa P. | 1 |
Holland, Paul W. | 1 |
Kim, Sooyeon | 1 |
More ▼ |
Publication Type
Journal Articles | 16 |
Reports - Research | 10 |
Reports - Evaluative | 5 |
Reports - Descriptive | 1 |
Education Level
Elementary Secondary Education | 1 |
Audience
Location
Israel | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 4 |
SAT (College Admission Test) | 3 |
What Works Clearinghouse Rating
Sinharay, Sandip; Holland, Paul W. – Journal of Educational Measurement, 2007
It is a widely held belief that anchor tests should be miniature versions (i.e., "minitests"), with respect to content and statistical characteristics, of the tests being equated. This article examines the foundations for this belief regarding statistical characteristics. It examines the requirement of statistical representativeness of…
Descriptors: Test Items, Comparative Testing
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias

Gressard, Risa P.; Loyd, Brenda H. – Journal of Educational Measurement, 1991
A Monte Carlo study, which simulated 10,000 examinees' responses to four tests, investigated the effect of item stratification on parameter estimation in multiple matrix sampling of achievement data. Practical multiple matrix sampling is based on item stratification by item discrimination and a sampling plan with moderate number of subtests. (SLD)
Descriptors: Achievement Tests, Comparative Testing, Computer Simulation, Estimation (Mathematics)

Wainer, Howard; And Others – Journal of Educational Measurement, 1992
Computer simulations were run to measure the relationship between testlet validity and factors of item pool size and testlet length for both adaptive and linearly constructed testlets. Making a testlet adaptive yields only modest increases in aggregate validity because of the peakedness of the typical proficiency distribution. (Author/SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation

Bridgeman, Brent; Rock, Donald A. – Journal of Educational Measurement, 1993
Exploratory and confirmatory factor analyses were used to explore relationships among existing item types and three new computer-administered item types for the analytical scale of the Graduate Record Examination General Test. Results with 349 students indicate constructs the item types are measuring. (SLD)
Descriptors: College Entrance Examinations, College Students, Comparative Testing, Computer Assisted Testing
Chen, Shu-Ying; Ankenman, Robert D. – Journal of Educational Measurement, 2004
The purpose of this study was to compare the effects of four item selection rules--(1) Fisher information (F), (2) Fisher information with a posterior distribution (FP), (3) Kullback-Leibler information with a posterior distribution (KP), and (4) completely randomized item selection (RN)--with respect to the precision of trait estimation and the…
Descriptors: Test Length, Adaptive Testing, Computer Assisted Testing, Test Selection

Wainer, Howard; And Others – Journal of Educational Measurement, 1991
Hierarchical (adaptive) and linear methods of testlet construction were compared. The performance of 2,080 ninth and tenth graders on a 4-item testlet was used to predict performance on the entire test. The adaptive test was slightly superior as a predictor, but the cost of obtaining that superiority was considerable. (SLD)
Descriptors: Adaptive Testing, Algebra, Comparative Testing, High School Students
Li, Yuan H.; Lissitz, Robert W. – Journal of Educational Measurement, 2004
The analytically derived asymptotic standard errors (SEs) of maximum likelihood (ML) item estimates can be approximated by a mathematical function without examinees' responses to test items, and the empirically determined SEs of marginal maximum likelihood estimation (MMLE)/Bayesian item estimates can be obtained when the same set of items is…
Descriptors: Test Items, Computation, Item Response Theory, Error of Measurement

Freedle, Roy; Kostin, Irene – Journal of Educational Measurement, 1990
The importance of item difficulty (equated delta) was explored as a predictor of differential item functioning of Black versus White examinees for 4 verbal item types using 13 Graduate Record Examination forms and 11 Scholastic Aptitude Test forms. Several significant racial differences were found. (TJH)
Descriptors: Black Students, College Bound Students, College Entrance Examinations, Comparative Testing

Dorans, Neil J.; And Others – Journal of Educational Measurement, 1992
The standardization approach to comprehensive differential item functioning is described and contrasted with the log-linear approach to differential distractor functioning and the item-response-theory-based approach to differential alternative functioning. Data from an edition of the Scholastic Aptitude Test illustrate application of the approach…
Descriptors: Black Students, College Entrance Examinations, Comparative Testing, Distractors (Tests)

Ben-Shakhar, Gershon; Sinai, Yakov – Journal of Educational Measurement, 1991
Gender differences in omitting items and guessing on multiple-choice tests were studied in Israel for 302 male and 302 female ninth graders and 150 male and 150 female university applicants. Females tended to omit more items and guess less often than did males. Implications for scoring are discussed. (SLD)
Descriptors: Aptitude Tests, Cognitive Ability, College Applicants, Comparative Testing

Bridgeman, Brent – Journal of Educational Measurement, 1992
Examinees in a regular administration of the quantitative portion of the Graduate Record Examination responded to particular items in a machine-scannable multiple-choice format. Volunteers (n=364) used a computer to answer open-ended counterparts of these items. Scores for both formats demonstrated similar correlational patterns. (SLD)
Descriptors: Answer Sheets, College Entrance Examinations, College Students, Comparative Testing

Martinez, Michael E. – Journal of Educational Measurement, 1991
Figural response items (FRIs) in science were administered to 347 fourth graders, 365 eighth graders, and 322 twelfth graders. Item and test statistics from parallel FRIs and multiple-choice questions illustrate FRIs' more difficult and more discriminating nature. Relevance of guessing to FRIs and diagnostic value of the item type are highlighted.…
Descriptors: Comparative Testing, Constructed Response, Elementary School Students, Elementary Secondary Education

Ryan, Katherine E. – Journal of Educational Measurement, 1991
The reliability of Mantel-Haenszel (MH) indexes across samples of examinees and sample sizes and their robustness to item context effects were investigated with data for 670 African-American and 5,015 white students from the Second International Mathematics Study. MH procedures can be used to detect differential item functioning. (SLD)
Descriptors: Black Students, Comparative Testing, Context Effect, Evaluation Criteria

Wise, Steven L.; And Others – Journal of Educational Measurement, 1992
Performance of 156 undergraduate and 48 graduate students on a self-adapted test (SFAT)--students choose the difficulty level of their test items--was compared with performance on a computer-adapted test (CAT). Those taking the SFAT obtained higher ability scores and reported lower posttest state anxiety than did CAT takers. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Difficulty Level
Previous Page | Next Page ยป
Pages: 1 | 2