NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 10 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Liu, Jinghua; Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2014
Maintaining score interchangeability and scale consistency is crucial for any testing programs that administer multiple forms across years. The use of a multiple linking design, which involves equating a new form to multiple old forms and averaging the conversions, has been proposed to control scale drift. However, the use of multiple linking…
Descriptors: Comparative Analysis, Reliability, Test Construction, Equated Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Steedle, Jeffrey T. – Online Submission, 2010
Tests of college learning are often administered to obtain value-added scores indicating whether score gains are below, near, or above typical performance for students of given entering academic ability. This study compares the qualities of value-added scores generated by the original Collegiate Learning Assessment value-added approach and a new…
Descriptors: Institutional Evaluation, Academic Achievement, Academic Ability, Outcomes of Education
Shaw, Emily J.; Mattern, Krista D. – College Board, 2009
This study examined the relationship between students' self-reported high school grade point average (HSGPA) from the SAT Questionnaire and their HSGPA provided by the colleges and universities they attend. The purpose of this research was to offer updated information on the relatedness of self-reported (by the student) and school-reported (by the…
Descriptors: High School Students, Grade Point Average, Accuracy, Aptitude Tests
Bailey, Roger L. – Measurement and Evaluation in Guidance, 1978
Scores from the Scholastic Aptitude Test and a new test of reading and writing, the California State University and Colleges English Placement Test, were examined to determine if unique properties for either test could be found. Both tests seem to be measuring the same underlying ability with few unique properties. (Author)
Descriptors: Aptitude Tests, College Students, Comparative Analysis, Higher Education
Petersen, Nancy S.; And Others – 1981
Three equating methods were compared in terms of magnitude of scale drift: equipercentile equating, linear equating, and item response theory (IRT) equating. A sample of approximately 2670 cases was selected for each pairing of a form of the Scholastic Aptitude Tests (SAT) and an anchor test. Of the two conventional equating methods,…
Descriptors: College Entrance Examinations, Comparative Analysis, Equated Scores, Latent Trait Theory
Peer reviewed Peer reviewed
Bennett, Randy Elliot; And Others – Special Services in the Schools, 1988
A study of Scholastic Aptitude Test scores for nine groups of students with disabilities taking special test administrations found differences in score levels among disability groups but no significant differences of measurement precision and no evidence of disadvantage for disabled students. (Author/MSE)
Descriptors: Adaptive Testing, College Entrance Examinations, Comparative Analysis, Disabilities
Camara, Wayne J. – College Entrance Examination Board, 2003
Previous research on differences in the reliability, validity, and difficulty of essay tests given under different timing conditions has indicated that giving examinees more time to complete an essay may raise their scores to a certain extent, but does not change the meaning of those scores, or the rank ordering of students. There is no evidence…
Descriptors: Essays, Comparative Analysis, Writing Tests, Timed Tests
Hendrickson, Amy; Patterson, Brian; Melican, Gerald – College Board, 2008
Presented at the Annual National Council on Measurement in Education (NCME) in New York in March 2008. This presentation explores how different item weighting can affect the effective weights, validity coefficents and test reliability of composite scores among test takers.
Descriptors: Multiple Choice Tests, Test Format, Test Validity, Test Reliability
Downey, Ronald G.
Previous research has studied the effects of different methods of item option weighting on the reliability and concurrent and predictive validity of achievement tests. Increases in reliability are generally found, but with mixed results for validity. Several methods of producing option weights, (i.e., Guttman internal and external weights and…
Descriptors: Achievement Tests, Comparative Analysis, Correlation, Grade Point Average