Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 9 |
Descriptor
College Entrance Examinations | 12 |
Test Reliability | 11 |
Reliability | 7 |
High Schools | 6 |
Item Response Theory | 6 |
Test Items | 6 |
Test Construction | 5 |
Comparative Analysis | 4 |
High School Students | 4 |
Scores | 4 |
Standardized Tests | 4 |
More ▼ |
Source
Author
Dorans, Neil J. | 2 |
Albanese, Mark A. | 1 |
Attali, Yigal | 1 |
Bennett, Randy Elliot | 1 |
Camara, Wayne J. | 1 |
Cameron, Robert G. | 1 |
Charlin, Bernard | 1 |
Covington, Helen C. | 1 |
Cui, Ying | 1 |
Dory, Valerie | 1 |
Eignor, Daniel R. | 1 |
More ▼ |
Publication Type
Reports - Evaluative | 21 |
Journal Articles | 11 |
Speeches/Meeting Papers | 4 |
Information Analyses | 1 |
Reports - Research | 1 |
Education Level
Higher Education | 8 |
Postsecondary Education | 4 |
Elementary Secondary Education | 1 |
Secondary Education | 1 |
Audience
Location
United States | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Katzman, John – New England Journal of Higher Education, 2014
It is so easy to criticize the SAT that most observers overlook the weaknesses of its architect, the College Board. This author contents that, until the latter is replaced, however, the former will never be fixed. The College Board has every incentive to create a complex, stressful, expensive college admissions system. Because it is accountable to…
Descriptors: Standardized Tests, Testing Programs, Program Administration, Cost Effectiveness
Dorans, Neil J. – Educational Measurement: Issues and Practice, 2012
Views on testing--its purpose and uses and how its data are analyzed--are related to one's perspective on test takers. Test takers can be viewed as learners, examinees, or contestants. I briefly discuss the perspective of test takers as learners. I maintain that much of psychometrics views test takers as examinees. I discuss test takers as a…
Descriptors: Testing, Test Theory, Item Response Theory, Test Reliability
Sinha, Ruchi; Oswald, Frederick; Imus, Anna; Schmitt, Neal – Applied Measurement in Education, 2011
The current study examines how using a multidimensional battery of predictors (high-school grade point average (GPA), SAT/ACT, and biodata), and weighting the predictors based on the different values institutions place on various student performance dimensions (college GPA, organizational citizenship behaviors (OCBs), and behaviorally anchored…
Descriptors: Grade Point Average, Interrater Reliability, Rating Scales, College Admission
Steedle, Jeffrey T. – Online Submission, 2010
Tests of college learning are often administered to obtain value-added scores indicating whether score gains are below, near, or above typical performance for students of given entering academic ability. This study compares the qualities of value-added scores generated by the original Collegiate Learning Assessment value-added approach and a new…
Descriptors: Institutional Evaluation, Academic Achievement, Academic Ability, Outcomes of Education
Dory, Valerie; Gagnon, Robert; Charlin, Bernard – Advances in Health Sciences Education, 2010
Case-specificity, i.e., variability of a subject's performance across cases, has been a consistent finding in medical education. It has important implications for assessment validity and reliability. Its root causes remain a matter of discussion. One hypothesis, content-specificity, links variability of performance to variable levels of relevant…
Descriptors: Medical Education, Trainees, English (Second Language), Error of Measurement
Isaacs, Emily; Molloy, Sean A. – College English, 2010
Focusing on writing placement at a particular university, the authors analyze the limits of SAT tests as a tool in this process. They then describe the writing program's adoption of a supplementary measure: a faculty committee's review of essays by students who may need to be reassigned to a different writing course. They describe how and why a…
Descriptors: Writing Strategies, Cutting Scores, Writing Instruction, Student Evaluation
Gierl, Mark J.; Cui, Ying; Zhou, Jiawen – Journal of Educational Measurement, 2009
The attribute hierarchy method (AHM) is a psychometric procedure for classifying examinees' test item responses into a set of structured attribute patterns associated with different components from a cognitive model of task performance. Results from an AHM analysis yield information on examinees' cognitive strengths and weaknesses. Hence, the AHM…
Descriptors: Test Items, True Scores, Psychometrics, Algebra
Whithaus, Carl; Harrison, Scott B.; Midyette, Jeb – Assessing Writing, 2008
This article examines the influence of keyboarding versus handwriting in a high-stakes writing assessment. Conclusions are based on data collected from a pilot project to move Old Dominion University's Exit Exam of Writing Proficiency from a handwritten format into a dual-option format (i.e., the students may choose to handwrite or keyboard the…
Descriptors: Writing Evaluation, Handwriting, Pilot Projects, Writing Tests
Dorans, Neil J. – Applied Psychological Measurement, 2004
How do scores from different tests relate to each other? Three types of score linkage are discussed: equating, concordance, and prediction of expected scores. Statistical indices, in conjunction with rational considerations, are needed to determine whether the highest level of linkage attainable between scores from two "tests" is the…
Descriptors: Prediction, College Entrance Examinations, Scores, Achievement Tests
Lawrence, Ida M. – 1995
This study examined to what extent, if any, estimates of reliability for a multiple choice test are affected by the presence of large item sets where each set shares common reading material. The purpose of this research was to assess the effect of local item dependence on estimates of reliability for verbal portions of seven forms of the old and…
Descriptors: Estimation (Mathematics), High Schools, Multiple Choice Tests, Reading Tests

Bennett, Randy Elliot; And Others – Special Services in the Schools, 1988
A study of Scholastic Aptitude Test scores for nine groups of students with disabilities taking special test administrations found differences in score levels among disability groups but no significant differences of measurement precision and no evidence of disadvantage for disabled students. (Author/MSE)
Descriptors: Adaptive Testing, College Entrance Examinations, Comparative Analysis, Disabilities
Wainer, Howard; And Others – 1991
It is sometimes sensible to think of the fundamental unit of test construction as being larger than an individual item. This unit, dubbed the testlet, must pass muster in the same way that items do. One criterion of a good item is the absence of differential item functioning (DIF). The item must function in the same way as all important…
Descriptors: Definitions, Identification, Item Bias, Item Response Theory
Cameron, Robert G. – 1989
The contribution of the Scholastic Aptitude Test (SAT) to education is described, and the most frequently asked questions about its use are addressed. The SAT became a prominent tool in college admissions when the members of the College Board recognized in the 1930's that the Board's program of very specific subject-matter tests was no longer…
Descriptors: College Bound Students, College Entrance Examinations, Educational Assessment, High Schools
Camara, Wayne J. – College Entrance Examination Board, 2003
Previous research on differences in the reliability, validity, and difficulty of essay tests given under different timing conditions has indicated that giving examinees more time to complete an essay may raise their scores to a certain extent, but does not change the meaning of those scores, or the rank ordering of students. There is no evidence…
Descriptors: Essays, Comparative Analysis, Writing Tests, Timed Tests
Previous Page | Next Page ยป
Pages: 1 | 2