NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 11 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Jue; Engelhard, George, Jr. – Educational and Psychological Measurement, 2019
The purpose of this study is to explore the use of unfolding models for evaluating the quality of ratings obtained in rater-mediated assessments. Two different judgmental processes can be used to conceptualize ratings: impersonal judgments and personal preferences. Impersonal judgments are typically expected in rater-mediated assessments, and…
Descriptors: Evaluative Thinking, Preferences, Evaluators, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A.; Engelhard, George, Jr. – Educational and Psychological Measurement, 2016
Mokken scale analysis is a probabilistic nonparametric approach that offers statistical and graphical tools for evaluating the quality of social science measurement without placing potentially inappropriate restrictions on the structure of a data set. In particular, Mokken scaling provides a useful method for evaluating important measurement…
Descriptors: Nonparametric Statistics, Statistical Analysis, Measurement, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Jue; Engelhard, George, Jr.; Wolfe, Edward W. – Educational and Psychological Measurement, 2016
The number of performance assessments continues to increase around the world, and it is important to explore new methods for evaluating the quality of ratings obtained from raters. This study describes an unfolding model for examining rater accuracy. Accuracy is defined as the difference between observed and expert ratings. Dichotomous accuracy…
Descriptors: Evaluators, Accuracy, Performance Based Assessment, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr.; Kobrin, Jennifer L.; Wind, Stefanie A. – International Journal of Testing, 2014
The purpose of this study is to explore patterns in model-data fit related to subgroups of test takers from a large-scale writing assessment. Using data from the SAT, a calibration group was randomly selected to represent test takers who reported that English was their best language from the total population of test takers (N = 322,011). A…
Descriptors: College Entrance Examinations, Writing Tests, Goodness of Fit, English
Peer reviewed Peer reviewed
Direct linkDirect link
Behizadeh, Nadia; Engelhard, George, Jr. – Assessing Writing, 2011
The purpose of this study is to examine the interactions among measurement theories, writing theories, and writing assessments in the United States from an historical perspective. The assessment of writing provides a useful framework for examining how theories influence, and in some cases fail to influence actual practice. Two research traditions…
Descriptors: Writing (Composition), Intellectual Disciplines, Writing Evaluation, Writing Tests
Gyagenda, Ismail S.; Engelhard, George, Jr. – 1998
The purpose of this study was to examine rater, domain, and gender influences on the assessed quality of student writing using weighted and unweighted scores. Twenty rates were randomly selected from a group of 87 operational raters contracted to rate essays as part of the 1993 field test of the Georgia High School Writing Test. All of the raters…
Descriptors: Essay Tests, Evaluators, High School Students, High Schools
Gyagenda, Ismail S.; Engelhard, George, Jr. – 1998
The purpose of this study was to describe the Rasch model for measurement and apply the model to examine the relationship between raters, domains of written compositions, and student writing ability. Twenty raters were randomly selected from a group of 87 operational raters contracted to rate essays as part of the 1993 field test of the Georgia…
Descriptors: Difficulty Level, Essay Tests, Evaluators, High School Students
Engelhard, George, Jr. – 1991
A many-faceted Rasch model (FACETS) is presented for the measurement of writing ability. The FACETS model is a multivariate extension of Rasch measurement models that can be used to provide a framework for calibrating both raters and writing tasks within the context of writing assessment. A FACETS model is described based on the current procedures…
Descriptors: Grade 8, Holistic Evaluation, Interrater Reliability, Item Response Theory
Peer reviewed Peer reviewed
Engelhard, George, Jr. – Journal of Educational Measurement, 1994
Rater errors (rater severity, halo effect, central tendency, and restriction of range) are described, and criteria are presented for evaluating rating quality based on a many-faceted Rasch (FACETS) model. Ratings of 264 compositions from the Eighth Grade Writing Test in Georgia by 15 raters illustrate the discussion. (SLD)
Descriptors: Criteria, Educational Assessment, Elementary Education, Elementary School Students
Peer reviewed Peer reviewed
Engelhard, George, Jr.; And Others – Journal of Educational Research, 1994
The influences of writing tasks and gender on the quality of student writing of black and white eighth graders were examined. Data from statewide writing assessments of 170,899 Georgia students indicated both writing tasks and student characteristics were significant predictors of writing quality. There were both racial and gender differences. (SM)
Descriptors: Blacks, Grade 8, Junior High School Students, Junior High Schools
Peer reviewed Peer reviewed
Engelhard, George, Jr. – Applied Measurement in Education, 1992
A Many-Faceted Rasch Model (FACETS) for measurement of writing ability is described, and its use in solving measurement problems in large-scale assessment is illustrated with a random sample of 1,000 students from Georgia's Eighth Grade Writing Test. It is a promising approach to assessment through written compositions. (SLD)
Descriptors: Educational Assessment, Essays, Evaluation Problems, Grade 8