NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)4
Since 2006 (last 20 years)11
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 13 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Jue; Engelhard, George, Jr. – Journal of Educational Measurement, 2019
Rater-mediated assessments exhibit scoring challenges due to the involvement of human raters. The quality of human ratings largely determines the reliability, validity, and fairness of the assessment process. Our research recommends that the evaluation of ratings should be based on two aspects: a theoretical model of human judgment and an…
Descriptors: Evaluative Thinking, Models, Measurement, Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr.; Perkins, Aminah – Measurement: Interdisciplinary Research and Perspectives, 2013
In this commentary, Englehard and Perkins remark that Maydeu-Olivares has presented a framework for evaluating the goodness of model-data fit for item response theory (IRT) models and correctly points out that overall goodness-of-fit evaluations of IRT models and data are not generally explored within most applications in educational and…
Descriptors: Goodness of Fit, Item Response Theory, Models, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Jue; Engelhard, George, Jr.; Wolfe, Edward W. – Educational and Psychological Measurement, 2016
The number of performance assessments continues to increase around the world, and it is important to explore new methods for evaluating the quality of ratings obtained from raters. This study describes an unfolding model for examining rater accuracy. Accuracy is defined as the difference between observed and expert ratings. Dichotomous accuracy…
Descriptors: Evaluators, Accuracy, Performance Based Assessment, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A.; Engelhard, George, Jr.; Wesolowski, Brian – Educational Assessment, 2016
When good model-data fit is observed, the Many-Facet Rasch (MFR) model acts as a linking and equating model that can be used to estimate student achievement, item difficulties, and rater severity on the same linear continuum. Given sufficient connectivity among the facets, the MFR model provides estimates of student achievement that are equated to…
Descriptors: Evaluators, Interrater Reliability, Academic Achievement, Music Education
Peer reviewed Peer reviewed
Direct linkDirect link
Chang, Mei-Lin; Engelhard, George, Jr. – Journal of Psychoeducational Assessment, 2016
The purpose of this study is to examine the psychometric quality of the Teachers' Sense of Efficacy Scale (TSES) with data collected from 554 teachers in a U.S. Midwestern state. The many-facet Rasch model was used to examine several potential contextual influences (years of teaching experience, school context, and levels of emotional exhaustion)…
Descriptors: Models, Teacher Attitudes, Self Efficacy, Item Response Theory
Engelhard, George, Jr.; Wind, Stefanie A. – College Board, 2013
The major purpose of this study is to examine the quality of ratings assigned to CR (constructed-response) questions in large-scale assessments from the perspective of Rasch Measurement Theory. Rasch Measurement Theory provides a framework for the examination of rating scale category structure that can yield useful information for interpreting the…
Descriptors: Measurement Techniques, Rating Scales, Test Theory, Scores
Kaliski, Pamela; Wind, Stefanie A.; Engelhard, George, Jr.; Morgan, Deanna; Plake, Barbara; Reshetar, Rosemary – College Board, 2012
The Many-Facet Rasch (MFR) Model is traditionally used to evaluate the quality of ratings on constructed response assessments; however, it can also be used to evaluate the quality of judgments from panel-based standard setting procedures. The current study illustrates the use of the MFR Model by examining the quality of ratings obtained from a…
Descriptors: Advanced Placement Programs, Achievement Tests, Item Response Theory, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Kaliski, Pamela K.; Wind, Stefanie A.; Engelhard, George, Jr.; Morgan, Deanna L.; Plake, Barbara S.; Reshetar, Rosemary A. – Educational and Psychological Measurement, 2013
The many-faceted Rasch (MFR) model has been used to evaluate the quality of ratings on constructed response assessments; however, it can also be used to evaluate the quality of judgments from panel-based standard setting procedures. The current study illustrates the use of the MFR model for examining the quality of ratings obtained from a standard…
Descriptors: Item Response Theory, Models, Standard Setting (Scoring), Science Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Randall, Jennifer; Cheong, Yuk Fai; Engelhard, George, Jr. – Educational and Psychological Measurement, 2011
To address whether or not modifications in test administration influence item functioning for students with disabilities on a high-stakes statewide problem-solving assessment, a sample of 868 students (with and without disabilities) from 74 Georgia schools were randomly assigned to one of three testing conditions (resource guide, calculator, or…
Descriptors: Item Response Theory, Models, Context Effect, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr.; Perkins, Aminah F. – Measurement: Interdisciplinary Research and Perspectives, 2011
Humphry (this issue) has written a thought-provoking piece on the interpretation of item discrimination parameters as scale units in item response theory. One of the key features of his work is the description of an item response theory (IRT) model that he calls the logistic measurement function that combines aspects of two traditions in IRT that…
Descriptors: Foreign Countries, Social Sciences, Item Response Theory, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr. – Measurement: Interdisciplinary Research and Perspectives, 2008
The major purpose of my focus article was to stimulate discussion regarding the concept of invariant measurement. My intent was to provide a historical lens for considering how our views of invariant measurement have evolved over time through the work of three key measurement theorists: Guttman, Rasch, and Mokken. The commentators have offered a…
Descriptors: Measurement, Item Response Theory, Models
Engelhard, George, Jr. – 1986
A model based on organizational theory was used in this study to discover educational goals emphasized by teachers in classrooms and to explore the impact of the social organization of schooling on these goals. Age-grade structure was examined as a possible organizational characteristic influencing moral and technical goals selected by teachers.…
Descriptors: Age Grade Placement, Catholic Schools, Classroom Techniques, Educational Objectives
Engelhard, George, Jr.; Myford, Carol M. – College Board, 2003
The purpose of this study was to examine, describe, evaluate, and compare the rating behavior of faculty consultants who scored essays written for the Advanced Placement English Literature and Composition (APĀ® ELC) Exam. Data from the 1999 AP ELC Exam were analyzed using FACETS (Linacre, 1998) and SAS. The faculty consultants were not all…
Descriptors: Advanced Placement, College Faculty, Consultants, Scoring