Publication Date
In 2025 | 12 |
Since 2024 | 187 |
Since 2021 (last 5 years) | 818 |
Since 2016 (last 10 years) | 1951 |
Since 2006 (last 20 years) | 4074 |
Descriptor
Item Response Theory | 5553 |
Test Items | 1817 |
Foreign Countries | 1196 |
Models | 1148 |
Psychometrics | 918 |
Scores | 782 |
Comparative Analysis | 761 |
Test Construction | 750 |
Simulation | 740 |
Statistical Analysis | 659 |
Difficulty Level | 570 |
More ▼ |
Source
Author
Sinharay, Sandip | 48 |
Wilson, Mark | 45 |
Cohen, Allan S. | 43 |
Meijer, Rob R. | 43 |
Tindal, Gerald | 42 |
Wang, Wen-Chung | 40 |
Alonzo, Julie | 37 |
Ferrando, Pere J. | 36 |
Cai, Li | 35 |
van der Linden, Wim J. | 35 |
Glas, Cees A. W. | 34 |
More ▼ |
Publication Type
Education Level
Location
Turkey | 94 |
Australia | 89 |
Germany | 79 |
United States | 74 |
Netherlands | 68 |
Taiwan | 59 |
Indonesia | 53 |
China | 51 |
Canada | 49 |
Japan | 38 |
Florida | 37 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 4 |
Meets WWC Standards with or without Reservations | 4 |
Ken A. Fujimoto; Carl F. Falk – Educational and Psychological Measurement, 2024
Item response theory (IRT) models are often compared with respect to predictive performance to determine the dimensionality of rating scale data. However, such model comparisons could be biased toward nested-dimensionality IRT models (e.g., the bifactor model) when comparing those models with non-nested-dimensionality IRT models (e.g., a…
Descriptors: Item Response Theory, Rating Scales, Predictive Measurement, Bayesian Statistics
Junhuan Wei; Qin Wang; Buyun Dai; Yan Cai; Dongbo Tu – Journal of Educational Measurement, 2024
Traditional IRT and IRTree models are not appropriate for analyzing the item that simultaneously consists of multiple-choice (MC) task and constructed-response (CR) task in one item. To address this issue, this study proposed an item response tree model (called as IRTree-MR) to accommodate items that contain different response types at different…
Descriptors: Item Response Theory, Models, Multiple Choice Tests, Cognitive Processes
Wind, Stefanie A. – Educational and Psychological Measurement, 2023
Rating scale analysis techniques provide researchers with practical tools for examining the degree to which ordinal rating scales (e.g., Likert-type scales or performance assessment rating scales) function in psychometrically useful ways. When rating scales function as expected, researchers can interpret ratings in the intended direction (i.e.,…
Descriptors: Rating Scales, Testing Problems, Item Response Theory, Models
Zeyuan Jing – ProQuest LLC, 2023
This dissertation presents a comprehensive review of the evolution of DIF analysis within educational measurement from the 1980s to the present. The review elucidates the concept of DIF, particularly emphasizing the crucial role of grouping for exhibiting DIF. Then, the dissertation introduces an innovative modification to the newly developed…
Descriptors: Item Response Theory, Algorithms, Measurement, Test Bias
Miguel A. García-Pérez – Educational and Psychological Measurement, 2024
A recurring question regarding Likert items is whether the discrete steps that this response format allows represent constant increments along the underlying continuum. This question appears unsolvable because Likert responses carry no direct information to this effect. Yet, any item administered in Likert format can identically be administered…
Descriptors: Likert Scales, Test Construction, Test Items, Item Analysis
Joseph A. Rios; Jiayi Deng – Educational and Psychological Measurement, 2024
Rapid guessing (RG) is a form of non-effortful responding that is characterized by short response latencies. This construct-irrelevant behavior has been shown in previous research to bias inferences concerning measurement properties and scores. To mitigate these deleterious effects, a number of response time threshold scoring procedures have been…
Descriptors: Reaction Time, Scores, Item Response Theory, Guessing (Tests)
Stephen L. Wright; Michael A. Jenkins-Guarnieri – Journal of Psychoeducational Assessment, 2024
The current study sought out to advance the Social Self-Efficacy and Social Outcome Expectations scale using multiple approaches to scale development. Data from 583 undergraduate students were used in two scale development approaches: Classic Test Theory (CTT) and Item Response Theory (IRT). Confirmatory factor analysis suggested a 2-factor…
Descriptors: Measures (Individuals), Expectation, Self Efficacy, Item Response Theory
Evaluating an Explicit Instruction Teacher Observation Protocol through a Validity Argument Approach
Johnson, Evelyn S.; Zheng, Yuzhu; Crawford, Angela R.; Moylan, Laura A. – Journal of Experimental Education, 2022
In this study, we examined the scoring and generalizability assumptions of an explicit instruction (EI) special education teacher observation protocol using many-faceted Rasch measurement (MFRM). Video observations of classroom instruction from 48 special education teachers across four states were collected. External raters (n = 20) were trained…
Descriptors: Direct Instruction, Teacher Education, Classroom Observation Techniques, Validity
Tyrone B. Pretorius; P. Paul Heppner; Anita Padmanabhanunni; Serena Ann Isaacs – SAGE Open, 2023
In previous studies, problem solving appraisal has been identified as playing a key role in promoting positive psychological well-being. The Problem Solving Inventory is the most widely used measure of problem solving appraisal and consists of 32 items. The length of the instrument, however, may limit its applicability to large-scale surveys…
Descriptors: Problem Solving, Measures (Individuals), Test Construction, Item Response Theory
Sa'ar Karp Gershon; Ella Anghel; Giora Alexandron – Education and Information Technologies, 2024
For Massive Open Online Courses to have trustworthy credentials, assessments in these courses must be valid, reliable, and fair. Item Response Theory provides a robust approach to evaluating these properties. However, for this theory to be applicable, certain properties of the assessment items should be met, among them that item difficulties are…
Descriptors: MOOCs, Item Response Theory, Physics, Advanced Placement Programs
Jianbin Fu; Xuan Tan; Patrick C. Kyllonen – Journal of Educational Measurement, 2024
This paper presents the item and test information functions of the Rank two-parameter logistic models (Rank-2PLM) for items with two (pair) and three (triplet) statements in forced-choice questionnaires. The Rank-2PLM model for pairs is the MUPP-2PLM (Multi-Unidimensional Pairwise Preference) and, for triplets, is the Triplet-2PLM. Fisher's…
Descriptors: Questionnaires, Test Items, Item Response Theory, Models
The Impact of Measurement Noninvariance across Time and Group in Longitudinal Item Response Modeling
In-Hee Choi – Asia Pacific Education Review, 2024
Longitudinal item response data often exhibit two types of measurement noninvariance: the noninvariance of item parameters between subject groups and that of item parameters across multiple time points. This study proposes a comprehensive approach to the simultaneous modeling of both types of measurement noninvariance in terms of longitudinal item…
Descriptors: Longitudinal Studies, Item Response Theory, Growth Models, Error of Measurement
Murat Tekin; Çetin Toraman; Aysen Melek Aytug Kosan – International Journal of Assessment Tools in Education, 2024
In the present study, we examined the psychometric properties of the data obtained from the Commitment to Profession of Medicine Scale (CPMS) with 4-point, 5-point, 6-point, and 7-point response sets based on Item Response Theory (IRT). A total of 2150 medical students from 16 different universities participated in the study. The participants were…
Descriptors: Psychometrics, Medical Students, Likert Scales, Data Collection
Gabrielle T. Lee; Xiaoyi Hu; Chun Shen – Journal of Autism and Developmental Disorders, 2024
The purpose of the study was to evaluate the effects of tact and match-to-sample instructions on the increase and maintenance of intraverbal responses to subcategorical questions (i.e., naming multiple items in a subcategory of a category). Three Chinese children on the autism spectrum (2 boys, 1 girl, aged 6-8 years old) participated in this…
Descriptors: Autism Spectrum Disorders, Children, Foreign Countries, Verbal Communication
William C. M. Belzak; Daniel J. Bauer – Journal of Educational and Behavioral Statistics, 2024
Testing for differential item functioning (DIF) has undergone rapid statistical developments recently. Moderated nonlinear factor analysis (MNLFA) allows for simultaneous testing of DIF among multiple categorical and continuous covariates (e.g., sex, age, ethnicity, etc.), and regularization has shown promising results for identifying DIF among…
Descriptors: Test Bias, Algorithms, Factor Analysis, Error of Measurement