Publication Date
In 2025 | 12 |
Since 2024 | 187 |
Since 2021 (last 5 years) | 818 |
Since 2016 (last 10 years) | 1951 |
Since 2006 (last 20 years) | 4074 |
Descriptor
Item Response Theory | 5553 |
Test Items | 1817 |
Foreign Countries | 1196 |
Models | 1148 |
Psychometrics | 918 |
Scores | 782 |
Comparative Analysis | 761 |
Test Construction | 750 |
Simulation | 740 |
Statistical Analysis | 659 |
Difficulty Level | 570 |
More ▼ |
Source
Author
Sinharay, Sandip | 48 |
Wilson, Mark | 45 |
Cohen, Allan S. | 43 |
Meijer, Rob R. | 43 |
Tindal, Gerald | 42 |
Wang, Wen-Chung | 40 |
Alonzo, Julie | 37 |
Ferrando, Pere J. | 36 |
Cai, Li | 35 |
van der Linden, Wim J. | 35 |
Glas, Cees A. W. | 34 |
More ▼ |
Publication Type
Education Level
Location
Turkey | 94 |
Australia | 89 |
Germany | 79 |
United States | 74 |
Netherlands | 68 |
Taiwan | 59 |
Indonesia | 53 |
China | 51 |
Canada | 49 |
Japan | 38 |
Florida | 37 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 4 |
Meets WWC Standards with or without Reservations | 4 |
Philippe Goldammer; Peter Lucas Stöckli; Yannik Andrea Escher; Hubert Annen; Klaus Jonas – Educational and Psychological Measurement, 2024
Indirect indices for faking detection in questionnaires make use of a respondent's deviant or unlikely response pattern over the course of the questionnaire to identify them as a faker. Compared with established direct faking indices (i.e., lying and social desirability scales), indirect indices have at least two advantages: First, they cannot be…
Descriptors: Identification, Deception, Psychological Testing, Validity
Joakim Wallmark; James O. Ramsay; Juan Li; Marie Wiberg – Journal of Educational and Behavioral Statistics, 2024
Item response theory (IRT) models the relationship between the possible scores on a test item against a test taker's attainment of the latent trait that the item is intended to measure. In this study, we compare two models for tests with polytomously scored items: the optimal scoring (OS) model, a nonparametric IRT model based on the principles of…
Descriptors: Item Response Theory, Test Items, Models, Scoring
Daniel Murphy; Sarah Quesen; Matthew Brunetti; Quintin Love – Educational Measurement: Issues and Practice, 2024
Categorical growth models describe examinee growth in terms of performance-level category transitions, which implies that some percentage of examinees will be misclassified. This paper introduces a new procedure for estimating the classification accuracy of categorical growth models, based on Rudner's classification accuracy index for item…
Descriptors: Classification, Growth Models, Accuracy, Performance Based Assessment
Justin L. Kern – Journal of Educational and Behavioral Statistics, 2024
Given the frequent presence of slipping and guessing in item responses, models for the inclusion of their effects are highly important. Unfortunately, the most common model for their inclusion, the four-parameter item response theory model, potentially has severe deficiencies related to its possible unidentifiability. With this issue in mind, the…
Descriptors: Item Response Theory, Models, Bayesian Statistics, Generalization
Hojung Kim; Changkyung Song; Jiyoung Kim; Hyeyun Jeong; Jisoo Park – Language Testing in Asia, 2024
This study presents a modified version of the Korean Elicited Imitation (EI) test, designed to resemble natural spoken language, and validates its reliability as a measure of proficiency. The study assesses the correlation between average test scores and Test of Proficiency in Korean (TOPIK) levels, examining score distributions among beginner,…
Descriptors: Korean, Test Validity, Test Reliability, Imitation
Jiayi Deng – ProQuest LLC, 2024
Test score comparability in international large-scale assessments (LSA) is of utmost importance in measuring the effectiveness of education systems and understanding the impact of education on economic growth. To effectively compare test scores on an international scale, score linking is widely used to convert raw scores from different linguistic…
Descriptors: Item Response Theory, Scoring Rubrics, Scoring, Error of Measurement
William R. Nugent – Measurement: Interdisciplinary Research and Perspectives, 2024
Symmetry considerations are important in science, and Group Theory is a theory of symmetry. Classical Measurement Theory is the most used measurement theory in the social and behavioral sciences. In this article, the author uses Matrix Lie (Lee) group theory to formulate a measurement model. Symmetry is defined and illustrated using symmetries of…
Descriptors: Item Response Theory, Measurement Techniques, Models, Simulation
Jianbin Fu; Patrick C. Kyllonen; Xuan Tan – Measurement: Interdisciplinary Research and Perspectives, 2024
Users of forced-choice questionnaires (FCQs) to measure personality commonly assume statement parameter invariance across contexts -- between Likert and forced-choice (FC) items and between different FC items that share a common statement. In this paper, an empirical study was designed to check these two assumptions for an FCQ assessment measuring…
Descriptors: Measurement Techniques, Questionnaires, Personality Measures, Interpersonal Competence
Wind, Stefanie A.; Jones, Eli – Educational Researcher, 2019
Teacher evaluation systems often include classroom observations in which raters use rating scales to evaluate teachers' effectiveness. Recently, researchers have promoted the use of multifaceted approaches to investigating reliability using Generalizability theory, instead of rater reliability statistics. Generalizability theory allows analysts to…
Descriptors: Teacher Evaluation, Observation, Generalizability Theory, Item Response Theory
May, Toni A.; Koskey, Kristin L. K.; Bostic, Jonathan D.; Stone, Gregory E.; Kruse, Lance M.; Matney, Gabriel – School Science and Mathematics, 2023
Determining the most appropriate method of scoring an assessment is based on multiple factors, including the intended use of results, the assessment's purpose, and time constraints. Both the dichotomous and partial credit models have their advantages, yet direct comparisons of assessment outcomes from each method are not typical with constructed…
Descriptors: Scoring, Evaluation Methods, Problem Solving, Student Evaluation
Combs, Adam – Journal of Educational Measurement, 2023
A common method of checking person-fit in Bayesian item response theory (IRT) is the posterior-predictive (PP) method. In recent years, more powerful approaches have been proposed that are based on resampling methods using the popular L*[subscript z] statistic. There has also been proposed a new Bayesian model checking method based on pivotal…
Descriptors: Bayesian Statistics, Goodness of Fit, Evaluation Methods, Monte Carlo Methods
Toker, Türker; Seidel, Kent – International Journal of Contemporary Educational Research, 2023
Although the Rasch model is used to measure latent traits like attitude or ability where there are multiple latent structures within the dataset it is best to use a technique called the Mixture Rasch Model (MRM) which is a combination of a Rasch model and a latent class analysis (LCA). This study used data from a survey for teachers, teacher…
Descriptors: Item Response Theory, Beginning Teachers, Teacher Competencies, Teacher Effectiveness
Weese, James D.; Turner, Ronna C.; Liang, Xinya; Ames, Allison; Crawford, Brandon – Educational and Psychological Measurement, 2023
A study was conducted to implement the use of a standardized effect size and corresponding classification guidelines for polytomous data with the POLYSIBTEST procedure and compare those guidelines with prior recommendations. Two simulation studies were included. The first identifies new unstandardized test heuristics for classifying moderate and…
Descriptors: Effect Size, Classification, Guidelines, Statistical Analysis
He, Yinhong – Journal of Educational Measurement, 2023
Back random responding (BRR) behavior is one of the commonly observed careless response behaviors. Accurately detecting BRR behavior can improve test validities. Yu and Cheng (2019) showed that the change point analysis (CPA) procedure based on weighted residual (CPA-WR) performed well in detecting BRR. Compared with the CPA procedure, the…
Descriptors: Test Validity, Item Response Theory, Measurement, Monte Carlo Methods
van der Linden, Wim J.; Belov, Dmitry I. – Journal of Educational Measurement, 2023
A test of item compromise is presented which combines the test takers' responses and response times (RTs) into a statistic defined as the number of correct responses on the item for test takers with RTs flagged as suspicious. The test has null and alternative distributions belonging to the well-known family of compound binomial distributions, is…
Descriptors: Item Response Theory, Reaction Time, Test Items, Item Analysis