NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)0
Since 2006 (last 20 years)13
Education Level
Higher Education3
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 44 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Egberink, Iris J. L.; Meijer, Rob R.; Tendeiro, Jorge N. – Educational and Psychological Measurement, 2015
A popular method to assess measurement invariance of a particular item is based on likelihood ratio tests with all other items as anchor items. The results of this method are often only reported in terms of statistical significance, and researchers proposed different methods to empirically select anchor items. It is unclear, however, how many…
Descriptors: Personality Measures, Computer Assisted Testing, Measurement, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Tendeiro, Jorge N.; Meijer, Rob R. – Journal of Educational Measurement, 2014
In recent guidelines for fair educational testing it is advised to check the validity of individual test scores through the use of person-fit statistics. For practitioners it is unclear on the basis of the existing literature which statistic to use. An overview of relatively simple existing nonparametric approaches to identify atypical response…
Descriptors: Educational Assessment, Test Validity, Scores, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Meijer, Rob R.; Egberink, Iris J. L. – Educational and Psychological Measurement, 2012
In recent studies, different methods were proposed to investigate invariant item ordering (IIO), but practical IIO research is an unexploited field in questionnaire construction and evaluation. In the present study, the authors explored the usefulness of different IIO methods to analyze personality scales and clinical scales. From the authors'…
Descriptors: Test Items, Personality Measures, Questionnaires, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Tendeiro, Jorge N.; Meijer, Rob R. – Applied Psychological Measurement, 2013
To classify an item score pattern as not fitting a nonparametric item response theory (NIRT) model, the probability of exceedance (PE) of an observed response vector x can be determined as the sum of the probabilities of all response vectors that are, at most, as likely as x, conditional on the test's total score. Vector x is to be considered…
Descriptors: Probability, Nonparametric Statistics, Goodness of Fit, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
Meijer, Rob R.; Tendeiro, Jorge N. – Journal of Educational and Behavioral Statistics, 2012
We extend a recent didactic by Magis, Raiche, and Beland on the use of the l[subscript z] and l[subscript z]* person-fit statistics. We discuss a number of possibly confusing details and show that it is important to first investigate item response theory model fit before assessing person fit. Furthermore, it is argued that appropriate…
Descriptors: Item Response Theory, Statistics, Goodness of Fit, Personality Measures
Peer reviewed Peer reviewed
Direct linkDirect link
Smits, Iris A. M.; Timmerman, Marieke E.; Meijer, Rob R. – Applied Psychological Measurement, 2012
The assessment of the number of dimensions and the dimensionality structure of questionnaire data is important in scale evaluation. In this study, the authors evaluate two dimensionality assessment procedures in the context of Mokken scale analysis (MSA), using a so-called fixed lowerbound. The comparative simulation study, covering various…
Descriptors: Simulation, Measures (Individuals), Program Effectiveness, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Meijer, Rob R.; de Vries, Rivka M.; van Bruggen, Vincent – Psychological Assessment, 2011
The psychometric structure of the Brief Symptom Inventory-18 (BSI-18; Derogatis, 2001) was investigated using Mokken scaling and parametric item response theory. Data of 487 outpatients, 266 students, and 207 prisoners were analyzed. Results of the Mokken analysis indicated that the BSI-18 formed a strong Mokken scale for outpatients and…
Descriptors: Measures (Individuals), Psychometrics, Depression (Psychology), Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Tendeiro, Jorge N.; Meijer, Rob R. – Applied Psychological Measurement, 2012
This article extends the work by Armstrong and Shi on CUmulative SUM (CUSUM) person-fit methodology. The authors present new theoretical considerations concerning the use of CUSUM person-fit statistics based on likelihood ratios for the purpose of detecting cheating and random guessing by individual test takers. According to the Neyman-Pearson…
Descriptors: Cheating, Individual Testing, Adaptive Testing, Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Egberink, Iris J. L.; Meijer, Rob R. – Assessment, 2011
The authors investigated the psychometric properties of the subscales of the Self-Perception Profile for Children with item response theory (IRT) models using a sample of 611 children. Results from a nonparametric Mokken analysis and a parametric IRT approach for boys (n = 268) and girls (n = 343) were compared. The authors found that most scales…
Descriptors: Profiles, Psychometrics, Item Response Theory, Self Concept
Peer reviewed Peer reviewed
Direct linkDirect link
Evers, Arne; Sijtsma, Klaas; Lucassen, Wouter; Meijer, Rob R. – International Journal of Testing, 2010
This article describes the 2009 revision of the Dutch Rating System for Test Quality and presents the results of test ratings from almost 30 years. The rating system evaluates the quality of a test on seven criteria: theoretical basis, quality of the testing materials, comprehensiveness of the manual, norms, reliability, construct validity, and…
Descriptors: Rating Scales, Documentation, Educational Quality, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Fox, Jean-Paul; Meijer, Rob R. – Applied Psychological Measurement, 2008
The authors discuss a new method that combines the randomized response technique with item response theory. This method allows the researcher to obtain information at the individual person level without knowing the true responses. With this new method, it is possible to compare groups of individuals by means of analysis of variance or regression…
Descriptors: Psychological Evaluation, Educational Assessment, Item Response Theory, Statistical Analysis
Peer reviewed Peer reviewed
Emons, Wilco H. M.; Meijer, Rob R.; Sijtsma, Klaas – Applied Psychological Measurement, 2002
Studied whether the theoretical sampling distribution of the U3 person-fit statistic is in agreement with the simulated sampling distribution under different item response theory models and varying item and test characteristics. Simulation results suggest that the use of standard normal deviates for the standardized version of the U3 statistic may…
Descriptors: Item Response Theory, Sampling, Simulation, Statistical Distributions
Peer reviewed Peer reviewed
Meijer, Rob R.; Sijtsma, Klaas – Applied Psychological Measurement, 2001
Reviews person-fit methods based on classical test theory and item response theory and the methods investigating particular types of response behavior on tests. Discusses similarities and differences among person-fit methods and their advantages and disadvantages. The usefulness of person-fit statistics for improving measurement depends on the…
Descriptors: Item Response Theory, Research Methodology
Peer reviewed Peer reviewed
Direct linkDirect link
Reise, Steven P.; Meijer, Rob R.; Ainsworth, Andrew T.; Morales, Leo S.; Hays, Ron D. – Multivariate Behavioral Research, 2006
Group-level parametric and non-parametric item response theory models were applied to the Consumer Assessment of Healthcare Providers and Systems (CAHPS[R]) 2.0 core items in a sample of 35,572 Medicaid recipients nested within 131 health plans. Results indicated that CAHPS responses are dominated by within health plan variation, and only weakly…
Descriptors: Item Response Theory, Psychometrics, Sample Size, Medical Care Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Emons, Wilco H. M.; Sijtsma, Klaas; Meijer, Rob R. – Psychological Methods, 2007
Short tests containing at most 15 items are used in clinical and health psychology, medicine, and psychiatry for making decisions about patients. Because short tests have large measurement error, the authors ask whether they are reliable enough for classifying patients into a treatment and a nontreatment group. For a given certainty level,…
Descriptors: Psychiatry, Patients, Error of Measurement, Test Length
Previous Page | Next Page ยป
Pages: 1  |  2  |  3