NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)0
Since 2006 (last 20 years)17
Education Level
Elementary Education1
Audience
Researchers1
Location
Netherlands1
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Showing 1 to 15 of 38 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Mittelhaëuser, Marie-Anne; Béguin, Anton A.; Sijtsma, Klaas – Journal of Educational Measurement, 2015
The purpose of this study was to investigate whether simulated differential motivation between the stakes for operational tests and anchor items produces an invalid linking result if the Rasch model is used to link the operational tests. This was done for an external anchor design and a variation of a pretest design. The study also investigated…
Descriptors: Item Response Theory, Simulation, High Stakes Tests, Pretesting
Peer reviewed Peer reviewed
Direct linkDirect link
Straat, J. Hendrik; van der Ark, L. Andries; Sijtsma, Klaas – Educational and Psychological Measurement, 2014
An automated item selection procedure in Mokken scale analysis partitions a set of items into one or more Mokken scales, if the data allow. Two algorithms are available that pursue the same goal of selecting Mokken scales of maximum length: Mokken's original automated item selection procedure (AISP) and a genetic algorithm (GA). Minimum…
Descriptors: Sampling, Test Items, Effect Size, Scaling
Peer reviewed Peer reviewed
Direct linkDirect link
Tijmstra, Jesper; Hessen, David J.; van der Heijden, Peter G. M.; Sijtsma, Klaas – Psychometrika, 2013
Most dichotomous item response models share the assumption of latent monotonicity, which states that the probability of a positive response to an item is a nondecreasing function of a latent variable intended to be measured. Latent monotonicity cannot be evaluated directly, but it implies manifest monotonicity across a variety of observed scores,…
Descriptors: Item Response Theory, Statistical Inference, Probability, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Zijlstra, Wobbe P.; van der Ark, L. Andries; Sijtsma, Klaas – Multivariate Behavioral Research, 2011
Exploratory Mokken scale analysis (MSA) is a popular method for identifying scales from larger sets of items. As with any statistical method, in MSA the presence of outliers in the data may result in biased results and wrong conclusions. The forward search algorithm is a robust diagnostic method for outlier detection, which we adapt here to…
Descriptors: Measures (Individuals), Statistical Analysis, Mathematics, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Bouwmeester, Samantha; Vermunt, Jeroen K.; Sijtsma, Klaas – Cognitive Development, 2012
We discuss the limitations of hypothesis testing using (quasi-) experiments in the study of cognitive development and suggest latent variable modeling as a viable alternative to experimentation. Latent variable models allow testing a theory as a whole, incorporating individual differences with respect to developmental processes or abilities in the…
Descriptors: Age, Testing, Individual Differences, Hypothesis Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Tijmstra, Jesper; Hessen, David J.; van der Heijden, Peter G. M.; Sijtsma, Klaas – Psychometrika, 2011
A new observable consequence of the property of invariant item ordering is presented, which holds under Mokken's double monotonicity model for dichotomous data. The observable consequence is an invariant ordering of the item-total regressions. Kendall's measure of concordance "W" and a weighted version of this measure are proposed as measures for…
Descriptors: Item Response Theory, Bayesian Statistics, Regression (Statistics), Models
Peer reviewed Peer reviewed
Direct linkDirect link
Ligtvoet, Rudy; van der Ark, L. Andries; Bergsma, Wicher P.; Sijtsma, Klaas – Psychometrika, 2011
We propose three latent scales within the framework of nonparametric item response theory for polytomously scored items. Latent scales are models that imply an invariant item ordering, meaning that the order of the items is the same for each measurement value on the latent scale. This ordering property may be important in, for example,…
Descriptors: Intelligence Tests, Measures (Individuals), Methods, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Sijtsma, Klaas – International Journal of Testing, 2009
This article reviews three topics from test theory that continue to raise discussion and controversy and capture test theorists' and constructors' interest. The first topic concerns the discussion of the methodology of investigating and establishing construct validity; the second topic concerns reliability and its misuse, alternative definitions…
Descriptors: Construct Validity, Reliability, Classification, Test Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Ligtvoet, Rudy; van der Ark, L. Andries; te Marvelde, Janneke M.; Sijtsma, Klaas – Educational and Psychological Measurement, 2010
This article discusses the concept of an invariant item ordering (IIO) for polytomously scored items and proposes methods for investigating an IIO in real test data. Method manifest IIO is proposed for assessing whether item response functions intersect. Coefficient H[superscript T] is defined for polytomously scored items. Given that an IIO…
Descriptors: Item Response Theory, Data Analysis, Evaluation Methods, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Sijtsma, Klaas – Psychometrika, 2009
The critical reactions of Bentler (2009, doi: 10.1007/s11336-008-9100-1), Green and Yang (2009a, doi: 10.1007/s11336-008-9098-4 ; 2009b, doi: 10.1007/s11336-008-9099-3), and Revelle and Zinbarg (2009, doi: 10.1007/s11336-008-9102-z) to Sijtsma's (2009, doi: 10.1007/s11336-008-9101-0) paper on Cronbach's alpha are addressed. The dissemination of…
Descriptors: Psychometrics, Reliability, Theory Practice Relationship, Structural Equation Models
Peer reviewed Peer reviewed
Direct linkDirect link
Evers, Arne; Sijtsma, Klaas; Lucassen, Wouter; Meijer, Rob R. – International Journal of Testing, 2010
This article describes the 2009 revision of the Dutch Rating System for Test Quality and presents the results of test ratings from almost 30 years. The rating system evaluates the quality of a test on seven criteria: theoretical basis, quality of the testing materials, comprehensiveness of the manual, norms, reliability, construct validity, and…
Descriptors: Rating Scales, Documentation, Educational Quality, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Conijn, Judith M.; Emons, Wilco H. M.; van Assen, Marcel A. L. M.; Sijtsma, Klaas – Multivariate Behavioral Research, 2011
The logistic person response function (PRF) models the probability of a correct response as a function of the item locations. Reise (2000) proposed to use the slope parameter of the logistic PRF as a person-fit measure. He reformulated the logistic PRF model as a multilevel logistic regression model and estimated the PRF parameters from this…
Descriptors: Monte Carlo Methods, Patients, Probability, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
van der Ark, L. Andries; Croon, Marcel A.; Sijtsma, Klaas – Psychometrika, 2008
Scalability coefficients play an important role in Mokken scale analysis. For a set of items, scalability coefficients have been defined for each pair of items, for each individual item, and for the entire scale. Hypothesis testing with respect to these scalability coefficients has not been fully developed. This study introduces marginal modelling…
Descriptors: Hypothesis Testing, Item Response Theory, Error of Measurement, Scaling
Peer reviewed Peer reviewed
Direct linkDirect link
van Ginkel, Joost R.; van der Ark, L. Andries; Sijtsma, Klaas – Multivariate Behavioral Research, 2007
The performance of five simple multiple imputation methods for dealing with missing data were compared. In addition, random imputation and multivariate normal imputation were used as lower and upper benchmark, respectively. Test data were simulated and item scores were deleted such that they were either missing completely at random, missing at…
Descriptors: Evaluation Methods, Psychometrics, Item Response Theory, Scores
Peer reviewed Peer reviewed
Emons, Wilco H. M.; Meijer, Rob R.; Sijtsma, Klaas – Applied Psychological Measurement, 2002
Studied whether the theoretical sampling distribution of the U3 person-fit statistic is in agreement with the simulated sampling distribution under different item response theory models and varying item and test characteristics. Simulation results suggest that the use of standard normal deviates for the standardized version of the U3 statistic may…
Descriptors: Item Response Theory, Sampling, Simulation, Statistical Distributions
Previous Page | Next Page »
Pages: 1  |  2  |  3