Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 21 |
Descriptor
Source
Author
Alonzo, Julie | 9 |
Tindal, Gerald | 9 |
Lai, Cheng Fei | 7 |
Adema, Jos J. | 6 |
Boekkooi-Timminga, Ellen | 6 |
van der Linden, Wim J. | 6 |
Armstrong, Ronald D. | 3 |
Stocking, Martha L. | 3 |
Avery, Marybell | 2 |
Dodd, Barbara G. | 2 |
Dyson, Ben | 2 |
More ▼ |
Publication Type
Reports - Evaluative | 85 |
Journal Articles | 34 |
Speeches/Meeting Papers | 18 |
Numerical/Quantitative Data | 10 |
Opinion Papers | 4 |
Information Analyses | 2 |
Books | 1 |
Collected Works - Proceedings | 1 |
Reports - Research | 1 |
Education Level
Elementary Education | 9 |
Elementary Secondary Education | 9 |
Grade 2 | 3 |
Grade 5 | 3 |
Kindergarten | 3 |
Early Childhood Education | 1 |
Grade 1 | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
Grade 6 | 1 |
Grade 7 | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Individuals with Disabilities… | 8 |
Assessments and Surveys
Law School Admission Test | 3 |
Armed Services Vocational… | 2 |
Graduate Management Admission… | 2 |
ACT Assessment | 1 |
California Achievement Tests | 1 |
Graduate Record Examinations | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Conejo, Ricardo; Guzmán, Eduardo; Trella, Monica – International Journal of Artificial Intelligence in Education, 2016
This article describes the evolution and current state of the domain-independent Siette assessment environment. Siette supports different assessment methods--including classical test theory, item response theory, and computer adaptive testing--and integrates them with multidimensional student models used by intelligent educational systems.…
Descriptors: Automation, Student Evaluation, Intelligent Tutoring Systems, Item Banks
Wetzel, Eunike; Hell, Benedikt; Passler, Katja – Journal of Career Assessment, 2012
Three test construction strategies are described and illustrated in the development of the Verb Interest Test (VIT), an inventory that assesses vocational interests using verbs. Verbs might be a promising alternative to the descriptions of occupational activities used in most vocational interest inventories because they are context-independent,…
Descriptors: Test Construction, Culture Fair Tests, Vocational Interests, Interest Inventories
Geerlings, Hanneke; van der Linden, Wim J.; Glas, Cees A. W. – Applied Psychological Measurement, 2013
Optimal test-design methods are applied to rule-based item generation. Three different cases of automated test design are presented: (a) test assembly from a pool of pregenerated, calibrated items; (b) test generation on the fly from a pool of calibrated item families; and (c) test generation on the fly directly from calibrated features defining…
Descriptors: Test Construction, Test Items, Item Banks, Automation
Arendasy, Martin E.; Sommer, Markus – Learning and Individual Differences, 2012
The use of new test administration technologies such as computerized adaptive testing in high-stakes educational and occupational assessments demands large item pools. Classic item construction processes and previous approaches to automatic item generation faced the problems of a considerable loss of items after the item calibration phase. In this…
Descriptors: Item Banks, Test Items, Adaptive Testing, Psychometrics
Preston, Kathleen; Reise, Steven; Cai, Li; Hays, Ron D. – Educational and Psychological Measurement, 2011
The authors used a nominal response item response theory model to estimate category boundary discrimination (CBD) parameters for items drawn from the Emotional Distress item pools (Depression, Anxiety, and Anger) developed in the Patient-Reported Outcomes Measurement Information Systems (PROMIS) project. For polytomous items with ordered response…
Descriptors: Item Response Theory, Models, Item Banks, Rating Scales
Yen, Yung-Chin; Ho, Rong-Guey; Laio, Wen-Wei; Chen, Li-Ju; Kuo, Ching-Chin – Applied Psychological Measurement, 2012
In a selected response test, aberrant responses such as careless errors and lucky guesses might cause error in ability estimation because these responses do not actually reflect the knowledge that examinees possess. In a computerized adaptive test (CAT), these aberrant responses could further cause serious estimation error due to dynamic item…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Response Style (Tests)
Zhu, Weimo; Fox, Connie; Park, Youngsik; Fisette, Jennifer L.; Dyson, Ben; Graber, Kim C.; Avery, Marybell; Franck, Marian; Placek, Judith H.; Rink, Judy; Raynes, De – Measurement in Physical Education and Exercise Science, 2011
The purpose of this study was to develop and calibrate an assessment system, or bank, using the latest measurement theories and methods to promote valid and reliable student assessment in physical education. Using an anchor-test equating design, a total of 30 items or assessments were administered to 5,021 (2,568 boys and 2,453 girls) students in…
Descriptors: Video Technology, Physical Education, Scoring Rubrics, Kindergarten
Fox, Connie; Zhu, Weimo; Park, Youngsik; Fisette, Jennifer L.; Graber, Kim C.; Dyson, Ben; Avery, Marybell; Franck, Marian; Placek, Judith H.; Rink, Judy; Raynes, De – Measurement in Physical Education and Exercise Science, 2011
In addition to validity and reliability evidence, other psychometric qualities of the PE Metrics assessments needed to be examined. This article describes how those critical psychometric issues were addressed during the PE Metrics assessment bank construction. Specifically, issues included (a) number of items or assessments needed, (b) training…
Descriptors: Measures (Individuals), Psychometrics, Interrater Reliability, Training
van der Linden, Wim J. – Applied Psychological Measurement, 2006
Two local methods for observed-score equating are applied to the problem of equating an adaptive test to a linear test. In an empirical study, the methods were evaluated against a method based on the test characteristic function (TCF) of the linear test and traditional equipercentile equating applied to the ability estimates on the adaptive test…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Format, Equated Scores
Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2009
In this technical report, we describe the development and piloting of a series of mathematics progress monitoring measures intended for use with students in grade 1. These measures, available as part of easyCBM [TM], an online progress monitoring assessment system, were developed in 2008 and administered to approximately 2800 students from schools…
Descriptors: Academic Achievement, Research Reports, Grade 1, Outcome Measures

van der Linden, Wim J.; Veldkamp, Bernard P.; Reese, Lynda M. – Applied Psychological Measurement, 2000
Presents an integer programming approach to item bank design that can be used to calculate an optimal blueprint for an item bank in order to support an existing testing program. Demonstrates the approach empirically using an item bank designed for the Law School Admission Test. (SLD)
Descriptors: Item Banks, Item Response Theory, Test Construction, Testing Programs

Armstrong, Ronald D.; Jones, Douglas H.; Kunce, Charles S. – Applied Psychological Measurement, 1998
Investigated the use of mathematical programming techniques to generate parallel test forms with passages and items based on item-response theory (IRT) using the Fundamentals of Engineering Examination. Generated four parallel test forms from the item bank of almost 1,100 items. Comparison with human-generated forms supports the mathematical…
Descriptors: Engineering, Item Banks, Item Response Theory, Test Construction
Eggen, Theo J. H. M.; Verschoor, Angela J. – Applied Psychological Measurement, 2006
Computerized adaptive tests (CATs) are individualized tests that, from a measurement point of view, are optimal for each individual, possibly under some practical conditions. In the present study, it is shown that maximum information item selection in CATs using an item bank that is calibrated with the one- or the two-parameter logistic model…
Descriptors: Adaptive Testing, Difficulty Level, Test Items, Item Response Theory
Huang, Chi-Yu; Kalohn, John C.; Lin, Chuan-Ju; Spray, Judith – 2000
Item pools supporting computer-based tests are not always completely calibrated. Occasionally, only a small subset of the items in the pool may have actual calibrations, while the remainder of the items may only have classical item statistics, (e.g., "p"-values, point-biserial correlation coefficients, or biserial correlation…
Descriptors: Classification, Computer Assisted Testing, Estimation (Mathematics), Item Banks
Kubinger, Klaus D. – International Journal of Testing, 2005
In this article, we emphasize that the Rasch model is not only very useful for psychological test calibration but is also necessary if the number of solved items is to be used as an examinee's score. Simplified proof that the Rasch model implies specific objective parameter comparisons is given. Consequently, a model check per se is possible. For…
Descriptors: Psychometrics, Psychological Testing, Item Banks, Item Response Theory