NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 151 to 165 of 262 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Alderson, J. Charles; Figueras, Neus; Kuijper, Henk; Nold, Guenter; Takala, Sauli; Tardieu, Claire – Language Assessment Quarterly, 2006
The Common European Framework of Reference (CEFR) is intended as a reference document for language education including assessment. This article describes a project that investigated whether the CEFR can help test developers construct reading and listening tests based on CEFR levels. If the CEFR scales together with the detailed description of…
Descriptors: Test Content, Listening Comprehension Tests, Classification, Test Construction
Childs, Ruth A.; Jaciw, Andrew P. – 2003
This Digest describes matrix sampling of test items as an approach to achieving broad coverage while minimizing testing time per student. Matrix sampling involves developing a complete set of items judged to cover the curriculum, then dividing the items into subsets and administering one subset to each student. Matrix sampling, by limiting the…
Descriptors: Item Banks, Matrices, Sampling, Test Construction
Leung, Chi-Keung; Chang, Hua-Hua; Hau, Kit-Tai – 2000
Item selection methods in computerized adaptive testing (CAT) can yield extremely skewed item exposure distribution in which items with high "a" values may be over-exposed while those with low "a" values may never be selected. H. Chang and Z. Ying (1999) proposed the a-stratified design (ASTR) that attempts to equalize item…
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Test Construction
Wei, Xin; Shen, Xuejun; Lukoff, Brian; Ho, Andrew Dean; Haertel, Edward H. – Online Submission, 2006
In 1998 and again in 2002, samples of eighth grade students in California were tested in reading as part of the state-level component of the National Assessment of Educational Progress (NAEP). In each of these years, all eighth graders in the state were also required to participate in the state's accountability testing, which included the reading…
Descriptors: Grade 8, Test Content, Reading Tests, Accountability
Kromrey, Jeffrey D.; Parshall, Cynthia G.; Yi, Qing – 1998
The effects of anchor test characteristics in the accuracy and precision of test equating in the "common items, nonequivalent groups design" were studied. The study also considered the effects of nonparallel based and new forms on the equating solution, and it investigated the effects of differential weighting on the success of equating…
Descriptors: Equated Scores, High Schools, Item Response Theory, Monte Carlo Methods
Leung, Chi-Keung; Chang, Hua-Hua; Hau, Kit-Tai – 2001
The multistage alpha-stratified computerized adaptive testing (CAT) design advocated a new philosophy of pool management and item selection using low discriminating items first. It has been demonstrated through simulation studies to be effective both in reducing item overlap rate and enhancing pool utilization with certain pool types. Based on…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Selection
Yang, Wen-Ling – 2000
This study investigated whether equating accuracy improves with an anchor test that is more representative of its corresponding total test and whether such content effect depends on the particular equating method used. Scoring outcomes of a professional examination for a medical specialty were used. A total of 1,092 examinees took one form, and…
Descriptors: Equated Scores, Item Response Theory, Licensing Examinations (Professions), Physicians
Peer reviewed Peer reviewed
Kingsbury, G. Gage; Zara, Anthony R. – Applied Measurement in Education, 1991
This simulation investigated two procedures that reduce differences between paper-and-pencil testing and computerized adaptive testing (CAT) by making CAT content sensitive. Results indicate that the price in terms of additional test items of using constrained CAT for content balancing is much smaller than that of using testlets. (SLD)
Descriptors: Adaptive Testing, Comparative Analysis, Computer Assisted Testing, Computer Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J.; Ariel, Adelaide; Veldkamp, Bernard P. – Journal of Educational and Behavioral Statistics, 2006
Test-item writing efforts typically results in item pools with an undesirable correlational structure between the content attributes of the items and their statistical information. If such pools are used in computerized adaptive testing (CAT), the algorithm may be forced to select items with less than optimal information, that violate the content…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Items, Item Banks
Yamamoto, Kentaro; Kulick, Edward – 1992
Test items are designed to be representative of the subject areas that they measure and to reflect the importance of specific domains or item types within those subject areas. Content validity is achieved by content specification and number of items in each content domain included in the design of the test. However, largely due to the normal…
Descriptors: Content Validity, Elementary Secondary Education, Field Tests, Mathematical Models
Luecht, Richard M. – 2003
This paper presents a multistage adaptive testing test development paradigm that promises to handle content balancing and other test development needs, psychometric reliability concerns, and item exposure. The bundled multistage adaptive testing (BMAT) framework is a modification of the computer-adaptive sequential testing framework introduced by…
Descriptors: Adaptive Testing, Computer Assisted Testing, High Stakes Tests, Mastery Tests
Georgia State Dept. of Education, Atlanta. – 1999
This document is a supplement to the Test Content Description for the Georgia High School Test in Mathematics. The sample items are representative of items on the actual Georgia High School Tests. The strands (major areas) being tested are: (1) Number and Composition (17% to 19% of the test); (2) Data Analysis (19-21% of the test); (3) Measurement…
Descriptors: High School Students, High Schools, Mathematics, Standardized Tests
Georgia State Dept. of Education, Atlanta. – 1999
This document is a supplement to the Test Content Description for the Georgia High School Test in Science. The sample items are representative of items on the actual Georgia High School Tests. The strands (major areas) being tested are: (1) Process/Research Skills (30% to 32% of the test); (2) Physical Science (33% to 35% of the test); and…
Descriptors: High School Students, High Schools, Sciences, Standardized Tests
US Department of Education, 2004
This demonstration booklet describes the 2005 assessments. The first three sections explain the purposes and content of the NAEP reading, mathematics, and science assessments and provide sample test questions at each grade level. The fourth section includes the directions and the background questions that all students will be asked to answer. The…
Descriptors: Grade 8, Integrity, National Competency Tests, Student Evaluation
Peer reviewed Peer reviewed
Eignor, Daniel R. – Journal of Educational Measurement, 1997
The authors of the "Guidelines," a task force of eight, intend to present an organized list of features to be considered in reporting or evaluating computerized-adaptive assessments. Apart from a few weaknesses, the book is a useful and complete document that will be very helpful to test developers. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Evaluation Methods, Guidelines
Pages: 1  |  ...  |  7  |  8  |  9  |  10  |  11  |  12  |  13  |  14  |  15  |  ...  |  18