NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 11 results Save | Export
Peer reviewed Peer reviewed
Holmes, Susan E. – Journal of Educational Measurement, 1982
Two tests were created from a standardized reading achievement test and vertically equated using a sample of third and fourth grade students. Based on differences in ability estimates for the same student, the Rasch model did not provide a satisfactory means of vertical equating. (Author/CM)
Descriptors: Achievement Tests, Elementary Education, Equated Scores, Latent Trait Theory
Peer reviewed Peer reviewed
Congdon, Peter J.; McQueen, Joy – Journal of Educational Measurement, 2000
Studied the stability of rater severity over an extended rating period by applying multifaceted Rasch analysis to ratings of 16 raters of writing performances of 8,285 elementary school students. Findings cast doubt on the practice of using a single calibration of rate severity as the basis for adjustment of person measures. (SLD)
Descriptors: Educational Assessment, Elementary Education, Elementary School Students, Interrater Reliability
Peer reviewed Peer reviewed
Loyd, Brenda H.; Hoover, H. D. – Journal of Educational Measurement, 1980
Three levels of a mathematics computation test were equated using the Rasch model. Sixth, seventh, and eighth graders were administered different levels of the test. Lack of consistency among equatings suggested that the Rasch model did not produce a satisfactory vertical equating of this computation test. (Author/RD)
Descriptors: Ability Grouping, Achievement Tests, Elementary Education, Equated Scores
Peer reviewed Peer reviewed
Hoskens, Machteld; De Boeck, Paul – Journal of Educational Measurement, 1995
With the nominal model of R. D. Bock as a framework, transformation matrices can be used to constrain the parameters of test item response categories to reflect their componential design. The approach is illustrated in a study of a spelling task completed by 385 Dutch third graders. (SLD)
Descriptors: Cognitive Tests, Elementary Education, Elementary School Students, Foreign Countries
Peer reviewed Peer reviewed
Tatsuoka, Kikumi K. – Journal of Educational Measurement, 1983
A newly introduced approach, rule space, can represent large numbers of erroneous rules of arithmetic operations quantitatively and can predict the likelihood of each erroneous rule. The new model challenges the credibility of the traditional right-or-wrong scoring procedure. (Author/PN)
Descriptors: Addition, Algorithms, Arithmetic, Diagnostic Tests
Peer reviewed Peer reviewed
Miller, M. David; Linn, Robert L. – Journal of Educational Measurement, 1988
Effects of instructional coverage variations on item characteristic functions were examined, using eighth-grade data from the Second International Mathematics Study (1985). Although some differences in item response curves were large, better performance was not necessarily related to greater learning opportunity. Item curve response differences…
Descriptors: Achievement Tests, Black Students, Elementary Education, Grade 8
Peer reviewed Peer reviewed
Fitzpatrick, Anne R.; Yen, Wendy M. – Journal of Educational Measurement, 1995
The psychometric characteristics of constructed response items referring to choice and nonchoice passages administered to approximately 150,000 students in grades 3, 5, and 8 were studied through item response theory methodology. Results indicated no consistent differences in the difficulty and discrimination of items referring to the two types of…
Descriptors: Constructed Response, Difficulty Level, Elementary Education, Elementary School Students
Peer reviewed Peer reviewed
Miller, M. David – Journal of Educational Measurement, 1986
An index of student patterns of item response, when aggregated to the class level, was shown to identify classes that have a poor match between test content and instructional coverage. The mean caution index for a class can best be interpreted knowing the within-class standard deviation of the index. (Author/LMO)
Descriptors: Classes (Groups of Students), Elementary Education, Error Patterns, Goodness of Fit
Peer reviewed Peer reviewed
Lane, Suzanne; And Others – Journal of Educational Measurement, 1996
Evidence from test results of 3,604 sixth and seventh graders is provided for the generalizability and validity of the Quantitative Understanding: Amplifying Student Achievement and Reasoning (QUASAR) Cognitive Assessment Instrument, which is designed to measure program outcomes and growth in mathematics. (SLD)
Descriptors: Achievement Tests, Cognitive Processes, Elementary Education, Elementary School Students
Peer reviewed Peer reviewed
Phillips, S. E.; Mehrens, William A. – Journal of Educational Measurement, 1987
This study investigates whether a linear factor analytic method commonly used to evaluate violation of the item response theory (IRT) unidimensionality assumption is sensitive to measurable curricular differences within a school district and to examine the possibility of differential item performance for groups of students receiving different…
Descriptors: Achievement Tests, Elementary Education, Elementary School Curriculum, Factor Analysis
Peer reviewed Peer reviewed
Crocker, Linda; And Others – Journal of Educational Measurement, 1988
Using generalizability theory as a framework, the problem of assessing the content validity of standardized achievement tests is considered. Four designs to assess test-item fit to a curriculum are described, and procedures for determining the optimal number of raters and schools in a content-validation decision-making study are considered. (TJH)
Descriptors: Achievement Tests, Content Validity, Decision Making, Elementary Education