Publication Date
In 2025 | 0 |
Since 2024 | 6 |
Since 2021 (last 5 years) | 33 |
Since 2016 (last 10 years) | 64 |
Since 2006 (last 20 years) | 87 |
Descriptor
Source
Author
Huggins-Manley, Anne Corinne | 3 |
Kam, Chester Chun Seng | 3 |
Plake, Barbara S. | 3 |
Ames, Allison J. | 2 |
Bejar, Isaac I. | 2 |
Belov, Dmitry I. | 2 |
Benson, Jeri | 2 |
Bolt, Daniel M. | 2 |
Braeken, Johan | 2 |
Bulut, Okan | 2 |
Cheng, Ying | 2 |
More ▼ |
Publication Type
Education Level
Secondary Education | 21 |
Higher Education | 13 |
Elementary Education | 10 |
Postsecondary Education | 9 |
Middle Schools | 8 |
Intermediate Grades | 7 |
Grade 3 | 5 |
Grade 4 | 5 |
Grade 5 | 5 |
High Schools | 5 |
Primary Education | 5 |
More ▼ |
Audience
Researchers | 10 |
Practitioners | 3 |
Teachers | 1 |
Location
Germany | 8 |
Canada | 6 |
Australia | 5 |
United States | 4 |
California | 3 |
Finland | 3 |
Netherlands | 3 |
United Kingdom | 3 |
Belgium | 2 |
China | 2 |
Denmark | 2 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Perrin, David W.; Kerasotes, Dean L. – 1979
It was hypothesized that using asterisks as attention focusing devices would cause students to read all asteriked test items more carefully and would improve test scores of undergraduate education students. Sixty-three undergraduates majoring in elementary or special education were administered a 36-item objective test. Asterisks were used to…
Descriptors: Difficulty Level, Higher Education, Objective Tests, Response Style (Tests)

McMorris, Robert F.; And Others – Journal of Educational Measurement, 1987
Consistency of gain from changing test answers was tested for students instructed about answer-changing research results, and composition of the gain was analyzed by examining the students' reasons for changing. Mean gain remained positive and consistent with gain for previously studied uninstructed groups; amount of change was also stable.…
Descriptors: Difficulty Level, Graduate Students, Higher Education, Instruction

Bardo, John W.; Yeager, Samuel J. – Perceptual and Motor Skills, 1982
Responses to various fixed test-response formats were examined for "reliability" due to systematic error; Cronbach's alphas up to .67 were obtained. Of formats tested, four-point Likert Scales were least affected while forms of lines and faces were most problematic. Possible modification in alpha to account for systematic bias is…
Descriptors: Higher Education, Measures (Individuals), Psychometrics, Response Style (Tests)

Veale, James R.; Foreman, Dale I. – Journal of Educational Measurement, 1983
Statistical procedures for measuring heterogeneity of test item distractor distributions, or cultural variation, are presented. These procedures are based on the notion that examinees' responses to the incorrect options of a multiple-choice test provide more information concerning cultural bias than their correct responses. (Author/PN)
Descriptors: Ethnic Bias, Item Analysis, Mathematical Models, Multiple Choice Tests

Frisbie, David A.; Brandenburg, Dale C. – Journal of Educational Measurement, 1979
Content-parallel questionnaire items in which response schemes varied in one of two ways--scale alternatives were all defined or only endpoints were defined, and alternatives were numbered or lettered--were investigated on a large sample of college freshmen. (Author/JKS)
Descriptors: Higher Education, Item Analysis, Questionnaires, Rating Scales
Swearingen, Dorothy L. – 1998
When response set is present, instead of responding to the intent of the question, the subject appears to be responding to a variable emanating from some personal characteristic. This threat to measurement reliability and validity warrants investigation of the source of response set so that questionnaire designers can minimize its occurrence. This…
Descriptors: Attitude Measures, College Students, Higher Education, Item Response Theory
Siskind, Theresa G.; Anderson, Lorin W. – 1982
The study was designed to examine the similarity of response options generated by different item writers using a systematic approach to item writing. The similarity of response options to student responses for the same item stems presented in an open-ended format was also examined. A non-systematic (subject matter expertise) approach and a…
Descriptors: Algorithms, Item Analysis, Multiple Choice Tests, Quality Control
Marshall, Sandra P. – 1981
The differences in responses of grade 6 boys and girls to selected items on a standardized test are examined. Data are the responses of approximately 36,000 children during 1978-79 to two test forms of the California Assessment Program's Survey of Basic Skills: Grade 6 (SBS:6). Performance on specific arithmetic items is studied and involves the…
Descriptors: Academic Achievement, Arithmetic, Basic Skills, Computation
Benson, Jeri; Wilcox, Stuart – 1981
What effect varying the phrasing of attitude statements would have across sex and grade was determined. Three content parallel forms were developed to assess student attitudes toward integration. Each form contained 15 items and differed only in terms of item phrasing either positive, negative or a combination of both. Six hundred and twenty-two,…
Descriptors: Age Differences, Intermediate Grades, Racial Integration, Response Style (Tests)

Poizner, Sharon B.; And Others – Applied Psychological Measurement, 1978
Binary, probability, and ordinal scoring procedures for multiple-choice items were examined. In two situations, it was found that both the probability and ordinal scoring systems were more reliable than the binary scoring method. (Author/CTM)
Descriptors: Confidence Testing, Guessing (Tests), Higher Education, Multiple Choice Tests

Huck, Schuyler W. – Journal of Educational Measurement, 1978
Providing examinees with advanced knowledge of the difficulty of an item led to an increase in test performance with no loss of reliability. This finding was consistent across several test formats. ( Author/JKS)
Descriptors: Difficulty Level, Feedback, Higher Education, Item Analysis

Velicer, Wayne F.; Stevenson, John F. – Applied Psychological Measurement, 1978
A Likert seven-choice response format for personality inventories allows finer distinctions by subjects than the traditional two-choice format. The Eysenck Personality Inventory was employed in the present study to test the hypothesis that use of the expanded format would result in a clearer and more accurate indication of test structure.…
Descriptors: Forced Choice Technique, Higher Education, Personality Measures, Rating Scales

Ace, Merle C.; Dawis, Rene V. – Educational and Psychological Measurement, 1973
Because no previous study was found in which both blank position in the item stem and positional placement of the correct response were studied simultaneously, it was decided to investigate the influence of these two factors, alone and in combination, on the difficulty level of verbal analogy items. (Authors)
Descriptors: Analysis of Variance, Data Analysis, Difficulty Level, Disadvantaged
Holland, Paul W.; Thayer, Dorothy T. – 1986
The Mantel-Haenszel procedure (MH) is a practical, inexpensive, and powerful way to detect test items that function differently in two groups of examinees. MH is a natural outgrowth of previously suggested chi square methods, and it is also related to methods based on item response theory. The study of items that function differently for two…
Descriptors: Comparative Analysis, Hypothesis Testing, Item Analysis, Latent Trait Theory
Montague, William E. – 1980
A number of examples are presented to illustrate a common flaw in the published research on learning, memory, and instruction. Experimental subjects--often college students--have certain expectations about the problems they will be asked to solve and about the questions that will appear on reading comprehension or recall tests; these expectations…
Descriptors: Advance Organizers, Correlation, Educational Research, Expectation