Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 14 |
Since 2006 (last 20 years) | 69 |
Descriptor
Source
Author
Publication Type
Education Level
Higher Education | 23 |
Postsecondary Education | 9 |
Secondary Education | 8 |
Elementary Education | 7 |
High Schools | 7 |
Elementary Secondary Education | 4 |
Grade 8 | 3 |
Middle Schools | 3 |
Grade 10 | 2 |
Grade 3 | 2 |
Grade 5 | 2 |
More ▼ |
Location
Texas | 3 |
United Kingdom | 3 |
Australia | 2 |
California | 2 |
Oregon | 2 |
United Kingdom (England) | 2 |
Utah | 2 |
Alaska | 1 |
Canada | 1 |
China | 1 |
Germany | 1 |
More ▼ |
Laws, Policies, & Programs
Elementary and Secondary… | 1 |
Elementary and Secondary… | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Han, Suhwa; Kang, Hyeon-Ah – Journal of Educational Measurement, 2023
The study presents multivariate sequential monitoring procedures for examining test-taking behaviors online. The procedures monitor examinee's responses and response times and signal aberrancy as soon as significant change is identifieddetected in the test-taking behavior. The study in particular proposes three schemes to track different…
Descriptors: Test Wiseness, Student Behavior, Item Response Theory, Computer Assisted Testing
Semere Kiros Bitew; Amir Hadifar; Lucas Sterckx; Johannes Deleu; Chris Develder; Thomas Demeester – IEEE Transactions on Learning Technologies, 2024
Multiple-choice questions (MCQs) are widely used in digital learning systems, as they allow for automating the assessment process. However, owing to the increased digital literacy of students and the advent of social media platforms, MCQ tests are widely shared online, and teachers are continuously challenged to create new questions, which is an…
Descriptors: Multiple Choice Tests, Computer Assisted Testing, Test Construction, Test Items
Livingston, Samuel A. – Educational Testing Service, 2020
This booklet is a conceptual introduction to item response theory (IRT), which many large-scale testing programs use for constructing and scoring their tests. Although IRT is essentially mathematical, the approach here is nonmathematical, in order to serve as an introduction on the topic for people who want to understand why IRT is used and what…
Descriptors: Item Response Theory, Scoring, Test Items, Scaling
Steedle, Jeffrey T.; Cho, Young Woo; Wang, Shichao; Arthur, Ann M.; Li, Dongmei – Educational Measurement: Issues and Practice, 2022
As testing programs transition from paper to online testing, they must study mode comparability to support the exchangeability of scores from different testing modes. To that end, a series of three mode comparability studies was conducted during the 2019-2020 academic year with examinees randomly assigned to take the ACT college admissions exam on…
Descriptors: College Entrance Examinations, Computer Assisted Testing, Scores, Test Format
Daniel R. Isbell; Benjamin Kremmel; Jieun Kim – Language Assessment Quarterly, 2023
In the wake of the COVID-19 boom in remote administration of language tests, it appears likely that remote administration will be a permanent fixture in the language testing landscape. Accordingly, language test providers, stakeholders, and researchers must grapple with the implications of remote proctoring on valid, fair, and just uses of tests.…
Descriptors: Distance Education, Supervision, Language Tests, Culture Fair Tests
Cui, Zhongmin; Liu, Chunyan; He, Yong; Chen, Hanwei – Journal of Educational Measurement, 2018
Allowing item review in computerized adaptive testing (CAT) is getting more attention in the educational measurement field as more and more testing programs adopt CAT. The research literature has shown that allowing item review in an educational test could result in more accurate estimates of examinees' abilities. The practice of item review in…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Test Wiseness
Leighton, Jacqueline P. – British Journal of Educational Psychology, 2019
Background and Aims: In educational measurement, performance assessments occupy a niche for offering a true-to-life format that affords the measurement of high-level cognitive competencies and the evidence to draw inferences about intellectual capital. However, true-to-life formats also introduce myriad complexities and can skew if not outright…
Descriptors: Performance Based Assessment, Cognitive Processes, Inferences, Accuracy
Wise, Steven L. – Educational Research and Evaluation, 2020
There has been increasing concern about the presence of disengaged test taking in international assessment programmes and its implications for the validity of inferences made regarding a country's level of educational attainment. This issue has received a growing research interest over the past 20 years, with notable advances in both the…
Descriptors: Learner Engagement, Test Wiseness, Student Motivation, Achievement Tests
Babcock, Sarah E.; Wilson, Claire A.; Lau, Chloe – Canadian Journal of School Psychology, 2018
This article describes and reviews The School Motivation and Learning Strategies Inventory (SMALSI™; Stroud & Reynolds, 2006), published by Western Psychological Services, a self-report inventory designed to assess academic motivation, as well as learning and study strategies. The test identifies 10 primary constructs, referred to broadly as…
Descriptors: Motivation, Measures (Individuals), Test Anxiety, Test Wiseness
Holland, Kristopher J.; Sheth, Nandita Baxi – Studies in Art Education: A Journal of Issues and Research in Art Education, 2018
In this article, we investigate the Visual Arts Educative Teacher Performance Assessment (edTPA) while integrating aspects of philosopher Jean-François Lyotard's thought in order to bring to light implicit assumptions made by the test. We expose disconnections between a mission for art education to create teachers as inquirers and the type of…
Descriptors: Visual Arts, Art Education, Performance Based Assessment, Preservice Teachers
Toker, Deniz – TESL-EJ, 2019
The central purpose of this paper is to examine validity problems arising from the multiple-choice items and technical passages in the Test of English as a Foreign Language Internet-based Test (TOEFL iBT) reading section, primarily concentrating on construct-irrelevant variance (Messick, 1989). My personal TOEFL iBT experience, along with my…
Descriptors: English (Second Language), Language Tests, Second Language Learning, Computer Assisted Testing
Francois, Chantal; Hood, Mia – Journal for Multicultural Education, 2021
Purpose: Scholars who advocate for equity-oriented educational practices have argued that the accountability era in the USA, now in place for two decades, has failed in its intended goal to improve student performance for traditionally marginalized student populations. This study aims to use a sociocultural lens to trace how a century-old…
Descriptors: Reading Tests, Performance Based Assessment, Standardized Tests, Accountability
Wise, Steven L.; Kingsbury, G. Gage – Journal of Educational Measurement, 2016
This study examined the utility of response time-based analyses in understanding the behavior of unmotivated test takers. For the data from an adaptive achievement test, patterns of observed rapid-guessing behavior and item response accuracy were compared to the behavior expected under several types of models that have been proposed to represent…
Descriptors: Achievement Tests, Student Motivation, Test Wiseness, Adaptive Testing
Higgins, Derrick; Heilman, Michael – Educational Measurement: Issues and Practice, 2014
As methods for automated scoring of constructed-response items become more widely adopted in state assessments, and are used in more consequential operational configurations, it is critical that their susceptibility to gaming behavior be investigated and managed. This article provides a review of research relevant to how construct-irrelevant…
Descriptors: Automation, Scoring, Responses, Test Wiseness
Rios, Joseph A.; Liu, Ou Lydia; Bridgeman, Brent – New Directions for Institutional Research, 2014
This chapter describes a study that compares two approaches (self-reported effort [SRE] and response time effort [RTE]) for identifying low-effort examinees in student learning outcomes assessment. Although both approaches equally discriminated from measures of ability (e.g., SAT scores), RTE was found to have a stronger relationship with test…
Descriptors: Student Evaluation, Educational Assessment, Reaction Time, Measures (Individuals)