NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 183 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Aditya Shah; Ajay Devmane; Mehul Ranka; Prathamesh Churi – Education and Information Technologies, 2024
Online learning has grown due to the advancement of technology and flexibility. Online examinations measure students' knowledge and skills. Traditional question papers include inconsistent difficulty levels, arbitrary question allocations, and poor grading. The suggested model calibrates question paper difficulty based on student performance to…
Descriptors: Computer Assisted Testing, Difficulty Level, Grading, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Beseiso, Majdi; Alzubi, Omar A.; Rashaideh, Hasan – Journal of Computing in Higher Education, 2021
E-learning is gradually gaining prominence in higher education, with universities enlarging provision and more students getting enrolled. The effectiveness of automated essay scoring (AES) is thus holding a strong appeal to universities for managing an increasing learning interest and reducing costs associated with human raters. The growth in…
Descriptors: Automation, Scoring, Essays, Writing Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Doewes, Afrizal; Kurdhi, Nughthoh Arfawi; Saxena, Akrati – International Educational Data Mining Society, 2023
Automated Essay Scoring (AES) tools aim to improve the efficiency and consistency of essay scoring by using machine learning algorithms. In the existing research work on this topic, most researchers agree that human-automated score agreement remains the benchmark for assessing the accuracy of machine-generated scores. To measure the performance of…
Descriptors: Essays, Writing Evaluation, Evaluators, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Johnson, Matthew S.; Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2020
One common score reported from diagnostic classification assessments is the vector of posterior means of the skill mastery indicators. As with any assessment, it is important to derive and report estimates of the reliability of the reported scores. After reviewing a reliability measure suggested by Templin and Bradshaw, this article suggests three…
Descriptors: Reliability, Probability, Skill Development, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Jin, Kuan-Yu; Wang, Wen-Chung – Journal of Educational Measurement, 2018
The Rasch facets model was developed to account for facet data, such as student essays graded by raters, but it accounts for only one kind of rater effect (severity). In practice, raters may exhibit various tendencies such as using middle or extreme scores in their ratings, which is referred to as the rater centrality/extremity response style. To…
Descriptors: Scoring, Models, Interrater Reliability, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Ravand, Hamdollah; Baghaei, Purya – International Journal of Testing, 2020
More than three decades after their introduction, diagnostic classification models (DCM) do not seem to have been implemented in educational systems for the purposes they were devised. Most DCM research is either methodological for model development and refinement or retrofitting to existing nondiagnostic tests and, in the latter case, basically…
Descriptors: Classification, Models, Diagnostic Tests, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Robert Schoen; Lanrong Li; Xiaotong Yang; Ahmet Guven; Claire Riddell – Society for Research on Educational Effectiveness, 2021
Many classroom-observation instruments have been developed (e.g., Gleason et al., 2017; Nava et al., 2019; Sawada et al., 2002), but a very small number of studies published in refereed journals have rigorously examined the quality of the ratings and the instrument using measurement models. For example, Gleason et al. developed a mathematics…
Descriptors: Item Response Theory, Models, Measurement, Mathematics Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Srour, F. Jordan; Karkoulian, Silva – International Journal of Social Research Methodology, 2022
The literature provides multiple measures of diversity along a single demographic dimension, but when it comes to studying the interaction of multiple diversity types (e.g. age, gender, and race), the field of useable measures diminishes. We present the use of decision trees as a machine learning technique to automatically identify the…
Descriptors: Diversity, Decision Making, Artificial Intelligence, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Mulligan, Shelley – Journal of Occupational Therapy, Schools & Early Intervention, 2017
Occupational performance assessments of children are essential for guiding occupational therapy intervention and for measuring the effectiveness of occupational therapy services for children. A review of relevant research and of occupational performance assessments designed for children was conducted to determine and describe how the occupational…
Descriptors: Occupational Therapy, Performance Tests, Children, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
O'Brien, Edward J.; Cook, Anne E. – Discourse Processes: A multidisciplinary journal, 2016
Common to all models of reading comprehension is the assumption that a reader's level of comprehension is heavily influenced by their standards of coherence (van den Broek, Risden, & Husbye-Hartman, 1995). Our discussion focuses on a subcomponent of the readers' standards of coherence: the coherence threshold. We situate this discussion within…
Descriptors: Reading Comprehension, Models, Rhetoric, Reading Ability
Peer reviewed Peer reviewed
Direct linkDirect link
Climie, Emma; Henley, Laura – British Journal of Special Education, 2016
School-based practitioners are often called upon to provide assessment and recommendations for struggling students. These assessments often open doors to specialised services or interventions and provide opportunities for students to build competencies in areas of need. However, these assessments often fail to highlight the abilities of these…
Descriptors: Student Evaluation, Alternative Assessment, Relevance (Education), Models
Peer reviewed Peer reviewed
Direct linkDirect link
Shu, Lianghua; Schwarz, Richard D. – Journal of Educational Measurement, 2014
As a global measure of precision, item response theory (IRT) estimated reliability is derived for four coefficients (Cronbach's a, Feldt-Raju, stratified a, and marginal reliability). Models with different underlying assumptions concerning test-part similarity are discussed. A detailed computational example is presented for the targeted…
Descriptors: Item Response Theory, Reliability, Models, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Katz, Daniel S. – Kappa Delta Pi Record, 2016
Including growth models based on student test scores in teacher evaluations effectively holds teachers individually accountable for students improving their test scores. While an attractive policy for state administrators and advocates of education reform, value-added measures have been fraught with problems, and their use in teacher evaluation is…
Descriptors: Teacher Evaluation, Models, Scores, Evaluation Criteria
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Dimitrov, Dimiter M.; von Eye, Alexander; Marcoulides, George A. – Educational and Psychological Measurement, 2013
A latent variable modeling method for evaluation of interrater agreement is outlined. The procedure is useful for point and interval estimation of the degree of agreement among a given set of judges evaluating a group of targets. In addition, the approach allows one to test for identity in underlying thresholds across raters as well as to identify…
Descriptors: Interrater Reliability, Models, Statistical Analysis, Computation
Mahalingam, Sheila; Abdollah, Faizal Mohd; Sahib, Shahrin – International Association for Development of the Information Society, 2014
The paper focus on learner centric attributes in a m-learning environment encounters the security measurements. In order to build up a systematic threat and countermeasure for protecting the learners as well as providing awareness and satisfaction in utilizing the mobile learning system, a security model need to be overhauled. The brief literature…
Descriptors: Electronic Learning, Computer Security, Reliability, Trust (Psychology)
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  12  |  13