Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 5 |
Descriptor
Grade 5 | 5 |
Natural Language Processing | 5 |
Scoring | 5 |
Automation | 4 |
Essays | 4 |
Elementary School Students | 3 |
Grade 4 | 3 |
Grade 6 | 3 |
Writing (Composition) | 3 |
Evaluation Methods | 2 |
Feedback (Response) | 2 |
More ▼ |
Source
American Educational Research… | 1 |
Grantee Submission | 1 |
International Journal of… | 1 |
Language Assessment Quarterly | 1 |
National Center for Research… | 1 |
Author
Chen, Dandan | 1 |
Correnti, R. | 1 |
Correnti, Richard | 1 |
E. E. Jang | 1 |
Hebert, Michael | 1 |
Howe, E. | 1 |
Iseli, Markus R. | 1 |
Kerr, Deirdre | 1 |
L. Hannah | 1 |
Litman, D. | 1 |
Litman, Diane | 1 |
More ▼ |
Publication Type
Reports - Research | 5 |
Journal Articles | 3 |
Speeches/Meeting Papers | 1 |
Education Level
Grade 5 | 5 |
Elementary Education | 4 |
Intermediate Grades | 4 |
Middle Schools | 4 |
Grade 4 | 3 |
Early Childhood Education | 2 |
Grade 3 | 2 |
Grade 6 | 2 |
Primary Education | 2 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Chen, Dandan; Hebert, Michael; Wilson, Joshua – American Educational Research Journal, 2022
We used multivariate generalizability theory to examine the reliability of hand-scoring and automated essay scoring (AES) and to identify how these scoring methods could be used in conjunction to optimize writing assessment. Students (n = 113) included subsamples of struggling writers and non-struggling writers in Grades 3-5 drawn from a larger…
Descriptors: Reliability, Scoring, Essays, Automation
L. Hannah; E. E. Jang; M. Shah; V. Gupta – Language Assessment Quarterly, 2023
Machines have a long-demonstrated ability to find statistical relationships between qualities of texts and surface-level linguistic indicators of writing. More recently, unlocked by artificial intelligence, the potential of using machines to identify content-related writing trait criteria has been uncovered. This development is significant,…
Descriptors: Validity, Automation, Scoring, Writing Assignments
Zhang, H.; Magooda, A.; Litman, D.; Correnti, R.; Wang, E.; Matsumura, L. C.; Howe, E.; Quintana, R. – Grantee Submission, 2019
Writing a good essay typically involves students revising an initial paper draft after receiving feedback. We present eRevise, a web-based writing and revising environment that uses natural language processing features generated for rubric-based essay scoring to trigger formative feedback messages regarding students' use of evidence in…
Descriptors: Formative Evaluation, Essays, Writing (Composition), Revision (Written Composition)
Rahimi, Zahra; Litman, Diane; Correnti, Richard; Wang, Elaine; Matsumura, Lindsay Clare – International Journal of Artificial Intelligence in Education, 2017
This paper presents an investigation of score prediction based on natural language processing for two targeted constructs within analytic text-based writing: 1) students' effective use of evidence and, 2) their organization of ideas and evidence in support of their claim. With the long-term goal of producing feedback for students and teachers, we…
Descriptors: Scoring, Automation, Scoring Rubrics, Natural Language Processing
Kerr, Deirdre; Mousavi, Hamid; Iseli, Markus R. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2013
The Common Core assessments emphasize short essay constructed response items over multiple choice items because they are more precise measures of understanding. However, such items are too costly and time consuming to be used in national assessments unless a way is found to score them automatically. Current automatic essay scoring techniques are…
Descriptors: Automation, Scoring, Essay Tests, Natural Language Processing