NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
What Works Clearinghouse Rating
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Xu, Jinfen; Li, Changying – Studies in Second Language Learning and Teaching, 2022
This study investigates how different form-focused instruction (FFI) timing impacts English as a foreign language (EFL) learners' grammar development. A total of 169 Chinese middle school learners were assigned to four conditions randomly: control, before-isolated FFI, integrated FFI, and after-isolated FFI. The three experimental groups received…
Descriptors: Intervention, Grammar, English (Second Language), Second Language Learning
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sasayama, Shoko; Garcia Gomez, Pablo; Norris, John M. – ETS Research Report Series, 2021
This report describes the development of efficient second language (L2) writing assessment tasks designed specifically for low-proficiency learners of English to be included in the "TOEFL® Essentials"™ test. Based on the can-do descriptors of the Common European Framework of Reference for Languages for the A1 through B1 levels of…
Descriptors: English (Second Language), Language Tests, Second Language Learning, Writing Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Terao, Takahiro; Ishii, Hidetoki – SAGE Open, 2020
This study aimed to compare selection patterns of distractors (incorrect options) according to test taker proficiency regarding Japanese students' summarization skills of an English paragraph. Participants included 414 undergraduate students, and the test comprised three summarization process types--deletion, generalization, and integration.…
Descriptors: Comparative Analysis, English (Second Language), Second Language Instruction, Second Language Learning
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Jahangard, Ali – MEXTESOL Journal, 2022
One of the most interesting studies on the role of L1 and contrastive analysis in vocabulary teaching is by Laufer and Girsai (2008). However, due to some methodological issues, their research findings are open to criticism and controversy. The current study aimed to replicate the research with a more rigorous design to re-investigate the…
Descriptors: Grammar, Vocabulary Development, Second Language Learning, Second Language Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Sato, Takanori; Ikeda, Naoki – Language Testing in Asia, 2015
Background: High-stakes tests have an immense washback effect on what students learn and affect the content of student learning. However, if students fail to recognize the abilities that the test developers intend to measure, they are less likely to learn what the test developers wish them to learn. This study aims to investigate test-taker…
Descriptors: High Stakes Tests, Testing Problems, Test Items, College Students
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr.; Kobrin, Jennifer L.; Wind, Stefanie A. – International Journal of Testing, 2014
The purpose of this study is to explore patterns in model-data fit related to subgroups of test takers from a large-scale writing assessment. Using data from the SAT, a calibration group was randomly selected to represent test takers who reported that English was their best language from the total population of test takers (N = 322,011). A…
Descriptors: College Entrance Examinations, Writing Tests, Goodness of Fit, English
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zhang, Mo; Breyer, F. Jay; Lorenz, Florian – ETS Research Report Series, 2013
In this research, we investigated the suitability of implementing "e-rater"® automated essay scoring in a high-stakes large-scale English language testing program. We examined the effectiveness of generic scoring and 2 variants of prompt-based scoring approaches. Effectiveness was evaluated on a number of dimensions, including agreement…
Descriptors: Computer Assisted Testing, Computer Software, Scoring, Language Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Bunch, Michael B. – Language Testing, 2011
Title III of Public Law 107-110 (No Child Left Behind; NCLB) provided for creation of assessments of English language learners (ELLs) and established, through the Enhanced Assessment Grant program, a platform from which four consortia of states developed ELL tests aligned to rigorous statewide content standards. Those four tests (ACCESS for ELLs,…
Descriptors: Test Items, Student Evaluation, Federal Legislation, Formative Evaluation
Lee, Yong-Won; Kantor, Robert; Mollaun, Pam – 2002
This paper reports the results of generalizability theory (G) analyses done for new writing and speaking tasks for the Test of English as a Foreign Language (TOEFL). For writing, a special focus was placed on evaluating the impact on the reliability of the number of raters (or ratings) per essay (one or two) and the number of tasks (one, two, or…
Descriptors: English (Second Language), Generalizability Theory, Reliability, Scores
Hendrickson, Amy; Patterson, Brian; Melican, Gerald – College Board, 2008
Presented at the Annual National Council on Measurement in Education (NCME) in New York in March 2008. This presentation explores how different item weighting can affect the effective weights, validity coefficents and test reliability of composite scores among test takers.
Descriptors: Multiple Choice Tests, Test Format, Test Validity, Test Reliability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Haberman, Shelby J. – ETS Research Report Series, 2004
Statistical and measurement properties are examined for features used in essay assessment to determine the generalizability of the features across populations, prompts, and individuals. Data are employed from TOEFL® and GMAT® examinations and from writing for Criterion?.
Descriptors: Language Tests, English (Second Language), Second Language Learning, Business Administration Education
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Broer, Markus; Lee, Yong-Won; Rizavi, Saba; Powers, Don – ETS Research Report Series, 2005
Three polytomous DIF detection techniques--the Mantel test, logistic regression, and polySTAND--were used to identify GRE® Analytical Writing prompts ("Issue" and "Argument") that are differentially difficult for (a) female test takers; (b) African American, Asian, and Hispanic test takers; and (c) test takers whose strongest…
Descriptors: Culture Fair Tests, Item Response Theory, Test Items, Cues