Publication Date
In 2025 | 1 |
Since 2024 | 18 |
Since 2021 (last 5 years) | 47 |
Since 2016 (last 10 years) | 126 |
Since 2006 (last 20 years) | 204 |
Descriptor
Test Content | 308 |
Test Items | 115 |
Foreign Countries | 96 |
Test Construction | 78 |
Test Validity | 65 |
Scores | 47 |
Language Tests | 45 |
Second Language Learning | 42 |
Student Evaluation | 42 |
Test Format | 40 |
Comparative Analysis | 38 |
More ▼ |
Source
Author
Sireci, Stephen G. | 4 |
Solano-Flores, Guillermo | 3 |
Steffen, Manfred | 3 |
Abedi, Jamal | 2 |
Agarwal, Pooja K. | 2 |
Bauer, Scott C. | 2 |
Binkley, Marilyn | 2 |
Borman, Walter C. | 2 |
Chang, Hua-Hua | 2 |
Cox, Shawna | 2 |
Dorans, Neil J. | 2 |
More ▼ |
Publication Type
Education Level
Audience
Teachers | 7 |
Practitioners | 5 |
Researchers | 2 |
Administrators | 1 |
Location
Australia | 8 |
Canada | 8 |
Turkey | 8 |
California | 7 |
Europe | 6 |
China | 5 |
United States | 5 |
Germany | 4 |
Hong Kong | 4 |
Iran | 4 |
Japan | 4 |
More ▼ |
Laws, Policies, & Programs
Every Student Succeeds Act… | 2 |
No Child Left Behind Act 2001 | 2 |
Individuals with Disabilities… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Kalayci, Nurdan; Cimen, Orhan – Educational Sciences: Theory and Practice, 2012
The aim of this study is to examine the questionnaires used to evaluate teaching performance in higher education institutes and called "Instructor and Course Evaluation Questionnaires (ICEQ)" in terms of questionnaire preparation techniques and components of curriculum. Obtaining at least one ICEQ belonging to any state and private…
Descriptors: Higher Education, Teaching Methods, Questionnaires, Learning Processes
Cigdem, Harun; Oncu, Semiral – EURASIA Journal of Mathematics, Science & Technology Education, 2015
This survey study examines an assessment methodology through e-quizzes administered at a military vocational college and subsequent student perceptions in spring 2013 at the "Computer Networks" course. A total of 30 Computer Technologies and 261 Electronic and Communication Technologies students took three e-quizzes. Data were gathered…
Descriptors: Foreign Countries, Military Schools, Military Training, Vocational Education
Ewing, Maureen; Packman, Sheryl; Hamen, Cynthia; Thurber, Allison Clark – Applied Measurement in Education, 2010
In the last few years, the Advanced Placement (AP) Program[R] has used evidence-centered assessment design (ECD) to articulate the knowledge, skills, and abilities to be taught in the course and measured on the summative exam for four science courses, three history courses, and six world language courses; its application to calculus and English…
Descriptors: Advanced Placement Programs, Equivalency Tests, Evidence, Test Construction
Choi, Bo Young; Park, Heerak; Nam, Suk Kyung; Lee, Jayoung; Cho, Daeyeon; Lee, Sang Min – Career Development Quarterly, 2011
The purpose of this study was to develop a Korean College Stress Inventory (KCSI), which is designed to measure Korean college students' experiences and symptoms of career stress. Even though there have been numerous scales related to career issues, few scales measure the career stress construct and its dimensions. Factor structure, internal…
Descriptors: College Students, Factor Structure, Psychometrics, Stress Variables
Sedki, S. Sam – Journal of International Education Research, 2011
Most professors use examinations as an important assessment tool to aid in determining the level of student subject matter comprehension. We also use the feedback from examinations as an indicator of the appropriateness and effectiveness of the teaching methodologies we are utilizing in the classroom. This paper is a follow-up to a 2006-2007 study…
Descriptors: Tests, Comparative Analysis, Teaching Methods, Comparative Education
Papageorgiou, Spiros; Stevens, Robin; Goodwin, Sarah – Language Assessment Quarterly, 2012
Listening comprehension tests typically include both monologic and dialogic input to measure listening ability. However, research as to which type of input is more challenging for examinees remains limited and has provided inconclusive results (Brindley & Slatyer, 2002; Read, 2002; Shohamy & Inbar, 1991). A better understanding of the…
Descriptors: Listening Comprehension Tests, Test Items, Content Analysis, Listening Comprehension
Sadler, D. Royce – Assessment & Evaluation in Higher Education, 2010
If a grade is to be trusted as an authentic representation of a student's level of academic achievement, one of the requirements is that all the elements that contribute to that grade must qualify as achievement, and not be something else. The implications of taking this proposition literally turn out to be far reaching. Many elements that are…
Descriptors: Student Evaluation, Academic Achievement, Integrity, Credits
McDaniel, Mark A.; Agarwal, Pooja K.; Huelser, Barbie J.; McDermott, Kathleen B.; Roediger, Henry L., III – Journal of Educational Psychology, 2011
Typically, teachers use tests to evaluate students' knowledge acquisition. In a novel experimental study, we examined whether low-stakes testing ("quizzing") can be used to foster students' learning of course content in 8th grade science classes. Students received multiple-choice quizzes (with feedback); in the quizzes, some target…
Descriptors: Feedback (Response), Course Content, Grade 8, Incidence
Thurlow, Martha; Rogers, Christopher; Christensen, Laurene – National Center on Educational Outcomes, University of Minnesota, 2010
The success of all students, including students with disabilities, on statewide assessments in mathematics and reading/English language arts has been examined closely. This is due, in part, to the role of these content areas in school accountability for the Elementary and Secondary Education Act (ESEA) known as "No Child Left Behind" (NCLB).…
Descriptors: Science Tests, Disabilities, Student Participation, Testing Accommodations
Hendrickson, Amy; Patterson, Brian; Ewing, Maureen – College Board, 2010
The psychometric considerations and challenges associated with including constructed response items on tests are discussed along with how these issues affect the form assembly specifications for mixed-format exams. Reliability and validity, security and fairness, pretesting, content and skills coverage, test length and timing, weights, statistical…
Descriptors: Multiple Choice Tests, Test Format, Test Construction, Test Validity
Ketelhut, Diane Jass; Nelson, Brian; Schifter, Catherine; Kim, Younsu – Education Sciences, 2013
Current science assessments typically present a series of isolated fact-based questions, poorly representing the complexity of how real-world science is constructed. The National Research Council asserts that this needs to change to reflect a more authentic model of science practice. We strongly concur and suggest that good science assessments…
Descriptors: Virtual Classrooms, Science Tests, Academic Standards, Middle School Students
Tourkin, Steven; Thomas, Teresa; Swaim, Nancy; Cox, Shawna; Parmer, Randall; Jackson, Betty; Cole, Cornette; Zhang, Bei – National Center for Education Statistics, 2010
The Schools and Staffing Survey (SASS) is conducted by the National Center for Education Statistics (NCES) on behalf of the United States Department of Education in order to collect extensive data on American public and private elementary and secondary schools. SASS provides data on the characteristics and qualifications of teachers and…
Descriptors: Elementary Secondary Education, National Surveys, Public Schools, Private Schools
Sawaki, Yasuyo; Kim, Hae-Jin; Gentile, Claudia – Language Assessment Quarterly, 2009
In cognitive diagnosis a Q-matrix (Tatsuoka, 1983, 1990), which is an incidence matrix that defines the relationships between test items and constructs of interest, has great impact on the nature of performance feedback that can be provided to score users. The purpose of the present study was to identify meaningful skill coding categories that…
Descriptors: Feedback (Response), Test Items, Test Content, Identification
Ferne, Tracy; Rupp, Andre A. – Language Assessment Quarterly, 2007
This article reviews research on differential item functioning (DIF) in language testing conducted primarily between 1990 and 2005 with an eye toward providing methodological guidelines for developing, conducting, and disseminating research in this area. The article contains a synthesis of 27 studies with respect to five essential sets of…
Descriptors: Test Bias, Evaluation Research, Testing, Language Tests
Hager, Karen D.; Slocum, Timothy A. – Education and Training in Developmental Disabilities, 2008
Alternate assessments are the means through which students with significant cognitive disabilities participate in accountability testing, thus measurement validity of alternate assessments is a critical aspect of state educational accountability systems. When evaluating the validity of assessment systems, it is important to take a broad view of…
Descriptors: Test Content, Student Evaluation, Alternative Assessment, Test Validity