Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 9 |
Since 2006 (last 20 years) | 19 |
Descriptor
Foreign Countries | 22 |
Test Items | 22 |
Item Response Theory | 18 |
Difficulty Level | 7 |
Test Bias | 6 |
Computer Assisted Testing | 5 |
Evaluation Methods | 5 |
Item Analysis | 5 |
Test Reliability | 5 |
Comparative Analysis | 4 |
Correlation | 4 |
More ▼ |
Source
Author
Ercikan, Kadriye | 2 |
Hung, Su-Pin | 2 |
Wang, Wen-Chung | 2 |
Blömeke, Sigrid | 1 |
Braeken, Johan | 1 |
Chang, Cheng-Chieh | 1 |
Chang, Wen-Chih | 1 |
Chao, Louis R. | 1 |
Chen, Chieh-Yu | 1 |
Chen, Deng-Jyi | 1 |
Chen, Hsueh-Chih | 1 |
More ▼ |
Publication Type
Journal Articles | 19 |
Reports - Research | 15 |
Reports - Evaluative | 5 |
Collected Works - Proceedings | 1 |
Dissertations/Theses -… | 1 |
Tests/Questionnaires | 1 |
Education Level
Audience
Location
Taiwan | 22 |
United States | 4 |
Hong Kong | 3 |
Australia | 2 |
Canada | 2 |
Japan | 2 |
Kuwait | 2 |
Qatar | 2 |
Singapore | 2 |
South Korea | 2 |
Botswana | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Progress in International… | 3 |
Remote Associates Test | 2 |
Trends in International… | 2 |
Graduate Record Examinations | 1 |
Program for International… | 1 |
Wechsler Adult Intelligence… | 1 |
What Works Clearinghouse Rating
Hung, Su-Pin; Wu, Ching-Lin – Creativity Research Journal, 2021
The Remote Associates Test, generally used in creativity research, has Chinese versions for the three levels of "radical-word-vocabulary." However, research has not been conducted on the influence of the item components on the difficulties among these Chinese Remote Associates Tests (CRATs). The present study selected six item components…
Descriptors: Creativity Tests, Chinese, Test Items, Difficulty Level
Chen, Yi-Hsin – Journal of Psychoeducational Assessment, 2022
The quality of diagnostic profiles and probability assignment depends on the validity of the proposed attributes and Q-matrix. The rule-space method (RSM), one of diagnostic classification models, provides the quality indices of diagnostic profiles, such as the classification rate and the squared Mahalanobis distance. The study aims to further…
Descriptors: Profiles, Probability, Classification, Construct Validity
Kao, Yu-Ting; Kuo, Hung-Chih – Interactive Learning Environments, 2023
This study implemented the principles of dynamic assessment (DA) with computer technology, iSpring Quiz Maker, to (1) identify the English listening difficulties of 172 L2 English learners; (2) diagnose their individual learning needs, and (3) promote their future potential abilities. Upon evaluating the participating junior high school students'…
Descriptors: Listening Comprehension Tests, English (Second Language), Second Language Learning, Second Language Instruction
A Feasible Guidance for Ordered Multiple-Choice Items in Students' Hierarchical Understanding Levels
Su, King-Dow – Journal of Baltic Science Education, 2019
This research focuses on students' 5 hierarchical levels of Ordered Multiple-Choice (OMC) items for their extensive conceptualized understanding in the particulate nature of matter (PNM) chemistry. The basic framework for OMC items is to link students' conceptual understanding levels with possible cognitive responses. Developed as the substantial…
Descriptors: Multiple Choice Tests, Science Tests, STEM Education, Test Items
Hung, Su-Pin; Huang, Po-Sheng; Chen, Hsueh-Chih – Creativity Research Journal, 2016
The remote association test (RAT) has been applied in various fields; however, evidence of construct validity for the original version and subsequent extensions of the RAT remains limited. This study aimed to elucidate the dimensionality and the relationship between item features and item difficulties for the RAT--Chinese Version (RAT-C) using the…
Descriptors: Creativity Tests, Difficulty Level, Test Items, Item Response Theory
Nix, John-Michael L. – International Journal of Listening, 2021
The relationship between second language (L2) listening presage and process variables was examined in a large cross-sectional study. A formal model of learning beliefs was adapted to create a structural equation model testing the interrelationships of individual difference (gender, practice, experience with native teachers) and trait variables…
Descriptors: Correlation, Listening Comprehension, Case Studies, Individual Differences
Tsai, Liang-Ting; Chang, Cheng-Chieh – Environmental Education Research, 2019
This study established a Chinese scale for measuring high school students' ocean literacy. This included testing its reliability, validity, and differential item functioning (DIF) with the aim of compensating for the lack of DIF tests focusing on current scales. The construct validity and reliability were verified and tested by analyzing the…
Descriptors: Foreign Countries, Measures (Individuals), Oceanography, Knowledge Level
Oliveri, María Elena; Ercikan, Kadriye; Zumbo, Bruno D.; Lawless, René – International Journal of Testing, 2014
In this study, we contrast results from two differential item functioning (DIF) approaches (manifest and latent class) by the number of items and sources of items identified as DIF using data from an international reading assessment. The latter approach yielded three latent classes, presenting evidence of heterogeneity in examinee response…
Descriptors: Test Bias, Comparative Analysis, Reading Tests, Effect Size
Chen, Chieh-Yu – ProQuest LLC, 2017
Investigating the psychometric properties of a screening instrument for young children is necessary to ascertain its quality and accuracy. In light of the important role culture plays on human beliefs and parenting styles, a newly translated and adapted test needs to be studied. Evaluating outcomes on a translated version of a test may reveal…
Descriptors: Psychometrics, Screening Tests, Questionnaires, Social Development
Oliveri, Maria Elena; Ercikan, Kadriye; Zumbo, Bruno – International Journal of Testing, 2013
In this study, we investigated differential item functioning (DIF) and its sources using a latent class (LC) modeling approach. Potential sources of LC DIF related to instruction and teacher-related variables were investigated using substantive and three statistical approaches: descriptive discriminant function, multinomial logistic regression,…
Descriptors: Test Bias, Test Items, Multivariate Analysis, Discriminant Analysis
Sung, Pei-Ju; Lin, Su-Wei; Hung, Pi-Hsia – Universal Journal of Educational Research, 2015
Task difficulty is a critical issue affecting test developers. Controlling or balancing the item difficulty of an assessment improves its validity and discrimination. Test developers construct tests from the cognitive perspective, by making the test constructing process more scientific and efficient; thus, the scores obtained more precisely…
Descriptors: Foreign Countries, Listening, Listening Comprehension, Listening Comprehension Tests
Reckase, Mark D.; Xu, Jing-Ru – Educational and Psychological Measurement, 2015
How to compute and report subscores for a test that was originally designed for reporting scores on a unidimensional scale has been a topic of interest in recent years. In the research reported here, we describe an application of multidimensional item response theory to identify a subscore structure in a test designed for reporting results using a…
Descriptors: English, Language Skills, English Language Learners, Scores
Huang, Hung-Yu; Wang, Wen-Chung – Educational and Psychological Measurement, 2013
Both testlet design and hierarchical latent traits are fairly common in educational and psychological measurements. This study aimed to develop a new class of higher order testlet response models that consider both local item dependence within testlets and a hierarchy of latent traits. Due to high dimensionality, the authors adopted the Bayesian…
Descriptors: Item Response Theory, Models, Bayesian Statistics, Computation
Yen, Yung-Chin; Ho, Rong-Guey; Laio, Wen-Wei; Chen, Li-Ju; Kuo, Ching-Chin – Applied Psychological Measurement, 2012
In a selected response test, aberrant responses such as careless errors and lucky guesses might cause error in ability estimation because these responses do not actually reflect the knowledge that examinees possess. In a computerized adaptive test (CAT), these aberrant responses could further cause serious estimation error due to dynamic item…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Response Style (Tests)
Braeken, Johan; Blömeke, Sigrid – Assessment & Evaluation in Higher Education, 2016
Using data from the international Teacher Education and Development Study: Learning to Teach Mathematics (TEDS-M), the measurement equivalence of teachers' beliefs across countries is investigated for the case of "mathematics-as-a fixed-ability". Measurement equivalence is a crucial topic in all international large-scale assessments and…
Descriptors: Comparative Analysis, Bayesian Statistics, Test Bias, Teacher Education
Previous Page | Next Page »
Pages: 1 | 2