A Comparison of IRT Model Combinations for Assessing Fit in a Mixed Format Elementary School Science Test

@article{Yilmaz2019ACO,
  title={A Comparison of IRT Model Combinations for Assessing Fit in a Mixed Format Elementary School Science Test},
  author={H. Bayram Yilmaz},
  journal={lnternational Electronic Journal of Elementary Education},
  year={2019}
}
  • H. B. Yilmaz
  • Published 2019
  • Psychology
  • lnternational Electronic Journal of Elementary Education
Open ended and multiple choice questions are commonly placed on the same tests; however, there is a discussion on the effects of using different item types on the test and item statistics. This study aims to compare model and item fit statistics in a mixed format test where multiple choice and constructed response items are used together. In this 25-item fourth grade science test administered to 2351 students in 35 schools in Turkey, items are calibrated separately and concurrently utilizing… Expand

Figures and Tables from this paper

Comparison of the Accuracy of Item Response Theory Models in Estimating Student’s Ability
This study aims to determine the item response theory model which is more accurate in estimating students' mathematical abilities. The models compared in this study are Multiple Choice Model andExpand
The information function of mixed-format test of physics learning outcomes measurement
This study investigates the function of item information on mixed-format tests on the measurement of physics learning outcomes. The item information function is an important attribute in the itemExpand

References

SHOWING 1-10 OF 43 REFERENCES
A Comparison of IRT and Rasch Procedures in a Mixed-Item Format Test
This study investigated the effects of test length (10, 20 and 30 items), scoring schema (proportion of dichotomous ad polytomous scoring) and item analysis model (IRT and Rasch) on the abilityExpand
The Contribution of Constructed Response Items to Large Scale Assessment: Measuring and Understanding Their Impact.
This article investigates several questions regarding the impact of different item formats on measurement characteristics. Constructed response (CR) items and multiple choice (MC) items obviouslyExpand
EQUATING OF MIXED-FORMAT TESTS IN LARGE-SCALE ASSESSMENTS
This study examined variations of the nonequivalent-groups equating design for mixed-format tests—tests containing both multiple-choice (MC) and constructed-response (CR) items—to determine whichExpand
Further Investigation of the Performance of S - X2: An Item Fit Index for Use With Dichotomous Item Response Theory Models
TLDR
Results from this study imply that S - X 2 may be a useful tool in detecting the misfit of one item contained in an otherwise well-fitted test, lending additional support to the utility of the index for use with dichotomous item response theory models. Expand
An Investigation of the Performance of the Generalized S-X2 Item-Fit Index for Polytomous IRT Models
Orlando and Thissen (2000, 2003) proposed an item-fit index, S-X, for dichotomous item response theory (IRT) models, which has performed better than traditional item-fit statistics such as Yen’sExpand
CONCURRENT CALIBRATION OF DICHOTOMOUSLY AND POLYTOMOUSLY SCORED TOEFL ITEMS USING IRT MODELS
In order to meet the needs of the Test of English as a Foreign Language (TOEFL®) constituencies, the TOEFL program is sponsoring a development project known as TOEFL 2000. Drawing from currentExpand
Performance of the generalized S-X2 item fit index for the graded response model
The utility of Orlando and Thissen’s (2000, 2003) S-X2 fit index was extended to the model-fit analysis of the graded response model (GRM). The performance of a modified S-X2 in assessing item-fit ofExpand
Estimation of latent ability using a response pattern of graded scores
Estimation of latent ability using the entire response pattern of free-response items is discussed, first in the general case and then in the case where the items are scored in a graded way,Expand
A PRIMER OF ITEM RESPONSE THEORY
Abstract : This book is an introduction to Item Response Theory (IRT) (also called Item Characteristic Curve Theory, or latent trait theory). It is written for the testing practitioner with minimumExpand
Comparison Between Dichotomous and Polytomous Scoring of Innovative Items in a Large-Scale Computerized Adaptive Test
This study explored the impact of partial credit scoring of one type of innovative items (multiple-response items) in a computerized adaptive version of a large-scale licensure pretest andExpand
...
1
2
3
4
5
...