Abstract
For didactic purposes, all of the preceding chapters have assumed that the metric of the ability scale was known. This metric had a midpoint of zero, a unit of measurement of 1, and a range from negative infinity to positive infinity. The numerical values of the item parameters and the examinee’s ability parameters have been expressed in this metric. While this has served to introduce you to the fundamental concepts of item response theory , it does not represent the actual testing situation. When test constructors write an item, they know what trait they want the item to measure and whether the item is designed to function among low-, medium-, or high-ability examinees. But it is not possible to determine the values of the item’s parameters a priori. In addition, when a test is administered to a group of examinees, it is not known in advance how much of the latent trait each of the examinees possesses. As a result, a major task is to determine the values of the item parameters and examinee abilities in a metric for the underlying latent trait. In item response theory, this task is called test calibration and it provides a frame of reference for interpreting test results. Test calibration is accomplished by administering a test to a group of N examinees and dichotomously scoring the examinees’ responses to the J items. Then mathematical procedures are applied to the item response data in order to create an ability scale that is unique to the particular combination of test items and examinees. The values of the item parameter estimates and the examinees’ estimated abilities are expressed in this metric. Once this is accomplished, the test has been calibrated and the test results can be interpreted via the constructs of item response theory.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Baker, F. B. (1992). Item response theory: Parameter estimation techniques. New York, NY: Dekker.
Baker, F. B., & Kim, S.-H. (2004). Item response theory: Parameter estimation techniques (2nd ed.). New York, NY: Dekker.
Linacre, J. M. (2015). A user’s guide to WINSTEPS MINISTEP Rasch-model computer programs. Chicago, IL: Winsteps.com.
Wingersky, M. S., Barton, M. A., & Lord, F. M. (1982). LOGIST user’s guide. LOGIST 5, version 1.0. Princeton, NJ: Educational Testing Service.
Wingersky, M. S., Patrick, R., & Lord, F. M. (1999). LOGIST user’s guide (version 7.1). Princeton, NJ: Educational Testing Service.
Wright, B. D., & Mead, R. J. (1976). BICAL: Calibrating items with the Rasch model. (Research Memorandum No. 23). Chicago, IL: Statistical Laboratory, Department of Education, University of Chicago.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this chapter
Cite this chapter
Baker, F.B., Kim, SH. (2017). Test Calibration. In: The Basics of Item Response Theory Using R. Statistics for Social and Behavioral Sciences. Springer, Cham. https://doi.org/10.1007/978-3-319-54205-8_7
Download citation
DOI: https://doi.org/10.1007/978-3-319-54205-8_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-54204-1
Online ISBN: 978-3-319-54205-8
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)