Abstract
In Chapter 6 we looked at elements of context validity that need to be considered at the test design stage and made a number of points in relation to test development which could potentially impact on the reliability of our tests. Hughes (2003: Chapter 5) examines a number of these specifically in relation to reliability. He provides a set of guidelines for making the test task itself more likely to produce reliable scores:
-
take enough samples of behaviour;
-
do not allow candidates too much freedom of choice;
-
write unambiguous items;
-
provide clear and explicit instructions;
-
ensure that tests are well laid out and perfectly legible;
-
make candidates familiar with format and testing techniques;
-
provide uniform and non-distracting conditions of administration;
-
use items that permit scoring which is as objective as possible;
-
make comparisons between candidates as direct as possible.
And in relation to the scoring of test performance itself:
-
provide a detailed scoring key;
-
train scorers;
-
agree acceptable responses and appropriate scores at the outset of scoring;
-
exclude items which do not discriminate well between weaker and stronger student;
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Further reading
Brown (1991) provides an accessible introduction to statistics for testers.
Bryman and Cramer (2001) provide a good introduction to SPSS.
Council of Europe (2001) provides a number of scales that might be useful as a basis for customizing to your own needs.
Crocker and Algina (1986) provide a comprehensive explanation of the statistics and concepts discussed in the chapter.
Hughes (2003: Chapter 5) provides a refreshingly accessible overview of all aspects of reliability, including worked examples of using IRT in Appendix 1.
Weigle (2002) provides numerous examples of scales for writing and Luoma (2004) does the same for speaking.
Author information
Authors and Affiliations
Copyright information
© 2005 Cyril J. Weir
About this chapter
Cite this chapter
Weir, C.J. (2005). Scoring Validity in Action. In: Language Testing and Validation. Research and Practice in Applied Linguistics. Palgrave Macmillan, London. https://doi.org/10.1057/9780230514577_10
Download citation
DOI: https://doi.org/10.1057/9780230514577_10
Publisher Name: Palgrave Macmillan, London
Print ISBN: 978-1-4039-1189-6
Online ISBN: 978-0-230-51457-7
eBook Packages: Palgrave Language & Linguistics CollectionEducation (R0)