Skip to main content
Top
Gepubliceerd in: Perspectives on Medical Education 1/2012

Open Access 01-03-2012 | Review Article

The use of progress testing

Auteurs: Lambert W. T. Schuwirth, Cees P. M. van der Vleuten

Gepubliceerd in: Perspectives on Medical Education | Uitgave 1/2012

share
DELEN

Deel dit onderdeel of sectie (kopieer de link)

  • Optie A:
    Klik op de rechtermuisknop op de link en selecteer de optie “linkadres kopiëren”
  • Optie B:
    Deel de link per e-mail
insite
ZOEKEN

Abstract

Progress testing is gaining ground rapidly after having been used almost exclusively in Maastricht and Kansas City. This increased popularity is understandable considering the intuitive appeal longitudinal testing has as a way to predict future competence and performance. Yet there are also important practicalities. Progress testing is longitudinal assessment in that it is based on subsequent equivalent, yet different, tests. The results of these are combined to determine the growth of functional medical knowledge for each student, enabling more reliable and valid decision making about promotion to a next study phase. The longitudinal integrated assessment approach has a demonstrable positive effect on student learning behaviour by discouraging binge learning. Furthermore, it leads to more reliable decisions as well as good predictive validity for future competence or retention of knowledge. Also, because of its integration and independence of local curricula, it can be used in a multi-centre collaborative production and administration framework, reducing costs, increasing efficiency and allowing for constant benchmarking. Practicalities include the relative unfamiliarity of faculty with the concept, the fact that remediation for students with a series of poor results is time consuming, the need to embed the instrument carefully into the existing assessment programme and the importance of equating subsequent tests to minimize test-to-test variability in difficulty. Where it has been implemented—collaboratively—progress testing has led to satisfaction, provided the practicalities are heeded well.

Introduction

Progress testing is becoming increasingly popular both in the Netherlands and internationally [19] after having been used for a long time only in those institutions where it was invented: the University of Missouri-Kansas City School of Medicine and Maastricht University in the Netherlands [10, 11]. The rapid spread of the concept, however, is not surprising because a longitudinal approach to assessment has an intrinsic appeal. It is intuitively more logical to assess students repeatedly and combine their results on these assessments to make predictions about future competence and/or performance. It is similar to a child’s development monitoring programme. In such programmes the child is weighed and measured at regular intervals and the outcomes are compared with population mean growth curves in order to detect and remedy problems as early as possible. This is probably also the reason why such an abundance of developmental and research papers on this topic have found their way to the literature in recent decades.
But it is not as straightforward as it looks; introducing progress testing involves not only a change in thinking about assessment but also an academic cultural change. Even more so, when collaboration on progress testing is sought; in such situations openness, non-competitiveness, exchange and mutual trust are essential. The purpose of this paper is to summarize the most important expectations and to accompany them with experiences from actual practice.

What is progress testing?

The many different descriptions of progress testing largely converge on the principle of longitudinal, repeated assessment of students’ functional knowledge. Often, a number of tests are set per academic year, each consisting of a large number of questions pitched at graduate level functional (relevant) knowledge. Each of these tests is sat by students of multiple or all year classes, and the results of each individual test are combined in a compensatory way to form the basis for a promotion decision at the end of the year. The test is comprehensive in that it consists of questions covering a broad domain of relevant medical knowledge, and it is organizationally founded on centralized test production, review, administration and analysis. Our description here is intentionally general because there are various different implementations possible, and more detailed descriptions are provided in the literature [1, 3, 5, 7, 11, 12].

Expectations and practicalities of progress testing

Reduction of examination stress

Because progress tests are longitudinal measurements it is assumed that students will experience less examination stress, because a one-off bad result cannot undo a series of good results [1113]. The—formative—collaborative progress test in the German speaking countries is even largely student led [5] and largely based on a bottom-up development. When McMaster formally evaluated their newly introduced progress test, a fair proportion (39%) of the students reported very little to no stress, a larger proportion (48%) reported limited stress and only a small proportion (27%) indicated moderate to high stress [3]. Yet, there is another side of the coin; if a single bad result cannot ruin a good series it is likewise difficult to make up for a bad series. This is particularly an issue when students are about to graduate, and all other examination requirements have been met, but they still have poor progress test results. A bad series of progress test results then has to be remediated, and one can safely assume that each of the subsequent sittings is a stressful event for those students, and in our experience in practice they are.

Repeat examinations become unnecessary

Another reported advantage of progress testing is that it renders resit examinations unnecessary. Resits are a burden for the organization; they have to be good quality examinations for only a small number of students. Also, they can lead students to adopt a minimalistic study approach; why study hard when there are always the resits [14]? But again, the side effect is that students in trouble have no quick repeat possibility, and may need to defer their graduation for some time, with very negative financial consequences.

Positive influence of student learning

Undisputed is the positive influence on student learning. This is actually why progress testing was originally developed [10, 11], and in the various implementations there is evidence to underpin this positive effect. In McMaster the test led students to study more continuously and to build a better knowledge base, preparing them better for the national licensing examinations [15]. The positive effect of progress testing can be seen clearly from curves showing the growth of medical knowledge. Not only can it be seen that the amount of functional knowledge grows continuously (without huge peaks and troughs), but also that the basic knowledge is retained over the year classes [3, 5, 11, 12, 1618]. Though such continuous growth occurred even if non-problem based learning or non-integrated curricula used progress testing [8, 9], growth curves were more irregular (with more peaks and troughs) when progress testing was not a summative element of the programme [19].
However, no assessment method can exert its influence on student learning in a vacuum; it always works in the context of the rest of the assessment programme [14, 20]. When progress testing was introduced in Maastricht and block tests were made formative, students changed their focus to continuous self-directed learning, but when the—mastery orientated—block test was made summative again, many students reverted to short-term memorization despite the progress test remaining unchanged.

Better predictive validity

Another assumed advantage is that longitudinal data collection is more predictive of future competence/performance than one-off measurements. For this, choices have to be made with respect to how to combine the information of subsequent tests. Some schools opt for a more continuous approach [3] and use regression techniques to make predictions, others acknowledge the discrete nature of the information and combine qualifications [5, 11, 13]. We feel that both are defensible choices but that equating or controlling for difficulty variation is a more pressing issue. Langer et al. [21] have elaborated on this problem and have suggested some solutions. Unfortunately, most solutions are not practical in a medical school setting [2125]. Equating techniques may be impossible to apply in the normal routine (the use of anchor items may induce students to memorize old tests) and item response theory (IRT) may simply require too much pretesting to be practical either. More feasible statistical smoothing techniques such as Bayesian models [24] or moving average techniques [22, 23] on the other hand may be too difficult to explain, especially to students whose original score has to be downgraded by the statistical procedures. This would seriously limit the already rocky base for university acceptance of the concept of progress testing.

Better reliability of decisions

Finally, longitudinal combination of results adds to the reliability of the decision. Research in the 1980s and onwards [26, 27] has made it clear that the sampling properties are much more important for reliability than how well structured the test is [28]. It is logical to assume that the combined result of four tests of 200 items each (in the case of Maastricht) is better than one big test, and a large test distributed over various occasions has better sampling than a one-off large test. Ricketts et al. [29] quantified this using generalizability theory and reported the standard errors of measurement (SEM) as a trade-off between number of items per test and number of tests per year. Their findings indicate that two tests of 200 items per year produce more reliable results (lower SEMs) than four tests of 100 items each, or even five tests of 100 items. So although there is value in having more occasions it is not simply more-occasions-is-better.
Another important discussion point in reliability is that most progress tests employ a correct-minus-incorrect (formula) scoring system. This is necessary because the tests are also administered to junior students. It is not considered desirable that our junior students—not being able to answer most of the questions—would be forced to guess on many items. Therefore, a question-mark option has to be offered with formula scoring. Whether or not this decreases the reliability of progress test scores is open to debate. When the test is taken under formula scoring conditions the number of correct reliabilities is higher—the difference being roughly 0.20 (unpublished results of the interuniversity progress test in the Netherlands)—but experimental studies where scores under formula scoring and number-right conditions were compared showed better reliabilities for the formula scoring [30, 31].

Comprehensive tests are less predictable for the test-savvy students

The comprehensiveness of the test content is often seen as an advantage too, because specific strategic revision does not work (what would you study if the whole of medical knowledge is sampled from?) [3, 11, 15, 32, 33]. So the longitudinality influences the imminence and threatening nature of the test [34] and the comprehensiveness influences the nature of assessable material in such a way that the best preparation is continuous learning [34]. But there is, again, another side to this, as it has to be very clear what the nature of assessable material is. In other words, what is relevant functional knowledge and what is not? This is an issue that still remains unresolved. It will take a feasible operationalization of ‘relevance’ for test writers, reviewers and users to be able to agree on the relevance of each item.

Curriculum independence and collaboration

A final advantage is the progress test’s curriculum independence. The fact that it is designed to test knowledge at graduate level makes it perfect for joint production, joint administration and joint research. The many emerging collaborations [1, 2, 59, 35] are proof of this. This is not to say that collaboration is easy or comes naturally. Schools for example are used to having complete ownership of their assessment material and collaboration means that they have to give up some of that ownership. Also coordination of test administrations, mutual dependency and division of labour may present considerable infrastructural and administrative hurdles [6].

Epilogue

Progress testing is definitely an important addition to the available assessment methods. It has become clear that in a programme of assessment it should not be used to replace current methods but to add to them [20, 36, 37]. Good knowledge of the pros and cons, the indications and contraindications, is a prerequisite for good usage of progress testing, and we hope this paper has contributed to this.

Essentials

  • Progress testing is a longitudinal test approach based on equivalent tests given at fixed intervals with the intention to assess the development on functional knowledge or competence
  • The biggest advantage of progress testing is that it minimizes test-driven learning strategies
  • Combining the results on the repeated tests increases both the reliability of pass–fail decisions and its predictive validity
  • A major concern with progress testing is ensuring the equivalence of the individual tests
  • When progress testing is used in a collaborative fashion—sharing test production and administration—it is not only more cost-effective but also a rich source for continuous benchmarking and quality improvement

Open Access

This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License ( https://​creativecommons.​org/​licenses/​by/​2.​0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Literatuur
1.
go back to reference Aarts R, Steidell K, Manuel BAF, Driessen EW. Progress testing in resource-poor countries: a case from Mozambique. Med Teach. 2010;32:461–3.PubMedCrossRef Aarts R, Steidell K, Manuel BAF, Driessen EW. Progress testing in resource-poor countries: a case from Mozambique. Med Teach. 2010;32:461–3.PubMedCrossRef
2.
go back to reference Bennett J, Freeman A, Coombs L, Kay L, Ricketts C. Adaptation of medical progress testing to a dental setting. Med Teach. 2010;32:500–2.PubMedCrossRef Bennett J, Freeman A, Coombs L, Kay L, Ricketts C. Adaptation of medical progress testing to a dental setting. Med Teach. 2010;32:500–2.PubMedCrossRef
3.
go back to reference Blake JM, Norman GR, Keane DR, Barber Mueller C, Cunnington J, Didyk N. Introducing progress testing in McMaster University’s problem-based medical curriculum: psychometric properties and effect on learning. Acad Med. 1996;71:1002–7.PubMedCrossRef Blake JM, Norman GR, Keane DR, Barber Mueller C, Cunnington J, Didyk N. Introducing progress testing in McMaster University’s problem-based medical curriculum: psychometric properties and effect on learning. Acad Med. 1996;71:1002–7.PubMedCrossRef
4.
go back to reference Freeman A, Van der Vleuten C, Nouns Z, Ricketts C. Progress testing internationally. Med Teach. 2010;32:451–5.PubMedCrossRef Freeman A, Van der Vleuten C, Nouns Z, Ricketts C. Progress testing internationally. Med Teach. 2010;32:451–5.PubMedCrossRef
5.
6.
go back to reference Schuwirth L, Bosman G, Henning R, Rinkel R, Wenink A. Collaboration on progress testing in medical schools in the Netherlands. Med Teach. 2010;32:476–9.PubMedCrossRef Schuwirth L, Bosman G, Henning R, Rinkel R, Wenink A. Collaboration on progress testing in medical schools in the Netherlands. Med Teach. 2010;32:476–9.PubMedCrossRef
7.
go back to reference Swanson D, Holtzman K, Butler A, et al. Collaboration across the pond: the multi-school progress testing project. Med Teach. 2010;32:480–5.PubMedCrossRef Swanson D, Holtzman K, Butler A, et al. Collaboration across the pond: the multi-school progress testing project. Med Teach. 2010;32:480–5.PubMedCrossRef
8.
go back to reference Van der Vleuten C, Schuwirth L, Muijtjens A, Thoben A, Cohen-Schotanus J, Van Boven C. Cross institutional collaboration in assessment: a case on progress testing. Med Teach. 2004;26:719–25.PubMedCrossRef Van der Vleuten C, Schuwirth L, Muijtjens A, Thoben A, Cohen-Schotanus J, Van Boven C. Cross institutional collaboration in assessment: a case on progress testing. Med Teach. 2004;26:719–25.PubMedCrossRef
9.
go back to reference Verhoeven B, Snellen-Balendong H, Hay I, et al. The versatility of progress testing assessed in an international context: a start for benchmarking global standardization? Med Teach. 2005;27:514–20.PubMedCrossRef Verhoeven B, Snellen-Balendong H, Hay I, et al. The versatility of progress testing assessed in an international context: a start for benchmarking global standardization? Med Teach. 2005;27:514–20.PubMedCrossRef
11.
go back to reference Van der Vleuten CPM, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem-based learning curriculum. Med Teach. 1996;18:103–10.CrossRef Van der Vleuten CPM, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem-based learning curriculum. Med Teach. 1996;18:103–10.CrossRef
12.
go back to reference Freeman A, Ricketts C. Choosing and designing knowledge assessments: experience at a new medical school. Med Teach. 2010;32:578–81.PubMedCrossRef Freeman A, Ricketts C. Choosing and designing knowledge assessments: experience at a new medical school. Med Teach. 2010;32:578–81.PubMedCrossRef
13.
go back to reference McHarg J, Bradley P, Chambelain S, Ricketts C, Searle J, McLachlan J. Assessment of progress tests. Med Educ. 2005;39:221–7.PubMedCrossRef McHarg J, Bradley P, Chambelain S, Ricketts C, Searle J, McLachlan J. Assessment of progress tests. Med Educ. 2005;39:221–7.PubMedCrossRef
14.
go back to reference Cohen-Schotanus J. Student assessment and examination rules. Med Teach. 1999;21:318–21.CrossRef Cohen-Schotanus J. Student assessment and examination rules. Med Teach. 1999;21:318–21.CrossRef
15.
go back to reference Norman G, Neville A, Blake J, Mueller B. Assessment steers learning down the right road: impact of progress testing on licensing examination performance. Med Teach. 2010;32:496–9.PubMedCrossRef Norman G, Neville A, Blake J, Mueller B. Assessment steers learning down the right road: impact of progress testing on licensing examination performance. Med Teach. 2010;32:496–9.PubMedCrossRef
16.
go back to reference Ricketts C, Freeman A, Coombes L. Standard setting for progress tests: combining external and internal standards. Med Educ. 2009;43:589–93.PubMedCrossRef Ricketts C, Freeman A, Coombes L. Standard setting for progress tests: combining external and internal standards. Med Educ. 2009;43:589–93.PubMedCrossRef
17.
go back to reference Verhoeven B, Verwijnen G, Scherpbier A, van der Vleuten C. Growth of medical knowledge. Med Educ. 2002;36:711–7.PubMedCrossRef Verhoeven B, Verwijnen G, Scherpbier A, van der Vleuten C. Growth of medical knowledge. Med Educ. 2002;36:711–7.PubMedCrossRef
18.
go back to reference Verhoeven BH, Verwijnen GM, Scherpbier AJJA, Schuwirth LWT, van der Vleuten CPM. Quality assurance in test construction: the approach of a multidisciplinary central test committee. Educ Health. 1999;12:49–60. Verhoeven BH, Verwijnen GM, Scherpbier AJJA, Schuwirth LWT, van der Vleuten CPM. Quality assurance in test construction: the approach of a multidisciplinary central test committee. Educ Health. 1999;12:49–60.
19.
go back to reference Albano MG, Cavallo F, Hoogenboom R, et al. An international comparison of knowledge levels of medical students: the Maastricht Progress Test. Med Educ. 1996;30:239–45.PubMedCrossRef Albano MG, Cavallo F, Hoogenboom R, et al. An international comparison of knowledge levels of medical students: the Maastricht Progress Test. Med Educ. 1996;30:239–45.PubMedCrossRef
20.
go back to reference van der Vleuten C, Schuwirth L. Assessing professional competence: from methods to programmes. Med Educ. 2005;39:309–17.PubMedCrossRef van der Vleuten C, Schuwirth L. Assessing professional competence: from methods to programmes. Med Educ. 2005;39:309–17.PubMedCrossRef
21.
go back to reference Langer M, Swanson D. Practical considerations in equating progress tests. Med Teach. 2010;32:509–12.PubMedCrossRef Langer M, Swanson D. Practical considerations in equating progress tests. Med Teach. 2010;32:509–12.PubMedCrossRef
22.
go back to reference Muijtjens A, Timmermans I, Donkers J, et al. Flexible electronic feedback using the virtues of progress testing. Med Teach. 2010;32:491–5.PubMedCrossRef Muijtjens A, Timmermans I, Donkers J, et al. Flexible electronic feedback using the virtues of progress testing. Med Teach. 2010;32:491–5.PubMedCrossRef
23.
go back to reference Muijtjens A, Schuwirth L, Cohen-Schotanus J, van der Vleuten C. Differences in knowledge development exposed by multi-curricular progress test data. Adv Health Sci Educ. 2008;13:593–605.CrossRef Muijtjens A, Schuwirth L, Cohen-Schotanus J, van der Vleuten C. Differences in knowledge development exposed by multi-curricular progress test data. Adv Health Sci Educ. 2008;13:593–605.CrossRef
24.
go back to reference Ricketts C, Moyeed R. Improving progress test score estimation using Bayesian statistics. Med Educ. 2011;45:570–7.PubMedCrossRef Ricketts C, Moyeed R. Improving progress test score estimation using Bayesian statistics. Med Educ. 2011;45:570–7.PubMedCrossRef
25.
go back to reference Schauber S, Nouns Z. Using the cumulative deviation method for cross-institutional benchmarking in the Berling progress test. Med Teach. 2010;32:471–5.PubMedCrossRef Schauber S, Nouns Z. Using the cumulative deviation method for cross-institutional benchmarking in the Berling progress test. Med Teach. 2010;32:471–5.PubMedCrossRef
26.
go back to reference Swanson DB, Norcini JJ. Factors influencing reproducibility of tests using standardized patients. Teach Learn Med. 1989;1:158–66.CrossRef Swanson DB, Norcini JJ. Factors influencing reproducibility of tests using standardized patients. Teach Learn Med. 1989;1:158–66.CrossRef
27.
go back to reference Van der Vleuten CPM, Swanson D. Assessment of clinical skills with standardized patients: state of the art. Teach Learn Med. 1990;2:58–76.CrossRef Van der Vleuten CPM, Swanson D. Assessment of clinical skills with standardized patients: state of the art. Teach Learn Med. 1990;2:58–76.CrossRef
28.
go back to reference Van der Vleuten CPM, Norman GR, De Graaf E. Pitfalls in the pursuit of objectivity: issues of reliability. Med Educ. 1991;25:110–8.PubMedCrossRef Van der Vleuten CPM, Norman GR, De Graaf E. Pitfalls in the pursuit of objectivity: issues of reliability. Med Educ. 1991;25:110–8.PubMedCrossRef
29.
go back to reference Ricketts C, Freeman A, Pagliuca G, Coombes L, Archer J. Difficult decisions for progress testing: how much and how often? Med Teach. 2010;32:513–5.PubMedCrossRef Ricketts C, Freeman A, Pagliuca G, Coombes L, Archer J. Difficult decisions for progress testing: how much and how often? Med Teach. 2010;32:513–5.PubMedCrossRef
30.
go back to reference Medema H. The effect of formula scoring versus number right scoring on partial knowledge and reliability in Progress testing. Maastricht: Department of Educational Development and Research, Maastricht University; 2010. p. 33. Medema H. The effect of formula scoring versus number right scoring on partial knowledge and reliability in Progress testing. Maastricht: Department of Educational Development and Research, Maastricht University; 2010. p. 33.
31.
go back to reference Muijtjens AMM, van Mameren H, Hoogenboom RJI, Evers JLH, Van der Vleuten C. The effect of a ‘don’t know’ option on test scores: number-right and formula scoring compared. Med Educ. 1999;33:267–75.PubMedCrossRef Muijtjens AMM, van Mameren H, Hoogenboom RJI, Evers JLH, Van der Vleuten C. The effect of a ‘don’t know’ option on test scores: number-right and formula scoring compared. Med Educ. 1999;33:267–75.PubMedCrossRef
32.
go back to reference Van Berkel HJM, Nuy HJP, Geerligs T. The influence of progress tests and block tests on study behaviour. Instr Sci. 1995;22:317–22.CrossRef Van Berkel HJM, Nuy HJP, Geerligs T. The influence of progress tests and block tests on study behaviour. Instr Sci. 1995;22:317–22.CrossRef
33.
go back to reference Van Til C. Voortgang in Voortgangstoetsing [Progress in Progress Testing]. Educational Research and Educational Development. University of Maastricht, Maastricht; 1998. Van Til C. Voortgang in Voortgangstoetsing [Progress in Progress Testing]. Educational Research and Educational Development. University of Maastricht, Maastricht; 1998.
34.
go back to reference Cilliers F, Schuwirth L, Herman N, Adendorff H, van der Vleuten C. A model of the sources, consequences and mechanism of impact of summative assessment on how students learn. Adv Health Sci Educ. 2011. Cilliers F, Schuwirth L, Herman N, Adendorff H, van der Vleuten C. A model of the sources, consequences and mechanism of impact of summative assessment on how students learn. Adv Health Sci Educ. 2011.
35.
go back to reference De Champlain A, Cuddy M, Scoles P, et al. Progress testing in clinical science education: results of a pilot project between the National Board of Medical Examiners and a US medical schools. Med Teach. 2010;32:503–8.PubMedCrossRef De Champlain A, Cuddy M, Scoles P, et al. Progress testing in clinical science education: results of a pilot project between the National Board of Medical Examiners and a US medical schools. Med Teach. 2010;32:503–8.PubMedCrossRef
36.
go back to reference Dijkstra J, Galbraith R, Hodges B, et al. Development and validation of guidelines for designing programmes of assessment: a modified Delphi-study, submitted. Dijkstra J, Galbraith R, Hodges B, et al. Development and validation of guidelines for designing programmes of assessment: a modified Delphi-study, submitted.
37.
go back to reference Dijkstra J, Van der Vleuten C, Schuwirth L. A new framework for designing programmes of assessment. Adv Health Sci Educ. 2010;15:379–93. Dijkstra J, Van der Vleuten C, Schuwirth L. A new framework for designing programmes of assessment. Adv Health Sci Educ. 2010;15:379–93.
Metagegevens
Titel
The use of progress testing
Auteurs
Lambert W. T. Schuwirth
Cees P. M. van der Vleuten
Publicatiedatum
01-03-2012
Uitgeverij
Bohn Stafleu van Loghum
Gepubliceerd in
Perspectives on Medical Education / Uitgave 1/2012
Print ISSN: 2212-2761
Elektronisch ISSN: 2212-277X
DOI
https://doi.org/10.1007/s40037-012-0007-2

Andere artikelen Uitgave 1/2012

Perspectives on Medical Education 1/2012 Naar de uitgave

Review Article

Workplace learning