Introduction
Methods
Data sources and searches
Study selection
Activities intended to promote or measure the clinical knowledge/skills of physicians in independent medical practice through a) courses or assessments delivered in any modality or venue, whether or not [continuing medical education] credit is awarded, or b) self-directed learning or self-assessment activities for which credit is awarded [19].
Data extraction and quality appraisal
Domain: Item | Operational adjustments | Level | Prevalence N (%) (N = 62) |
---|---|---|---|
Study design | Added option for economic modeling studies (score 1.5) | 1‑group post-only (1) | 6 (10%) |
1‑group pre-post, or modeling (1.5) | 20 (32%) | ||
2‑group non-randomized (2) | 16 (26%) | ||
2‑group randomized (3) | 20 (32%) | ||
Sampling: No. of institutions studied | No change | 1 (0.5) | 54 (87%) |
2 (1) | 1 (2%) | ||
>2 (1.5) | 7 (11%) | ||
Sampling: Response rate | For cost data: Data derived from large record sets unlikely to reflect bias (e.g., institutional electronic health record or regional claims database) count as high (score 1.5) | <50% or not specified (0.5) | 24 (39%) |
50–74% (1) | 7 (11%) | ||
≥75% or large record | 31 (50%) | ||
Type of data (data source) | For cost data: Details of resource quantitation (both data source and quantity [number of units, not just total cost]) count as high (score 3). Cost alone counts as low (score 1) | Self-reported data, or cost without resource quantitation (1) | 8 (13%) |
Objective measurement, or cost with data source and quantity (3) | 54 (87%) | ||
Validation of evaluation instrument: Content | For cost data: “The degree to which the cost estimation encompasses all aspects of the true cost, encompassing processes to both identify and measure cost” [15]. Evidence could include use of a formal framework (e.g., the Ingredients Method) or the involvement of experts in planning, empiric identification and selection of relevant resources (e.g., time-motion studies or process mapping), and substantiation that a robust data source was used to select, quantitate, or price resources (e.g., detailed description of a computer database) | Reported (1) | 8 (13%) |
Validation of evaluation instrument: Internal structure | For cost data: “The degree to which the cost estimate is reproducible if the same method is followed” [15]. Evidence could include replicability of the valuation or analysis (e.g., robust examination of the uncertainty of input parameter estimates [sensitivity analysis], independent valuation of costs by two investigators [inter-rater reliability], or comparing cost estimates derived at two different time points [temporal stability]) | Reported (1) | 9 (15%) |
Validation of evaluation instrument: Relations with other variables | For cost data: “The degree to which the cost estimate relates to cost estimates formed using alternative approaches” [15]. Evidence could include examining predicted associations among results obtained using alternative approaches to economic modeling (e.g., sensitivity analysis comparing different base assumptions, valuation methods, statistical models, or economic theories) | Reported (1) | 1 (2%) |
Data analysis: Appropriateness | For cost data: The following count as “appropriate” (score 1): cost effectiveness ratio, net benefit, or other similar analysis of cost data | Inappropriate for study design (0) | 37 (60%) |
Appropriate (1) | 25 (40%) | ||
Data analysis: Complexity | For cost data: The following count as “beyond descriptive” (score 2): cost effectiveness ratio, net benefit, visual display of cost-effectiveness | Descriptive analysis only (1) | 37 (60%) |
Beyond descriptive analysis (2) | 25 (40%) | ||
Outcomes | For cost outcomes: As per Foo, we distinguished education costs in a “test setting” or a “real setting,” namely: “Test settings are those in which the context does not match how the intervention would be utilized in actual practice (e.g., a hypothetical program that was not actually implemented). Real settings are where the intervention is evaluated in a context similar to its anticipated utilization in practice (e.g., an evaluation of a program that is taught to real students)” [15]. However, we assigned points differently than Foo: score 1.5 for cost of education in test setting, score 2 for cost of education in real setting, score 3 for health care costs. Outcomes estimated from previously published research (including health care costs and non-cost outcomes) also score 1.5 | Knowledge, skills, or education costs in a “test” or hypothetical training setting, or estimated from literature (1.5) | 1 (2%) |
Behaviors in practice or education costs in a “real” training setting (2) | 25 (40%) | ||
Patient effects, including health care costs (3) | 36 (58%) |
Data synthesis and analysis
Results
Participants and CPD topics
Reporting quality overall: CHEERS elements
Reporting quality of abstracts
Methodological quality
-
Perspective: 19 studies (31%) noted the economic perspective.
-
Comparison: The reference case recommends comparison “ideally be made with a plausible alternate active intervention”; this was done in 17 of 62 studies (27%). Comparison with a no-intervention group (which typically offers less useful insights than an active comparison) was done in 22 studies (35%). Single-group designs, which “contribute little to generalizable knowledge and are not preferred,” were used in 26 studies (42%).
-
Description of intervention(s): Details were reported above.
-
Non-economic measures of effectiveness: 56 studies (90%) reported effectiveness outcomes, including participation rates or satisfaction (18 [29%]), knowledge/skills (10 studies [16%]), physician behaviors (39 [63%]), and effects on patients (30 [48%]).
-
Training costs: As per inclusion criteria, all studies reported training costs. The reference case recommends including “time invested by faculty, support staff, and learners in all valuations of training cost.” Time was reported for faculty, staff, learners, or any of these, in 21 (34%), 16 (26%), 16 (26%), and 29 (47%) studies, respectively.
-
Valuation: Details were reported above for methods of selection, quantitation, and pricing. Additionally, 13 studies (21%) explicitly excluded potentially relevant resources such as startup costs, capital expenses, faculty or learner time, and donated materials.
-
Analysis: The reference case recommends reporting combined “expressions of cost and impact (e.g., cost-effectiveness ratio, net value, or benefit-cost ratio)”; this was done in 25 (44%) of the 57 studies reporting effectiveness outcomes. Statistical comparisons of costs were performed in only 1 study (2%). Descriptive numeric analyses of costs, such as cost-effectiveness ratios and net benefits, were more common but still infrequent (17 [27%] and 9 [15%], respectively). Visual analyses (cost-effectiveness plane, decision acceptability curve) were employed in 2 studies (3%).
-
Sensitivity analyses: 9 studies (15%) reported sensitivity analyses exploring uncertainty in training cost estimates (6 other studies reported sensitivity analyses for clinical cost outcomes). Probabilistic sensitivity methods were used in 4 studies (6%). Two studies (3%) conducted subgroup analyses exploring heterogeneity among training subgroups (e.g., across implementation sites).
-
Discounting: 9 studies (15%) employed discounting in cost valuations, and 3 others (5%) explained why discounting was unnecessary in their study.