INTRODUCTION

Despite significant advances in patient safety since the seminal 1999 Institute of Medicine report, To Err is Human,1 the health care system continues to grapple with an excess of medical errors.2 , 3 Diagnostic errors are implicated in a huge proportion of cases, affecting millions of individuals and causing an estimated 40,000–80,000 deaths each year in the United States alone.4 There are many causes of diagnostic error, but cognitive factors (e.g., faulty medical knowledge, data gathering and synthesis) resulting in diagnostic reasoning failures rank among the most important.5 , 6 As medical school training provides the foundation for the development of clinical reasoning skills, undergraduate medical education can potentially play a critical role in improving diagnostic accuracy and mitigating diagnostic error. This recognition, along with the current national spotlight on reducing diagnostic errors, naturally prompts an examination of how clinical reasoning is taught within medical schools.

Traditionally, medical schools begin clinical reasoning instruction in the preclinical years through a series of “doctoring” courses focused on topics such as patient interviewing, the physical examination, and pathophysiology. However, these courses typically do not explicitly address the clinical reasoning process in a structured fashion. Interactions with patients and exposure to resident and attending role models during the clinical years thus serve as the major means of fostering clinical reasoning capabilities in learners. This unstandardized process depends on the teaching skills of role models. Stated another way, clinical reasoning ability is often “caught” by learners as opposed to being explicitly “taught” to them.

Recent advances in cognitive psychology pertaining to diagnostic reasoning have provided theoretical insights for developing a framework for clinical reasoning curricula. As a result, reports of clinical reasoning courses have emerged in the medical education literature,7,8,9,10, 11 yet data regarding the national state of clinical reasoning curricula in medical schools are lacking. Such data are needed to enable educators and policymakers to propose optimal solutions at the medical school level to help mitigate medical errors. To address this gap in the literature, we developed a set of questions, administered through a national survey of designated institutional representatives of the Clerkship Directors of Internal Medicine (CDIM), to investigate the content and format of clinical reasoning teaching in US medical schools. Additionally, we sought to determine clerkship directors’ attitudes toward clinical reasoning curricula and to identify barriers to the adoption of such curricula.

METHODS

Design

We used a cross-sectional survey methodology to obtain data about current clinical reasoning curricula and clerkship director attitudes regarding these curricula at US medical schools.

Setting and Participants

We used data from the 2015 CDIM annual survey of US medical schools. The designated CDIM institutional representatives at 123 allopathic medical schools were contacted via email and provided with a web-based questionnaire. Non-responders received up to three follow-up requests by email and telephone.

Survey Development and Data Collection

The 2015 CDIM survey was developed from member submissions in response to an open invitation extended in the fall of 2014 for research proposals addressing important topics and questions regarding medical student education. The CDIM Survey and Scholarship Committee reviewed all proposals and selected five topics based on relevance to the current educational climate. The section on clinical reasoning education was one of five components of the 2015 survey. Other sections of the survey, which comprised a total of 126 questions, explored longitudinal integrated clerkships, milestones and Core Entrustable Professional Activities for Entering Residency (CEPAER), direct observation of learners, and duties of clerkship directors. Another section of the survey collected demographic information including the respondent’s academic rank, age, sex, and primary role in internal medicine education.

Three of the authors (JR, RT, SJD), who recently conducted a literature review and served as editors on a textbook regarding clinical reasoning teaching,12 developed the initial draft of the survey questions pertaining to clinical reasoning. The questions were reviewed and revised by members of the Survey and Scholarship Committee in close collaboration with the authors. Once the survey questions were edited, a draft of the entire survey was reviewed and revised by the CDIM Council. The survey items were also pilot-tested with members of the survey committee and the CDIM Council, followed by final revisions. The final survey was emailed using SurveyMonkey® on May 20, 2015, to the designated CDIM representative at each member institution. These representatives are typically current internal medicine clerkship directors, or serve in a role that has direct oversight of the internal medicine clerkship. The list of designated institutional representatives is updated annually, and every effort is made to ensure that the survey is sent to the person at each institution with detailed knowledge of the internal medicine clerkship. All data were anonymized. The survey closed on August 17, 2015. The study was approved by the institutional review board at the Washington DC VA Medical Center.

The topics and terms of the clinical reasoning questions were chosen based on a literature review and on the expertise of the author group. Eight questions with multiple sub-items on clinical reasoning education were developed, and included free text as well as rating scales, including modified Likert scales (online Appendix 1). Questions were focused on foundational topics of clinical reasoning, including the underlying theory (dual process, illness scripts), reasoning strategies (pre-test probability, thresholds to test and treat, heuristics, likelihood ratios, decision analysis), and common causes of diagnostic error (premature closure, cognitive bias). Premature closure and cognitive bias were both selected, because they are considered common causes of error and are familiar concepts to medical educators. The questions focused on the respondents’ perceptions of three aspects of these topics: current and ideal curricula, importance, and student knowledge.

Statistical Analysis

The results were analyzed in a de-identified, confidential format. Institutional respondents were used as the unit of analysis. Unanswered items were not included in the analysis. Descriptive statistics were performed on all responses and calculated using a standard statistical software program (SPSS version 12; SPSS Inc., Chicago, Illinois).

RESULTS

Ninety-five of 123 (77%) institutional representatives completed the CDIM survey. The institutional geographic distribution represented every region of the country, with 59 (62%) public and 36 (38%) private medical schools responding. The majority of respondents identified their role (role reported by n = 90) as clerkship director, co-clerkship director, or associate clerkship director (n = 84). The other six comprised five vice chairs for education and one assistant dean. The number of responses by sex were almost equal (n = 87), with 43 (49%) women and 44 (51%) men. The academic rank distribution (n = 91) was as follows: 43 associate professors (47%), 22 professors (24%), 25 assistant professors (28%), and one instructor (1.0%). The majority of respondents (n = 60) were between the ages of 35 and 55 years. Four respondents did not complete any portion of the clinical reasoning education section and were excluded from our analysis, resulting in a 74% response rate (n = 91 respondents).

The majority of respondents felt that a structured curriculum in clinical reasoning should be taught in all phases of medical education, including the preclinical years (3.9, SD 0.7), clinical clerkships (4.4, SD 0.7), and the fourth year (4.3, SD 0.7, Table 1). Most reported that students had a fair or poor understanding of all clinical reasoning concepts presented, with very few (< 5%) answering that students had an excellent understanding of these concepts (Table 2). Eighty percent of respondents used small group case discussions to teach clinical reasoning, with 96% agreeing that it should be taught in this way. Clinical reasoning teaching most commonly occurred in attending rounds (76%) and morning report (66%). Nearly all respondents felt that these were the two best venues for teaching it (96 and 91%, respectively; Table 3).

Table 1 Perceived Importance of the Timing of a Structured Curriculum to Explicitly Teach Concepts of Clinical Reasoning
Table 2 Student Understanding of Clinical Reasoning Concepts upon Entering Clinical Clerkships
Table 3 Venues for Teaching Clinical Reasoning Topics in Internal Medicine Clerkships (n = 91)

Respondents endorsed the importance of teaching clinical reasoning concepts in the clerkship years, rating all potential topics as at least moderately important (Table 4). The most important topics were premature closure (3.9, SD 1.0), pre-test probability estimation (3.8, SD 1.0), thresholds to test and treat (3.8, SD 1.0), and cognitive bias (3.7, SD 1.0). Formal decision analysis (2.60, SD 1.0), the use of likelihood ratios (3.2, SD 0.9), and dual process theory (3.3, SD 0.9) were considered the least important. Despite the expressed importance of teaching clinical reasoning concepts in the clinical years, most sites did not have sessions specifically devoted to these topics. Only pre-test probability estimation was specifically taught in a majority of sites (54.8%), with the other topics ranging from 16.7 to 48.8%.

Table 4 Importance of Teaching Clinical Reasoning Topics and Current Coverage in Internal Medicine Clerkships

The time dedicated to clinical reasoning in the internal medicine clerkships ranged from 0 to 32 h, with an average of 6.4 h over the course of the clerkship. Most clerkship directors (67%) believed that more time should be dedicated to clinical reasoning teaching in the clerkship. Respondents identified a lack of curricular time (87%) and a lack of faculty qualified to teach these concepts (70%) as barriers. Few respondents (16%) believed that the perception that clinical reasoning cannot be taught was a significant barrier.

DISCUSSION

These results from a national survey of important institutional educators provide the first empirical evidence regarding the content of, venues for, and barriers to clinical reasoning teaching in undergraduate medical education. Our North American medical school survey showed that the majority of CDIM institutional representatives felt that clinical reasoning should be taught in all phases of undergraduate medical education, which supports similar recommendations by authoritative bodies such as the Institute of Medicine2 and the Coalition to Improve Diagnosis.12

Respondents’ reports that students with poor to fair knowledge of clinical reasoning receive an average of only 6.4 h of dedicated clinical reasoning teaching highlight the need for a specific emphasis on the development of clinical reasoning skills throughout the educational process. Our study supports the notion that learners may not be able to gain proficiency in clinical reasoning solely through patient interactions and exposure to role models. Although these modalities will always remain a cornerstone for the development of clinical reasoning abilities, our data suggest that this process should be nurtured and supported by specific curricular programs in all phases of medical education, and should be a specific and explicit goal of medical education.

Our survey attempted to address why medical schools are lagging in developing curricula dedicated to teaching clinical reasoning. Clerkship directors identified barriers to implementing additional instruction in clinical reasoning, including lack of curricular time and qualified faculty, with curricular time reported as the most significant barrier. The role of the physician in medicine has evolved dramatically over the past decade, and medical schools are now tasked with incorporating curricular initiatives centered on quality, patient safety, and systems engineering into an already dense educational agenda.13 Adding further instruction in clinical reasoning will undoubtedly be a challenge. Faculty expertise is also a significant issue. Although many clinical faculty may be experts in their fields, they may not be experts in teaching clinical reasoning. Most physicians can teach clinical reasoning by modeling the process and “thinking out loud,” but certain topics may be more difficult for the traditional clinician educator who may have never explicitly learned about them. An additional problem is the lack of a gold standard for assessing clinical reasoning skills, making it difficult to determine the efficacy of educational interventions. Current research in clinical reasoning assessment may remedy this issue in the future.14

Most of the topics surveyed were considered important to teach, with estimation of pre-test probability, thresholds for testing and treatment, premature closure, and cognitive biases leading the list. This is a logical prioritization by clerkship directors, as the first two relate to differential diagnosis and decision making, two essential aspects of clinical reasoning. The assigned importance of premature closure and cognitive bias is also not surprising in light of the increased focus on the advantages and disadvantages of heuristics in both the popular press15 , 16 and the medical literature.17 , 18 Cognitive error is the most common reason for diagnostic error,5 but continued research into how these types of errors may be reduced is needed before a robust evidence-based curriculum can be developed.17,18,19,20, 21

Notably, respondents felt that clinical reasoning teaching should be increased in all venues mentioned. However, they felt least strongly about web-based learning. We speculate that this relates to concerns regarding the learning transfer of clinical reasoning skills from online educational activities to actual patient encounters. Learning transfer poses a challenge in all of these venues, although discussion and examination of actual patients in attending rounds and physical diagnosis sessions, respectively, would be expected to enhance it. Further study is needed to explore the issue of learning transfer in clinical reasoning teaching activities.

Limitations of our survey include a lack of information regarding institutional characteristics, which could result in non-response bias. However, our response rate suggests that the results include a broad spectrum of institutions in terms of size, location, and academic mission. Survey length constraints prevented the inclusion of terminology definitions, so some respondents may not have understood the meaning of the clinical reasoning terms. However, given the recent emphasis on clinical reasoning and diagnostic error in the medical education literature, we suspect that most experienced medical educators are familiar with these terms.

Other limitations include the exclusive focus on internal medicine, a specialty in which clinical reasoning is considered an essential skill. Therefore, our data likely provide a “best-case scenario” estimation of clinical reasoning teaching and are not generalizable to other specialties. Furthermore, clerkship directors are typically more knowledgeable about the clinical than the preclinical years in undergraduate medical education, potentially limiting their ability to accurately describe the scope of instruction in clinical reasoning. However, the survey was distributed to designated institutional representatives who were expected to confer with the appropriate colleagues as needed to answer the survey items. Finally, clerkship directors were not asked to prioritize clinical reasoning relative to other curricular content, so the results may overestimate the importance of clinical reasoning teaching. However, most medical educators view clinical reasoning as an essential skill for the practicing clinician, and the fact that the Institute of Medicine views education on diagnostic error as an important priority suggests that clerkship directors are aligned with national thought leaders in their views.

These survey data suggest that students lack knowledge of key concepts in clinical reasoning and that a greater emphasis should be placed on clinical reasoning teaching in both the pre-clinical and clinical years. Currently, however, limited curricular time devoted to clinical reasoning and a lack of faculty expertise in this area are major barriers. Given the pressing need to improve diagnostic accuracy and reduce diagnostic error, we believe these data sound a clarion call for medical schools to re-prioritize resources toward building clinical reasoning curricula and developing faculty expertise. Further research to explore interventions for overcoming existing barriers is imperative. Ultimately, we believe that the medical education community should aim to develop a set of national standards and objectives regarding the teaching of clinical reasoning, which medical schools can use to guide their curricular design efforts.