Evidence-Based Medicine
‘In God we trust; all others must bring data.’
—W. EDWARDS DEMING, PHYSICIST AND QUALITY IMPROVEMENT PIONEER
Evidence-based medicine (EBM) is defined as the conscientious, explicit, and judicious use of the current best evidence in making decisions about the care of individual patients.1 A simpler concept would be treating patients based on data rather than the surgeon’s thoughts or beliefs. EBM represents the concept that medical practice can be largely dictated by evidence gained from the scientific method. Given that the practice of medicine has historically been based on knowledge handed down from mentor to apprentice, the concepts of EBM embody a new paradigm, replacing the traditional paradigm that was based on authority. In a global sense, it describes a methodology for evaluating the validity of clinical research and applying those results to the care of patients.
History
The initial groundwork forming the framework for EBM can be considered the earliest scientists who pursued explanatory truth instead of accepting beliefs. The process of discovering truth became replicable for aspiring scientists when the scientific method was developed. No one individual can be credited for creating the scientific method because it is the result of the progressive recognition for a natural process of acquiring facts. However, the earliest publication alluding to the steps of current scientific methodology may be found in Book of Optics published in 1021 by Ibn al-Haytham (Alhazen).2 His investigations were based on experimental evidence. Furthermore, his experiments were systematic and repeatable as he demonstrated that rays of light are emitted from objects rather than from the eyes. In Western literature, Roger Bacon wrote about a repeating cycle of observation, hypothesis, and experimentation in the 1200s. Influenced by the contributions of many scientists and philosophers, Francis Bacon delineated a recognizable form of the scientific method in the 1620 publication of Novum Organum Scientificum.3 He suggested that mastery of the world in which man lives is dependent on careful understanding. Moreover, this understanding is based entirely on the facts of this world and not, as the ancients portrayed it, in philosophy. Nearly 400 years later, we find ourselves coming to the same conclusions in the practice of medicine and surgery. We now understand that facts and truths transcend experience, and that facts about optimal care can be gained through experimentation more reliably than from beliefs generated through experience.
The establishment of the scientific method in the pursuit of proven truths was fundamental to the development of EBM as an entity. The current scientific method consists of several steps that are outlined in Box 78-1. However, the concepts of EBM are not simply encompassed by the application of these steps to attain facts, but also address the ability to understand the value of the results generated from experimentation. Under the auspices of EBM, investigators are burdened with the first five steps, from inquiry to experimentation, while all caregivers must develop a deep understanding of experimental methodology, analysis of data, and how conclusions are drawn to be able to place appropriate value on published studies to influence their practice.
The movement urging physicians to utilize proven facts in the development of decision-making algorithms began in 1972 with the publication of the revolutionary book Effectiveness and Efficiency: Random Reflections on Health Services.4 The author, Archie Cochrane, a Scottish epidemiologist working in the UK National Health Service, has likely had the greatest influence on the development of an organized means of guiding care through data and results. The book demonstrates disdain for the scientific establishment, devalues expert opinion, and shows that physicians should systematically question what is the best care for the patient. Most impressively, Cochrane calls for an international registry of randomized controlled trials and for explicit quality criteria for appraising published research. At the time of his passing in 1988, these aspirations had not fully materialized. However, the field of medicine is fortunate that Cochrane’s prophetic ideas have precipitated the maturation of centers of EBM research that make up the global not-for-profit organization called the Cochrane Collaboration. The product of this collaboration is The Cochrane Library, which is a collection of seven databases that contain high-quality, independent evidence to inform health care decision-making.5 The most clinically utilized database is the Cochrane Database of Systematic Reviews containing reviews on the highest level of evidence on which to base clinical treatment decisions. There are now over 550,000 entries.
After Cochrane and other proponents of EBM showed the importance of comparative data, it took over a decade for the amorphous clouds of these concepts to solidify into tangible terms and usable methods. The methodology of specifically qualifying data and judging the relative merits of available studies did not begin to appear in the literature until the early 1990s. In a paper published in 1990 by David Eddy on the role of guidelines in medical decision-making, the term evidence-based first appeared in the literature.6 Much of the framework currently used to determine best available evidence was established by David Sackett and Gordon Guyatt at the McMaster University-based research group called the Evidence-Based Medicine Working Group. In 1992, JAMA published the group’s landmark paper titled, ‘Evidence-Based Medicine: A New Approach to Teaching the Practice of Medicine,’ and the term evidence-based medicine was born.7 A cementing moment in the paradigm shift occurred when the Centre for Evidence-Based Medicine was established in Oxford, England, in 1995 as the first of several centers. Over the past decade, centers or departments focusing on clinical research and EBM have been developed at universities and hospitals around the world, including our own (www.cmhclinicaltrials.com). EBM principles are now represented in the core curriculum that is integral to three of the six general competencies outlined by the Accreditation Council for Graduate Medical Education that oversees the accredited residency training programs in the United States. The concept of EBM is no longer a movement of progressive physicians, but rather a basic guiding principle of medical training and practice.
Levels of Evidence
Level 1—evidence is supported by prospective, randomized trials
Level 2—evidence is supported by cohort studies, outcomes data, or low-quality prospective trials
Level 3—evidence comprises case-control studies
Level 4—evidence is based on case series
Level 5—evidence is expert opinion or beliefs based on rational principles.
The quality of data that is conveyed within each level and study type is clearly a wide spectrum. The complete delineation in the levels of evidence as defined by the Oxford Centre for Evidence-Based Medicine is outlined in Table 78-1.
TABLE 78-1
Levels of Evidence as Defined by the Oxford Centre for Evidence-Based Medicine
Level | Type of Evidence |
1a | Systematic review of randomized trials displaying homogeneity |
1a− | Systematic review of randomized trials displaying worrisome heterogeneity |
1b | Individual randomized controlled trials (with narrow confidence interval) |
1b− | Individual randomized controlled trials (with a wide confidence interval) |
1c | All or none randomized controlled trials |
2a | Systematic reviews (with homogeneity) of cohort studies |
2a− | Systematic reviews of cohort studies displaying worrisome heterogeneity |
2b | Individual cohort study or low-quality randomized controlled trials (<80% follow-up) |
2b− | Individual cohort study or low-quality randomized controlled trials (<80% follow-up/wide confidence interval) |
2c | ‘Outcomes’ research; ecological studies |
3a | Systematic review (with homogeneity) of case-control studies |
3a− | Systematic review of case-control studies with worrisome heterogeneity |
3b | Individual case-control study |
4 | Case series (and poor quality cohort and case-control studies) |
5 | Expert opinion without explicit critical appraisal, or based on physiology, bench research, or ‘first principles’ |
As can be seen by the levels listed in Table 78-1, the strength of evidence improves significantly by the application of prospective data collection. In clinical medicine, and particularly in the practice of surgery, many aspects of trial design are not feasible such as blinding, placebo treatments, independent follow-up evaluation, and others. However, if one accepts these limitations and conducts a trial with prospective evaluation, the results remain markedly more meaningful than a retrospective case-control comparative series that compares surgeons and/or timeframes against one another.
The review of several studies can gain strength over an individual study, which is valid in many models and fields of medicine. However, one should be cautioned about the real strengths of such a meta-analysis before considering it to have a high level of evidence. The strength of these combined reviews is derived from the strength of the individual trials providing the numbers for the analysis. In the best scenario, such a combined review is composed of multiple prospective trials with similar design that each compares the effect of two treatments on a specific outcome. However, in the field of pediatric surgery, multiple prospective trials with similar designs that address the same disease with the same interventions are nonexistent. Also, surgeons should not overvalue the influence of combined reviews derived from retrospective studies. Such reviews should be interpreted as a mosaic of the individual studies. Any attempt by authors to combine a number of retrospective studies to harness statistical power is fraught with hazard.
Grades of Recommendation
The quality of the evidence as defined in Tables 78-1 and 78-2 applies a score of strength for each individual contribution in the literature. However, on many topics, particularly common clinical scenarios, there is an abundance of published studies such that the appropriate care cannot be guided by a single study. The total body of available information places caregivers in the difficult position of evaluating the published principles from the multiple sources that make up practice guidelines. The caregiver applying these clinical practice guidelines and other recommendations needs to know how much confidence can be placed in the recommendations from a conglomerate of citations. Strength of recommendation scales were born from this need.8–12 Given that the level of evidence indicates the extent to which one can be confident that an estimate of effect is correct, the strength or grade of recommendation indicates the extent to which one can be confident that adherence to the recommendation will do more good than harm.13 As with the levels of evidence, there are multiple published grading scales. The grading format used by the Oxford Centre for Evidence-Based Medicine is outlined in Table 78-2.
TABLE 78-2
Grades of Recommendation as Defined by the Oxford Centre for Evidence-Based Medicine
Grade | Level of Evidence |
A | Consistent level 1 studies |
B | Consistent level 2 or 3 studies or extrapolations from level 1 studies |
C | Level 4 studies or extrapolations from level 2 or 3 studies |
D | Level 5 evidence or troublingly inconsistent or inconclusive studies of any level |