Evidence-based medicine

Evidence-based medicine is "the conscientious, explicit and judicious use of current best evidence in making decisions about the care of individual patients.". Alternative definitions are "the process of systematically finding, appraising, and using contemporaneous research findings as the basis for clinical decisions" or "evidence-based medicine (EBM) requires the integration of the best research evidence with our clinical expertise and our patient's unique values and circumstances." Better known as EBM, evidence based medicine has roots in clinical epidemiology and the scientific method, and emerged in the early 1990's in response to discoveries about variations and deficiencies in medical care to help healthcare providers and policy makers evaluate the efficacy of different treatments.

Evidence-based practice is not restricted to medicine; dentistry, nursing and other allied health science are adopting "evidence-based medicine" as well as alternative medical approaches, such as acupuncture. Evidence-Based Health Care or evidence-based practice extends the concept of EBM to all health professions, including management and policy.

Classification
Two types of evidence-based medicine have been proposed:


 * Evidence-based guidelines, EBM at the organizational or institutional level, which involves producing guidelines, policy, and regulations.

and
 * Evidence-based individual decision making, EBM as practiced by an individual health care provider when treating an individual patient. There is concern that evidence-based medicine focuses excessively on the physician-patient dyad and as a result miss many opportunities to improve healthcare.

Steps in EBM

 * See also Teaching evidence-based medicine

Ask
To "ask" and formulate a well-structured clinical question, i.e. one that is directly relevant to the identified problem, and which is constructed in a way that facilitates searching for an answer. The question should have four 'PICO' elements : the patient or problem (P); the medical intervention or exposure (e.g., a cause for a disease) (I); the comparison intervention or exposure (C); and the clinical outcomes (O). The better focused the question is, the more relevant and specific the search for evidence will be.

Acquisition of evidence
The ability to "acquire" evidence, also called information retrieval, in a timely manner may improve healthcare. Unfortunately, doctors may be led astray when acquiring information because of difficulties in selecting best articles, or because individual trials may be flawed or their outcomes may not be fully representative.

Research conflicts on the ability of end-users of MEDLINE. One study found that users were almost as likely to misinterpret articles found as correctly interpret them. A second study that was in vitro found that searching for evidence was helpful.

One proposed structure for a comprehensive evidence search is the 5S search strategy, which starts with the search of "summaries" (textbooks).

Appraisal of evidence
To "Appraise" the quality of the evidence in a study is very important, as one third of the results of even the most visible medical research is eventually either attenuated or refuted. There are many reasons for this ; two of the most common being publication bias and conflict of interest (see article on Medical ethics). These two problems interact, as conflict of interest often leads to publication bias. Complicating the appraisal process further, many (if not all) studies contain some design flaws, and even when there are no clear methodological flaws, any outcome of a test that is evaluated by statistical test has a margin of error: this means that some positive outcomes will be "false positives".

It is also very common to read only the abstract of an article. However, many abstracts contain errors compared to the body of the article. These are usually errors of omission rather than contradiction, but there is a danger that positive findings are over-emphasised in the abstract.

"Levels of evidence" are used for describing the strength of a research study. Guides exist for assessing a body of literature; these are discussed below, Assessing a body of research evidence.

Publication bias Publication bias is "the influence of study results on the chances of publication [in academic journals] and the tendency of investigators, reviewers, and editors to submit or accept manuscripts for publication based on the direction or strength of the study findings. Publication bias has an impact on the interpretation of clinical trials and meta-analyses. Bias can be minimized by insistence by editors on high-quality research, thorough literature reviews, acknowledgment of conflicts of interest, modification of peer review practices, etc."
 * Common threats to validity

Conflict of interest

The presence of a conflict of interest has many effects in medical publishing.

Statistical analysis Common problems include small sample sizes in subgroup analyses, problems of "multiple comparisons" when several outcomes are being assessed, and biasing of study populations by selection criteria.

Problems may arise from interim analyses of randomized controlled trials which may lead to exaggerated estimations of treatment effects.

The outcome of a study is often summarized by a "P-value" that expresses the likelihood that an observed difference between treatment groups reflects a true difference in treatment effectiveness; the P value is a statistical calculation of the chance that the observed difference reflects the chance outcome of random sampling. Some have argued that focusing on P values neglects other important sources of knowledge and information that should properly be used to assess the likely efficacy of a treatment In particular, some argue that the P-value should be interpreted in light of how plausible is the hypothesis based on the totality of prior research and physiologic knowledge.

Application
It is important to "apply" the best practices found to the correct situation. One common problem in applying evidence is that both patients and healthcare professionals often have difficulties with health numeracy and probabilistic reasoning. Another problem is successful clinical reasoning at the bedside. Successful reasoning is associated with the use of pattern matching and recognizing clinical findings that are "pivot" or specific to certain diseases. A third problem is to identify exactly which patients will benefit from the new practices. Extrapolating study results to the wrong patient populations (over-generalization)  and not applying study results to the correct population (under-utilization)  can both increase adverse outcomes.

The problem of over-generalizating study results may be more common among specialist physicians. Two studies found specialists were more likely to adopt cyclooxygenase 2 inhibitor drugs before the drug rofecoxib was withdrawn by its manufacturers because of its unanticipated adverse effects. One of the studies went on to state:
 * "using COX-2s as a model for physician adoption of new therapeutic agents, specialists were more likely to use these new medications for patients likely to benefit but were also significantly more likely to use them for patients without a clear indication".

Similarly, orthopedists provide more expensive care for back pain, but without measurably increased benefit compared to other types of practitioners. Some of the reason that subspecialists may be more likely to adopt inadequately studied innovations is that they read from a spectrum of journals that have less consistent quality. Articles from specialty journals have been noted to persist in supporting claims in the medical literature that have been refuted.

The problem of under-utilizing study results may be more common when physicians are practising outside their expertise. For example, specialist physicians are less likely to under-utilize specialty care, while primary care physicians are less likely to under-utilize  preventive care.

Metrics used in EBM

 * Diagnosis


 * Sensitivity and specificity
 * Likelihood ratios


 * Causation
 * Relative measures


 * Odds ratio
 * Relative risk ratio
 * Relative risk reduction


 * Absolute measures


 * Absolute risk reduction
 * Number needed to treat (NNT)
 * Number needed to screen
 * Number needed to harm


 * Health policy


 * Cost per year of life saved
 * Cost of Preventing an Event (COPE) . For example, to prevent a major vascular event n a high-risk adult, the number needed to treat is 19, the number of years of treatment are 5, and the daily cost of the generic drug is 68 cents. The COPE is 19 * 5 * ( 365 * .68) which equals $23,579 in the United States.
 * Years (or months or days) of life saved. "A gain in life expectancy of a month from a preventive intervention targeted at populations at average risk and a gain of a year  from a preventive intervention targeted at populations at elevated risk can both be considered large."

Original research studies: levels of evidence
'Levels of evidence' were first proposed in 1986 to create a framework for judging the strength of research. An example of levels of evidence is:
 * Level 1 - Randomized controlled trials
 * Level 2 - Cohort studies
 * Level 3 - Case control studies
 * Level 4 - Case series and case reports

These levels of evidence have been criticized as only representing one of two views of medical science. Vandenbroucke proposes that the two views of medical research are:
 * 1) Evaluation of interventions. In this view, the randomized controlled trial is most important.
 * 2) Discovery and explanation. In this view, the randomized controlled trial is not supreme and observational studies, subgroup analyses, and secondary analyses are important.

In practice, randomized controlled trials are available to support only 21% to 53% of principle therapeutic decisions. Due to this, evidence-based medicine has evolved to accept lesser levels of evidence when randomized controlled trials are not available.

The Grade Working Group recognizes that assessing medical evidence involves attention to other dimensions than study design, and expanded upon the levels of evidence in 2004. These dimensions include: study design; study quality; consistency of study results; and directness (e.g. how close does the research study mirror clinical practice)

Assessing a body of research evidence
Several systems have been proposed to assess whether a body of research evidence establishes causality.

In the 1880's Koch proposed postulates (Koch's Postulates) that suggest association. However, even though the postulated focus on establishing casuality among infections diseases, examples have emerged that violate his postulates.

The Bradford Hill criteria emerged were proposed in 1965:
 * Strength or magnitude of association?
 * Consistency of association across studies?
 * Specificity of association?
 * Temporality of association?
 * Plausibility based on biological knowledge?
 * Biological gradient: or dose-response relationship?
 * Coherence? Does the proposed association explain other observations?
 * Experimental evidence?
 * Analogy?

More explicit methods have been proposed by groups such as the U.S. Preventive Services Task Force (USPSTF) (see yellow text box), the American College of Chest Physicians and the Grade Working Group The Grade Working Group combines the quality of research studies with their clinical relevance. For example, this allows comparing a randomized control trial in a population that is not exactly like the population of interest versus a cohort study that is from a relevant population. The concept that there are multiple dimensions in assessing evidence has been expanded to a proposal that assessing evidence is not a linear process, but is circular.

The legal system has focused on assessing the quality of the expert who interprets evidence for the court rather than the quality of the evidence itself. Methods to assess expertise have included the Frye test and the Daubert standard. Occasionally, courts have used "Daubert panels" to assess evidence in large tort claims.

Systematic review
A systematic review is a summary of healthcare research that involves a thorough literature search and critical appraisal of individual studies to identify the valid and applicable evidence. It often, but not always, uses appropriate techniques (meta-analysis) to combine studies, and may grade the quality of the particular pieces of evidence according to the methodology used, and according to strengths or weaknesses of the study design. While many systematic reviews are based on an explicit quantitative meta-analysis of available data, there are also qualitative reviews which nonetheless adhere to the standards for gathering, analyzing and reporting evidence.

Clinical practice guidelines
Clinical practice guidelines are defined as "Directions or principles presenting current or future rules of policy for assisting health care practitioners in patient care decisions regarding diagnosis, therapy, or related clinical circumstances. The guidelines may be developed by government agencies at any level, institutions, professional societies, governing boards, or by the convening of expert panels. The guidelines form a basis for the evaluation of all aspects of health care and delivery."

Incorporating evidence into clinical care
see also Teaching evidence-based medicine

Medical informatics
Practicing clinicians cite the lack of time for keeping up with emerging medical evidence that may change clinical practice. Medical informatics is an essential adjunct to EBM, and focuses on creating tools to access and apply the best evidence for making decisions about patient care.

Before practicing EBM, informaticians (or informationists) must be familiar with medical journals, literature databases, medical textbooks, practice guidelines, and the growing number of other evidence-based resources, like the Cochrane Database of Systematic Reviews and Clinical Evidence.

Similarly, for practicing medical informatics properly, it is essential to have an understanding of EBM, including the ability to phrase an answerable question, locate and retrieve the best evidence, and critically appraise and apply it.

Criticisms of EBM


There are a number of criticisms of EBM. Most generally, EBM has been criticized as an attempt to define knowledge in medicine in the same way that was done unsuccessfully by the logical positivists in epistemology, "trying to establish a secure foundation for scientific knowledge based only on observed facts" and not recognizing the fallible nature of knowledge in general. The problem of relying on empiric evidence as a foundation for knowledge was recognized over 100 years ago and is known as the "Problem of Induction" or "Hume's Problem".

A general problem with EBM is that it seeks to make recommendations for treatment that (on balance) are likely to provide the best treatment for most patients. However what is the best treatment for most patients is not necessarily the best treatment for a particular individual patient. The causes of disease, and the patient responses to treatment all vary considerably, and are affected for example by the individual's genetic make-up, their particular history, and by factors of individual lifestyle. To take these properly into account requires the clinical experience of the treating physician, and over-reliance upon recommendations based upon statistical outcomes of treatments given in a standardised way to large populations may not always lead to the best care for a particular individual.

An early criticism of EBM is that it will be a guise for rationing resources or other goals that are not in the interest of the patient. In 1994, the American Medical Association helped introduce the "Patient Protection Act" in Congress to reduce the power of insurers to use guidelines to deny payment for a medical services. As a possible example, Milliman Care Guidelines state they produce "evidence-based clinical guidelines since 1990". In 2000, an academic pediatrician sued Milliman for using his name as an author on a practice guidelines that he stated were "dangerous"  A similar suit disputing the origin of care decisions at Kaiser has been filed. The outcomes of both suits are not known.
 * Ulterior motives===

Conversely, clinical practice guidelines by the Infectious Disease Society of America are being investigated by Connecticut's attorney general on grounds that the guidelines, which do not recognize a chronic form of Lyme disease, are anticompetitive.

EBM is a set of techniques derived from clinical epidemiology, but a common criticism is that epidemiology can show association but not causation. While clinical epidemiology has its role in inspiring clinical decisions if it is complemented with testable hypotheses on disease, many critics consider that EBM is a form of clinical epidemiology which became so common in health care systems, and imposed such an empiricist bias on medical research, that it has undermined the notion of causal inference in clinical practice. It is argued that it has even become condemnable to use common sense, as was cleverly illustrated in a systematic review of randomized controlled trials studying the effects of parachutes against gravitational challenges (free falls).
 * EBM not recognizing the limits of clinical epidemiology===

Complexity science

Complexity science and chaos theory are proposed as further explaining the nature of medical knowledge and education.