While evidence-based medicine has increasingly broad-based support in health care, it remains difficult to get physicians to actually practice it. Across most domains in medicine, practice has lagged behind knowledge by at least several years. The authors believe that the key tools for closing this gap will be information systems that provide decision support to users at the time they make decisions, which should result in improved quality of care. Furthermore, providers make many errors, and clinical decision support can be useful for finding and preventing such errors. Over the last eight years the authors have implemented and studied the impact of decision support across a broad array of domains and have found a number of common elements important to success. The goal of this report is to discuss these lessons learned in the interest of informing the efforts of others working to make the practice of evidence-based medicine a reality.
Delivering outstanding medical care requires providing care that is both high-quality and safe. However, while the knowledge base regarding effective medical therapies continues to improve, the practice of medicine continues to lag behind, and errors are distressingly frequent.1
Regarding the gaps between evidence and practice, Lomas et al.2 evaluated a series of published guidelines and found that it took an average of approximately five years for these guidelines to be adopted into routine practice. Moreover, evidence exists that many guidelines—even those that are broadly accepted—are often not followed.3–7 For example, approximately 50% of eligible patients do not receive beta blockers after myocardial infarction,8 and a recent study found that only 33% of patients had low-density lipoprotein (LDL) cholesterol levels at or below the National Cholesterol Education Program recommendations.5 Of course, in many instances, relevant guidelines are not yet available, but even in these instances, practitioners should consider the evidence if they wish to practice evidence-based medicine, and a core part of practicing evidence-based medicine is considering guidelines when they do exist.
Although we strive to provide the best possible care, many studies within our own institution have identified gaps between optimal and actual practice. For example, in a study designed to assess the appropriateness of antiepileptic drug monitoring, only 27% of antiepileptic drug levels had an appropriate indication and, among these, half were drawn at an inappropriate time.9 Among digoxin levels, only 16% were appropriate in the inpatient setting, and 52% were appropriate in the outpatient setting.10 Of clinical laboratory tests, 28% were ordered too early after a prior test of the same type to be clinically useful.11 For evaluation of hypothyroidism or hyperthyroidism, the initial thyroid test performed was not the thyroid-stimulating hormone level in 52% of instances.12 Only 17% of diabetics who needed eye examinations had them, even after visiting their primary care provider.13 The Centers for Disease Control and Prevention (CDC) guidelines for vancomycin use were not followed 68% of the time.14 Safety also is an issue: in one study, we identified 6.5 adverse drug events per 100 admissions, and 28% were preventable15; for example, many patients received medications to which they had a known allergy. Clearly, there are many opportunities for improvement.
We believe that decision support delivered using information systems, ideally with the electronic medical record as the platform, will finally provide decision makers with tools making it possible to achieve large gains in performance, narrow gaps between knowledge and practice, and improve safety.16,17 Recent reviews have suggested that decision support can improve performance, although it has not always been effective.18,19 These reviews have summarized the evidence that computerized decision support works, in part, based on evidence domain. While this perspective has been very useful and has suggested, for example, that decision support focusing on preventive reminders and drug doses has been more effective than decision support targeting assistance regarding diagnosis, it does not tell one how best to deliver it.
In all the areas discussed above, we have attempted to intervene with decision support to improve care with some successes20,21 and many partial or complete failures.22–24 For the purposes of this report, we consider decision support to include passive and active referential information as well as reminders, alerts, and guidelines. Many others also have evaluated the impact of decision support,19,20 and we are not attempting to provide a comprehensive summary of how decision support can improve care but rather to provide our perspective on what worked and what did not.19 Thus, the goal of this report is to present generic lessons from our experiences that may be useful to others, including informaticians, systems developers, and health care organizations.
Brigham and Women's Hospital (BWH) is a 720-bed tertiary care hospital. The hospital has an integrated hospital information system, accessed via networked desktop personal computers, that provides clinical, administrative, and financial functions.25,26 A physician order entry application was implemented initially in 1993.27,28 Physicians enter all patient orders into this application, with the majority being entered in coded form. The information system in general, and the physician order entry system in particular, delivers patient-specific decision support to clinicians in real time. Most active decision support to date has focused on drugs,29 laboratory testing,30,22 and radiology procedures.24 In addition, a wide array of information is available online for physicians to consult, including literature searching, Scientific American Medicine, and the Physician's Desk Reference, among others; these applications are used hundreds of times daily. In the ambulatory practices associated with the hospital, we have developed an electronic medical record, which is the main record used in most practices31 and which includes an increasing amount of decision support.32
We currently are in the process of developing additional applications for the Partners network, which includes BWH and Massachusetts General Hospital, several smaller community hospitals, and Partners Community Healthcare, a network of more than 1,000 physicians across the region. These include a Longitudinal Medical Record, which will serve as a network-wide record across the continuum of care and will include ambulatory order entry, an enterprise master patient index, and a clinical data repository.31
Ten Commandments for Effective Clinical Decision Support
1. Speed Is Everything
We have found repeatedly,33 as have others,34 that the speed of an information system is the parameter that users value most. If the decision support is wonderful, but takes too long to appear, it will be useless. When infrastructure problems slow the speed of an application, user satisfaction declines markedly. Our goal is subsecond “screen flips” (the time it takes to transition from one screen to the next), which appears anecdotally to be the threshold that is important to our users. While this may be a difficult standard to achieve, it should be a primary goal.
Evidence supporting this comes, in part, from user surveys regarding computerized physician order entry. In one such survey, we found that the primary determinant of user satisfaction was speed and that this rated much higher than quality improvement aspects.35 In fact, users perceived physician order entry primarily as an efficiency technology,35 even though we found in a formal time–motion study that it took users significantly longer to write orders using the computer than with paper, in part, because many screens were involved.36 Others have had similar results.37 Thus, while the hospital administration and clinical leadership's highest priorities are likely to be costs and quality, the top priority of users will be the speed of the information system.
2. Anticipate Needs and Deliver in Real Time
It is not enough for the information a provider needs to simply be available someplace in the system—applications must anticipate clinician needs and bring information to clinicians at the time they need it (Fig. 1). All health professionals in the United States face increasing time pressure and can ill afford to spend even more time seeking bits of information. Simply making information accessible electronically, while better than nothing, has little effect.38 Systems can serve the critical role of gathering and making associations between pieces of information that clinicians might miss because of the sheer volume of data, for example, emphasizing a low-potassium level in a patient receiving digoxin.
Optimal clinical decision support systems should also have the capability to anticipate the subtle “latent needs” of clinicians in addition to more obvious needs. “Latent needs” are needs that are present but have not been consciously realized. Decision support provided through the computer can fill many such latent needs, for example, notifying the clinician to lower a drug dose when a patient's kidney function worsens.39 Another example of a latent need is a situation that occurs in which one order or piece of information suggests that an action should follow (Fig. 2).40 Overhage et al.40 have referred to these orders or information as corollary orders. As did Overhage et al.,40 our group has also found that displaying suggested orders across a wide range of order types substantially increased the likelihood that the desired action will occur.41
3. Fit into the User's Workflow
Success with alerts, guidelines, and algorithms depends substantially on integrating suggestions with practice.38 We have built a number of “stand-alone” guidelines for a variety of conditions, including sleep apnea, for example. However, use counts have been very low, even for excellent guidelines. The vancomycin guideline mentioned above was available for passive consultation, yet was rarely used; only after bringing the guideline to the user on a single screen at the time the clinician was in the process of ordering vancomycin did we see an impact. Understanding clinician workflow, particularly when designing applications for the outpatient setting, is critical.
4. Little Things Can Make a Big Difference
The point here is that usability matters—a lot. Developers must make it easy for a clinician to “do the right thing.” In the human factors world, usability testing has had a tremendous impact on improving systems,42 and what appear to be nuances can make the difference between success and failure. While it should be obvious that clinical computing systems are no different, usability testing has not necessarily been a routine part of designing them. We have had many experiences in which a minor change in the way screens were designed had a major impact on provider actions. For example, in one application, when the default was set to have clinicians enter the diagnosis as free text rather than from a code list below, a much larger proportion of diagnoses were entered as free text than when we had defaulted to the most likely coded alternative. This, in turn, had down-stream consequences regarding providing decision support, for example, providers did not get reminders about diabetes if diabetes was entered as free text. In a more dramatic example, the Regenstrief group found that displaying computerized reminders suggesting vaccinations for inpatients had no impact when the reminders were easy to ignore, but when the screen flow was altered to make it harder for physicians to ignore the reminder suggestions, they saw a large positive impact.43
5. Recognize that Physicians Will Strongly Resist Stopping
Across a wide array of interventions, we have found that physicians strongly resist suggestions not to carry out an action when we do not offer an alternative, even if the action they are about to carry out is virtually always counterproductive. In a study of decision support regarding abdominal radiography,24 suggestions that no radiograph be ordered at all were accepted only 5% of the time, even though studies ordered when the alerts were overridden yielded almost no useful findings. Similarly, for tests that were clearly redundant with another test of the same kind performed earlier that day, we found that clinicians overrode reminders a third of the time, even though such results were never useful (Fig. 3).44
In counter detailing about drugs, we have found also that if clinicians have strong beliefs about a medication, and either no alternative or an unpalatable alternative is offered, clinicians routinely override suggestions not to order the ketorolac original medication. One example was for intravenous ketorolac, which clinicians believed was more effective than oral nonsteroidals for pain relief, despite little evidence to support this.
Our general approach has been to allow clinicians to exercise their own judgment and override nearly all reminders and to “get past” most guidelines. However, situations arise in which this may not be desirable. For example, we noted several years ago that we were using approximately $600,000 of human growth hormone per year, even though we are not a pediatric hospital. A drug utilization evaluation found that most of this was being used by surgeons in the intensive care unit who believed that giving daily growth hormone to patients with “failure to wean” from the ventilator helped them get off the ventilator more rapidly. However, an examination of the evidence found that there was only one abstract to support this belief. Subsequently, the Pharmacy and Therapeutics Committee formed a group that developed a single-screen, computerized guideline for this medication. Because the medication is never needed urgently and the costs involved were so large ($180 per dose), the guideline stated that one of the indications on the screen had to be present, or the user would have to apply in writing to the chairman of the Pharmacy and Therapeutics Committee (Fig. 4). This had a large impact, and utilization decreased to about one third the former level. However, over time, utilization began again to climb. An evaluation by pharmacy found that users had begun to “game the system” or to state that one of the indications under the guideline was present when, in fact, it was not. However, because of order entry, which requires physicians to log in, it was possible to identify the ordering clinician involved. Targeted discussions with these physicians quickly resulted in utilization returning to previous levels. Thus, in such situations, ongoing monitoring may be required.
6. Changing Direction Is Easier than Stopping
In many situations, we have found that the computer is an enormously powerful tool for getting physicians to “change direction,” and enormous savings can result. Changing physician behavior in this way is especially effective when the issue at hand is one attribute of an order the physician probably does not have strong feelings about, such as the dose, route, or frequency of a medication or the views in a radiographic study. For example, the Pharmacy and Therapeutics Committee at our institution determined that intravenous ondansetron was as effective if given at a lower dose three times daily rather than the previous routine of four times daily after doing an equivalence study and surveying patients (unpublished data). Simply changing the default dose and frequency on the ordering screen had a dramatic effect on physician behavior. In the four weeks before changing the default frequency, 89.7% of the orders for ondansetron were for four times daily and 5.9% were for three times daily. In the four weeks after changing the default frequency, 13.7% of ondansetron orders were for four times daily, and 75.3% were for three times daily. The cost savings associated with this change were approximately $250,000 in the first year after intervention alone.41
Another example of decision support successfully changing physician direction involves abdominal radiograph orders (Fig. 5). For abdominal films, for certain indications, only one view is needed, while for other indications both upright and flat films are needed.24 In one study, the computer order entry system asked the clinician to provide his or her indication, then suggested the appropriate views if the choice of views did not match the indication. Clinicians accepted these suggestions almost half the time.24
7. Simple Interventions Work Best
Our experience has been that if you cannot fit a guideline on a single screen, clinicians will not be happy about using it. Writers of paper-based guidelines do not have such constraints and tend to go on at some length. While there is some utility in having the complete backup guideline as reference material, as McDonald and Overhage have pointed out,45 such guidelines are not usable online without modification, especially during the provision of routine care. An example would be “use aspirin in patients status-post myocardial infarction unless otherwise contraindicated.” We have built a number of computerized guidelines,14,22,24,46 and have repeatedly found this issue to be important. In one study, when clinicians ordered several frequently overused rheumatologic tests on inpatients, we developed a Bayesian program that asked clinicians to enter probabilities of disease.46 However, clinicians needed to provide several pieces of information to get to the key decision support. Eleven percent of intervention orders were cancelled versus less than 1% in the controls. But, we found that in instances in which they cancelled, physicians never got to the decision support piece of the intervention—all the effect appeared to be due to a barrier effect.
Another guideline we evaluated was an adaptation of the CDC guideline for prescribing vancomycin.14 Getting the guideline to fit on a single screen required substantial condensation and simplification, but we believe this was a key factor in convincing clinicians to use it. In a randomized, controlled trial evaluating the reduction of vancomycin use using computer guidelines, intervention physicians wrote 32% fewer orders than control physicians and had 28% fewer patients for whom a vancomycin order was initiated or renewed. The duration of vancomycin therapy was also 36% lower for patients of the intervention physicians than for patients of the control physicians.14 Based on a small number of chart reviews, these decreases were clinically appropriate, although a significant amount of inappropriate use persisted.
8. Ask for Additional Information Only When You Really Need It
To provide advanced decision support—and especially to implement guidelines—one frequently needs data that are not already in the information system and can be obtained only from the clinician. Examples are the weight of a patient about to be prescribed a nephrotoxic drug and the symptom status of a patient with congestive heart failure. Some situations are perceived as sufficiently risky by clinicians that they will be willing to provide a number of pieces of information, for example, if they are prescribing chemotherapy. However, in many others, the clinician may not want to get a piece of information that is not immediately at hand, such as the weight or whether a young woman could be pregnant. In a trial we performed of guided medication dosing for inpatients with renal insufficiency,39 we needed to have the patient weight to make suggestions (other variables including age, gender, and creatinine level were all available). Getting consensus that it was acceptable to demand this was in many ways the most difficult part of the project. In approximately one third of patients, no weight was entered at first, although this proportion eventually fell. Overall, our experience has been that the likelihood of success in implementing a computerized guideline is inversely proportional to the number of extra data elements needed. Plans must be made to cover situations in which the provider does not give a piece of information for some reason, and, over time, it makes sense to try to make sure that key pieces of data (such as weight) are collected as part of routine care.
9. Monitor Impact, Get Feedback, and Respond
If reminders are to be delivered, there should be a reasonable probability that they will be followed, although this probability should likely vary substantially according to the type of reminder. For strongly “action-oriented” suggestions, we try to have clinicians respond positively more than approximately 60% of the time—this is a threshold we reached empirically and somewhat arbitrarily. For other situations, for example when suggesting that a specific type of utilization be avoided, lower thresholds may be reasonable. It is clear that if numerous suggestions are delivered that rarely are useful, clinicians will simply extinguish and miss even very important suggestions.47 For example, with drug–drug interactions in pharmacies, so many notifications are delivered to pharmacists with similar priority levels that pharmacists have been found to override a third of reminders about a life-threatening interaction.48 Therefore, we carefully evaluate and prune our knowledge base. For drug–drug interactions, we have downgraded the severity level for many. Achieving the right balance between over-and underalerting is difficult and should be an important topic for further research.
Many of our interventions did not turn out as expected with lower impact for a variety of reasons. We have learned that it is essential to track impact early when supplying decision support by assessing how often suggestions are followed, and then make appropriate midcourse corrections. For instance, when presenting reminders to use aspirin in patients with coronary artery disease, we found that a major reason for failure to comply was that patients were already receiving warfarin, which we should have anticipated but did not.
10. Manage and Maintain Your Knowledge-based Systems
Maintaining the knowledge within the system and managing the individual pieces of the system are critical to successful delivery of decision support. We have found it useful to track the frequency of alerts and reminders and user responses and have someone, usually in information systems, evaluate the resulting reports on a regular basis. Thus, if it becomes clear that a drug–drug interaction is suddenly coming up tens of times per day yet is always being overridden, an appropriate corrective action can be taken. The effort required to monitor and address issues in such systems is considerable and is easy to underestimate. It is also critical to keep up with the pace of change of medical knowledge. We have attempted to assign each area of decision support to an individual, and require the individual to assess their assignment periodically to ensure that the knowledge base remains applicable.
The costs of medical care continue to rise, and society is no longer willing to give the medical profession a blank check. An early consequence of rising costs was the increasing penetration of managed care, which has attempted to minimize the provision of unnecessary care, while providing the care that is important. However, even in most managed care situations, many unnecessary things get done, while other effective interventions do not get carried out in part because the number of potentially beneficial things to accomplish is so large that physicians cannot effectively keep track of them all. Information systems represent a critical and underused tool for managing utilization and improving both efficiency and quality.1
Especially when combined with an electronic record, decision support is one of the most potent ways to change physician behavior. Some find this approach threatening and fear the loss of physician autonomy.49 While such fears may not be entirely unfounded, we think the computer can essentially provide a “better cockpit” for clinicians, which can help them avoid errors, be more thorough, and stay closer to the findings of the evidence base. A one-dimensional scale of “degrees of computerization” has been suggested: (1) The computer offers no assistance; humans must do it all. (2) The computer offers a complete set of action alternatives, and (3) narrows the selection down to a few, (4) suggests one, and (5) executes that selection if the human approves, or (6) allows the human a restricted time to veto before automatic execution, or (7) executes automatically, then necessarily informs the human, or (8) informs him or her after execution only if he or she asks, or (9) informs him or her after execution only if it, the computer, decides to. (10) The computer decides everything and acts autonomously, ignoring the human.50
Most of medicine in the United States is still at level 1, which we believe is far from optimal. We have found that clinical decision support is most likely to be accepted if it approaches level 5, and while there are comparatively few situations in which this can be achieved, they are very high yield. For instance, there are many “consequent actions,” in which one action suggests that another is very likely indicated, for example, ordering aminoglycoside levels after ordering aminoglycosides, and physicians are much more likely to carry out such actions when appropriate suggestions are made.40 In addition, there are many straightforward situations in which it makes sense to make it as easy as possible for a clinician to carry out an action, for example, ordering a mammogram for a 55-year-old woman. In other instances, level 2 or 3 may be optimal.
Several important, recent trials from England have identified no benefit with computer-based guidelines in chronic diseases in primary care; the studies targeted hypertension, angina, and asthma.51–,53 In the most recent of these evaluations, the investigators concluded that key issues were the timing of triggers, ease of use of the system, and helpfulness of the content.53 These fit with our comments regarding workflow and making messages highly directive. Our own results with chronic diseases and complex guidelines have been largely similar.38 These results are disappointing because chronic disease management is so important; perhaps the biggest challenge is identifying accurately where the patient is in their care so that helpful suggestions can be made. This remains a vitally important frontier in decision support.
A number of features we believed would be valued have either caused major problems, or received relatively little use, or both. Regarding ordering, a major issue was implementation of “free text” ordering—orders that were written simply using free text. While we developed a parser that worked reasonably well, orders written using this mechanism often proved difficult to categorize according to type. As a consequence, relevant decision support was not delivered. Moreover, in these instances, the computer did not know to which ancillary area to send such orders and they thus ended up in a generic pile. However, the clinician had the impression that his or her order had been communicated. A related feature with similar problems was personal order sets. Because of the way the application was designed, such order sets ended up bypassing most decision support. To our surprise, neither free text orders nor personal order sets were used frequently or were highly valued.35 As a result, we no longer allow either free text orders or personal order sets, although divisions and departments can develop order sets.
Another tool that we thought would be highly valued was links to referential information supporting the decision support. Although this has been used to date less often than we would have hoped, users do rate it as important to them, and part of the reason these get relatively little use may be that the information we have displayed has, as of yet, been quite limited. We currently are building a tool to allow real-time links to Web-based evidence, which should be much more comprehensive and may be more highly valued.
Another surprise came in delivering reminders about redundant tests.22 We found that these reminders were effective when delivered, but that the overall impact of the decision support was much lower than expected, because the system was often bypassed by test orders going directly to the laboratory, and because tests ordered using order sets were not put through these screens.
Limitations and Conclusions
Our findings are limited in that many of our experiences to date come from a single, large, tertiary care institution and one large integrated delivery system, so that issues in other types of institutions may vary. Further, a major part of our user group represents residents in training, although many attending physicians use the system on a regular basis.
We conclude that decision support provided using information systems represents a powerful tool for improving clinical care and patient outcomes, and we are hopeful that these thoughts will be useful to others building such systems. Moving toward more evidence-based practice has the potential to improve quality and safety while simultaneously reducing costs. We believe that implementation of computerized decision support through electronic medical records will be the key to actually accomplishing this. However, much remains to be learned about how to best influence physician behavior using decision support—especially around implementing complex guidelines—and this is likely to become an increasingly important area of research as we struggle to provide higher-quality care at lower cost.