The quality and safety of health care are under increasing scrutiny. Recent studies suggest that medical errors, practice variability, and guideline noncompliance are common, and that cognitive error contributes significantly to delayed or incorrect diagnoses. These observations have increased interest in understanding decision-making psychology.

Many nonrational (i.e., not purely based in statistics) cognitive factors influence medical decisions and may lead to error. The most well-studied include heuristics, preferences for certainty, overconfidence, affective (emotional) influences, memory distortions, bias, and social forces such as fairness or blame.

Although the extent to which such cognitive processes play a role in anesthesia practice is unknown, anesthesia care frequently requires rapid, complex decisions that are most susceptible to decision errors. This review will examine current theories of human decision behavior, identify effects of nonrational cognitive processes on decision making, describe characteristic anesthesia decisions in this context, and suggest strategies to improve decision making.

THE quality and safety of health care are under increasing scrutiny. Studies have suggested that health care is plagued with errors,1  unexplained practice variability,2–4  and guideline noncompliance.5,6  These observations have led to increased interest in understanding decision-making cognitive processes and improving educational strategies for teaching decision-making skills.7,8 * Most anesthesiology journals have not described decision-making processes in a systematic way, and we understand little about how decisions may be improved or harmed by cognitive factors.

The incidence of diagnostic error varies across physician specialty, with rates ranging from 2 to 12% in diagnostic radiology and pathology, 12 to 15% in emergency medicine,9  and up to 50% with respect to diagnosing the cause of death.10  Although the incidence of erroneous decision making in anesthesiology is not known, reports from the American Society of Anesthesiologists closed claims registry suggest that more than half of diagnosis-related adverse events in obstetric anesthesia were related to a delay in diagnosis or treatment.11  Most decision researchers believe that specialties characterized by a high degree of time pressure, data uncertainty, stress, and distractors may have an even greater incidence of errors.12  In some estimates, more than two thirds of missed or delayed diagnoses are caused in part by cognitive errors in decision making.13 

In principle, medical decision making should be relatively straightforward. A constellation of clinical findings should generate a limited differential of known clinical conditions, ordered by their probability of occurrence. Diagnostic tests or responses to empiric therapy would then refine the list until only a few candidates exist with (usually) a clear favorite.

Abundant evidence, however, suggests that real-world medical decision making is beset with variability and complexity. Physicians often fail to agree on the interpretation of diagnostic test results,14–16  are inconsistent in their approach to management,17–19  and arrive at different diagnoses in the presence of identical information.13  Even for clinical conditions with a widely accepted theoretical framework and established diagnostic and therapeutic strategies, a startling amount of unexplained practice variability exists.20 

Noncompliance with evidence-based guidelines developed by expert panels is high,5  further highlighting the need to understand physician decision making. Noncompliance observed in simulated preoperative evaluation by anesthesiology trainees and experts shows the need to assess decision behavior in addition to medical knowledge.21,22 

Although characteristics of human decision behavior have been studied in economics, business, and psychology, the extent to which anesthesia care is affected by cognitive decision errors is not known. It is also unclear whether key perioperative anesthesia decisions may be improved by application of human decision research from outside disciplines. Importantly, nonrational decision factors should not be viewed as uniformly good or bad.

This review has three sections. First, this article highlights leading models of decision making and explores bias, heuristic, and nonrational cognitive processes that impact them (fig. 1). Then, to illustrate how such nonrational factors may affect anesthesia practice, we will analyze several anesthesia-related decisions. Finally, we will describe strategies to help recognize and/or recover from decision-making errors.

Fig. 1.

Influences on decision making and diagnostic error. A variety of nonrational factors (i.e., factors not based purely in statistics or logic) influence decisions; these factors are themselves neither good nor bad. This figure highlights factors discussed in this review but is not comprehensive. Importantly, decisions may also utilize rational processes. This graphic is not intended to compare the proportion of cognitive effort or time that is rational compared with nonrational.

Fig. 1.

Influences on decision making and diagnostic error. A variety of nonrational factors (i.e., factors not based purely in statistics or logic) influence decisions; these factors are themselves neither good nor bad. This figure highlights factors discussed in this review but is not comprehensive. Importantly, decisions may also utilize rational processes. This graphic is not intended to compare the proportion of cognitive effort or time that is rational compared with nonrational.

Close modal

Expected Utility

The longest lived and most widely accepted formal model of rational decision making is termed “expected utility” (EU).23,24  Developed in the 17th century, the EU model argues that humans decide among a set of choices by calculating the expected benefit from each choice (multiplying the probability of the outcome resulting from each choice by the payoff for that outcome) and selecting the option with the highest “expected value” (EV). As an illustration, consider a choice between two gambles involving a coin toss. The first would pay $2 for getting “heads” but nothing for “tails.” The second would pay $1 for heads and $0.50 for tails. The first gamble would thus have an (EV) = 0.5 × $2 + 0.5 × 0 or $1, and the second an EV = 0.5 × $1 + 0.5 × $0.50 or $0.75. According to EU theory, the human should choose the first gamble due to the larger EV. Note that EU theory assumes complete probabilistic knowledge of all potential choices, the payoffs associated with each choice, and a well-organized and consistent ordering of preferences for each payoff.

Because these assumptions rarely apply in most real-world settings, decision researchers rapidly realized that such a model would be unlikely to predict human behavior. In 1978 the economist Herbert Simon won a Nobel prize for proposing that humans instead followed a modified, “bounded” rationality wherein “good enough” solutions replaced optimal ones when the effort of obtaining information was significant, and measurable subgoals took the place of goals more difficult to assess.25 

Because medical decisions often involve incomplete data and outcome uncertainty, EU theory may likewise have limited applicability in a clinical care setting.26,27  Not only are all the possible solutions difficult to identify, but also are payoffs diverse and uncertain, and identifying tradeoffs in choosing the optimum payoff can be difficult. Moreover, physicians and patients may vary in their ordering of preferred outcomes.28 

Bayesian Probability

This decision-making model adapts EU theory to permit new information to change existing odds. Because clinical medicine follows a similar dynamic trajectory, Bayesian approaches have been advocated for medical decision strategies.29  As with EU, a decision begins by constructing a list of initial decision possibilities with their relative likelihoods. Unlike EU, however, new information from diagnostic tests or therapeutic interventions is used to modify the initial probability of each decision possibility. In this way, the pretest probability of a disease and the results of diagnostic testing (or therapy) both affect subsequent assessments of disease probability. By using sequential testing and observing the results of therapy, physicians should eventually arrive at the correct diagnosis or decision.

Because of its similarity to clinical medical management, Bayesian approaches to decision making are frequently cited as “evidence based” and are widely taught in medical school.30  However, Bayesian analysis suffers from many of the same limitations as EU theory. Pretest and posttest probabilities may frequently be elusive for individual patients, test results and responses to therapy may be difficult to interpret, likelihood ratios may be unavailable, and patient preferences may differ.

Formalized Pattern-matching

This decision model, developed to cope with limited statistical information, is the classic “case conference” approach taught in medical school, and leverages the remarkable human ability to identify by pattern-matching.31,32  In this approach, physicians begin by aggregating clinical observations into groups to reduce their number. As an example, prolonged hypotension, lactic acidosis, and hepatic failure may be aggregated into a composite “feature” of the case (inadequate end-organ perfusion). Then, pattern-matching is used to identify features of the case, or “pivots” that stand out, are unique, or do not fit. In this example, a pivot might be “if hypoperfusion was sufficient to cause hepatic failure, why was renal function not affected?” Because a pivot is often unique and may itself trigger diagnostic possibilities, the clinician then uses pivots to temporarily ignore nonessential features, and construct a list of diagnostic possibilities (causes of primary liver failure). The list of possibilities is then pruned by identifying incompatible features (“because of the rapid onset, cirrhosis is unlikely”), diagnostic testing/response to initial therapy, and by further pattern-matching (“in most cases of hepatic artery thrombosis, pain is common”). Candidate diagnoses are compared and those that are poor matches are often dropped.

Although surprisingly effective, the “pattern-matching” approach makes extensive use of cognitive shortcuts in place of statistical logic. Pattern-matching is heavily dependent on a mental pattern library that may vary in extensiveness with clinical experience and differs between individuals. In addition, factors other than frequency of occurrence may affect physician estimates of likelihood. By relegating pretest (and posttest) probabilities to a minor role, the process of selecting from the final, pruned list of choices thus becomes statistically vulnerable to error.

Heuristics

Although Simon’s “bounded rationality” explained some human behavior that failed to follow EU principles, many other examples of human decisions that systematically deviated from EU theory soon surfaced. In 1982 the psychologist Daniel Kahneman won a Nobel prize for the systematic identification and characterization of human decision behaviors that (under certain conditions) violated rational EU doctrine. Kahneman33  theorized that these decision behaviors represented cognitive “shortcuts” used preferentially by humans to reduce the cognitive cost of decision making. He termed these shortcuts heuristics.

Heuristics are used frequently in medicine with the goal of making decisions more quickly or efficiently, when attempting to solve very complex problems, or when complete information is not available.34,35  An anesthesia example might be “I always check a preoperative potassium (K+) level in patients on hemodialysis and treat if the K+ is greater than 5.5 meq/l.” Note that this heuristic simplifies preoperative evaluation by eliminating the need to individually determine which patients need a preoperative K+ check or to assess which K+ levels are appropriate for which procedures.

Similarly, a mild, transient fall in blood pressure with induction may not prompt an exhaustive differential diagnostic evaluation. Most anesthesiologists would instead agree that a heuristic of initially attributing the hypotension to anesthetic induction appropriately preserves attention for other important tasks. Decisions in anesthesia practice are not only complex and often rooted in uncertainty, but must frequently be made in stressful conditions, under time pressure, and with high stakes. For these reasons, heuristics may be useful to anesthesiologists by allowing attention to be effectively allocated and distributed.

Although heuristics are often effective at finding reasonable solutions with reduced decision effort, they may not always lead to optimal solutions, and can be misapplied or fooled, depending on contextual features. In a classic example, Tversky and Kahneman36  described a meek, tidy man who loves books and asked whether he was more likely to be a salesman or librarian. Both undergraduates and graduates diagnosed him as a librarian, ignoring the statistical paucity of male librarians relative to male salesmen. In the induction/hypotension example above, if the anesthesiologist fails to notice a developing rash or bronchospasm related to anaphylaxis, the mental shortcut of attributing hypotension to anesthetic induction likewise can fail.

Three heuristics identified by Kahneman can lead to neglect of baseline incidence information in medicine. The first is demonstrated with Kahneman’s librarian example. Termed the “representativeness” heuristic, this strategy triages diagnostic decisions based on resemblance to a mental model rather than statistical probability. In addition to misdiagnosis by the sheer strength of the pattern match (as in the librarian example), representativeness can also cause diagnosticians to ignore probabilistic logic:

Example 1:37 

Which statement is more likely?

  1. Mr. F has had one or more heart attacks.

  2. Mr. F is over 55 yr old, and has had one or more heart attacks.

Answer 2 is clearly a smaller subset of answer 1, so statistically answer 1 is correct. But answer 2 more closely matches the mental image of a patient with a heart attack, and was chosen more frequently in studies of college and graduate students. An example of such a question in anesthesia practice might be:

Which is more common?

  1. An acute decrease in end-tidal carbon dioxide (mmHg) on anesthesia induction.

  2. A known deep venous thrombosis and acute decrease in end-tidal carbon dioxide (mmHg) on anesthesia induction.

The preference for using the representativeness heuristic over statistical reasoning is sufficiently strong such that humans will use it even when the statistically correct choice is readily apparent, as in Example 238 :

A group is composed of 70 engineers and 30 lawyers. One member of the group is 30 yr old, married, with no children. He is skilled and motivated, and is likely to be quite successful in his field. He is well liked by his colleagues. What is the likelihood that he is an engineer?

Although the first sentence provides the correct answer, approximately 50% of college and graduate students chose a number other than 70%. For anesthesiologists, an example might be:

Mr. Smith is obese and has hyperlipidemia, hypertension, and a family history of heart disease. Is his likelihood of a perioperative major adverse cardiac event higher than normal?

(Note that none of the above descriptors are independently predictive of perioperative major adverse cardiac event.)39,40 

The following vignette demonstrates how pattern-matching heuristics can lead to diagnostic error in anesthesia practice:

A gravida-5, para-4 postpartum woman gets out of bed on postoperative day 1 after her third cesarean section and becomes acutely hypotensive with shortness of breath. She is morbidly obese, has 1+ pitting edema on exam, and has refused her last three doses of subcutaneous heparin (including one that morning). She rapidly loses consciousness and progresses to pulseless electrical activity. She is given tissue plasminogen activator for a presumptive diagnosis of pulmonary embolus.

Although many anesthesiologists might choose pulmonary embolus as a more likely diagnosis than postpartum bleeding statistically postpartum bleeding is 100 times more likely, even given the patient’s pattern-matching features (obesity and heparin refusal).41,42  As well, her symptoms are consistent with both pulmonary embolus and significant blood loss. Of course, any particular patient may have any outcome, regardless of statistical estimates.

A second commonly applied heuristic in medicine43  was named the “availability” heuristic by Kahneman and describes the tendency to equate the memorableness of an event with its likelihood of occurrence. Because a shark attack is vivid and memorable, for example, its occurrence may be judged more common than statistically predicted. In medicine, numerous factors may affect the memorableness of an event including a bad outcome, an emotional attachment to a patient, its resemblance to published case reports, and whether it was anticipated (among others). Experts and novices may be affected by the availability heuristic in slightly different ways. For the expert, extensive clinical experience increases the likelihood of personally encountering a rare event. That event then becomes memorable because of its rarity, or because an associated adverse outcome generates a strong emotional impact. After experiencing massive bleeding during mediastinoscopy, for example, an anesthesiologist may insist on crossmatching blood for all subsequent mediastinoscopies despite a low statistical likelihood of bleeding requiring transfusion. For the novice, rare conditions may come easily to mind either because they are interesting, exciting, and novel or because they are easily recalled from classroom learning. For both expert and novice, high-visibility attention (Morbidity and Mortality Conferences, media coverage, or lawsuits), may lead to increased “memorableness” and perceived frequency without any change in real statistical likelihood.

A third heuristic that is less well studied in medicine but may also play a role in medical decision making is the “anchoring” heuristic. Also first characterized by Tversky and Kahneman,36  the anchoring heuristic notes that when humans make an initial estimate of a value or state, and adjust that estimate in response to new information, the initial estimate, or starting point, affects subsequent estimates. As an example, Tversky and Kahneman36  found that asking students to guess the product of 8 × 7 × 6 × 5 × 4 × 3 × 2 × 1 resulted in a higher estimate than asking another group to estimate the product 1 × 2 × 3 × 4 × 5 × 6 × 7 × 8.

In medicine, anchoring bias may lead to failure to adjust initial prognoses based on later information44,45  or failure to modify initial diagnoses due to new events.35  An example of anchoring bias in anesthesia might be a failure to adequately appreciate blood loss during a case in which no bleeding is expected. Another example of anchoring may occur during difficult airway management when repeated instrumentation efforts cause the effectiveness of mask ventilation to deteriorate. Anchoring on the initial “easy mask” conditions may lead to delay in recognition that the patient’s clinical status is changing.

Most physicians are probably familiar with the potential for decisional error due to the availability heuristic. The oft-quoted medical axiom, “When you hear hoofbeats, think of horses, not zebras,” is intended to counterbalance the availability heuristic by reminding the decider to consider pretest probability when ranking diagnostic possibilities. By remembering that common symptoms (hoofbeats) are usually caused by common diseases (horses), physicians may thus prioritize more common (rather than less common) diagnostic possibilities for specific symptoms.

Dual Process Reasoning

A more modern understanding of human decision behavior includes characteristics of both the EU and heuristic-driven decision models.46,47  This hybrid model is called “dual process reasoning,” and asserts that humans may toggle back and forth between an intuitive autonomous decision strategy (also called “type I” processing; fig. 2) and a deliberate statistical/analytic strategy (called “type II”).48–50  Although the factors that govern which processing strategy is used for which decision are incompletely understood, decisions that must be made rapidly and involve nonquantifiable cues, high stakes, and uncertain data are likely to trigger type I-driven decision making whereas decisions that do not involve time pressure and can be analyzed using quantifiable cues or data, are likely to induce type II approaches. Thus, life-saving decisions made under high-stakes conditions and time pressure may be more likely to trigger an intuitive strategy. Cognitive psychologists estimate that we spend a considerable amount of decision-making time in intuitive mode, and although deviations from purely rational processes may occur in either mode, they are less likely to be noticed by the thinking in the subconscious intuitive mode. Intuitive mode may include “hardwired” instincts or processes that are implicitly learned or learned via repetition. Some error prevention strategies discussed in Educational Strategies to Improve Decision Making attempt to cue a toggle between the intuitive mode and the analytic mode. Importantly, we must emphasize that intuitive processes themselves are not errors, and in fact lead to fast and accurate decisions much of the time.

Fig. 2.

Dual process model of reasoning: This illustrates how intuitive processes (type I) and analytical processes (type II) interact to influence diagnostic thinking. Some type I processes go directly to end decisions without any overrides, toggling, or calibration and represent largely unexamined decision-making. Explicit effort allows for toggling between type I and type II processes. Repetition of analytic processes until they become automatic is the basis of skill acquisition. This model does not account for proportion of time spent in, nor superiority of, one process over another. Error may be made in either system at any point, including the starting point (i.e., patterns may be “recognized” incorrectly). (Adapted with permission from Croskerry P, Singhal G, Mamede S: Cognitive debiasing 1: Origins of bias and theory of debiasing. BMJ Qual Saf 2013; Oct; 22(suppl 2):ii58–64. Adaptations are themselves works protected by copyright. So in order to publish this adaptation, authorization must be obtained both from the owner of the copyright in the original work and from the owner of copyright in the translation or adaptation.)

Fig. 2.

Dual process model of reasoning: This illustrates how intuitive processes (type I) and analytical processes (type II) interact to influence diagnostic thinking. Some type I processes go directly to end decisions without any overrides, toggling, or calibration and represent largely unexamined decision-making. Explicit effort allows for toggling between type I and type II processes. Repetition of analytic processes until they become automatic is the basis of skill acquisition. This model does not account for proportion of time spent in, nor superiority of, one process over another. Error may be made in either system at any point, including the starting point (i.e., patterns may be “recognized” incorrectly). (Adapted with permission from Croskerry P, Singhal G, Mamede S: Cognitive debiasing 1: Origins of bias and theory of debiasing. BMJ Qual Saf 2013; Oct; 22(suppl 2):ii58–64. Adaptations are themselves works protected by copyright. So in order to publish this adaptation, authorization must be obtained both from the owner of the copyright in the original work and from the owner of copyright in the translation or adaptation.)

Close modal

Sensemaking

Another modern approach to understanding human decision behavior is to reframe analysis of decisions from discrete choices made by human deciders to a dynamic situational assessment of contextual features. This “sensemaking” approach argues that making a decision first requires an effort to understand the ongoing event, and that such an effort involves initial and evolving impressions, dynamic feedback, and attention shifting to identify and decipher pieces of information. Our early impressions bias and inform what we subsequently give attention to or discount. This regulation of attention, in turn, influences what we think, including the biases and heuristics we apply going forward. Sensemaking has been described as the “complex cognition of the experience of now and then,” resulting in a “too-lateness of human understanding,”51  and can be thought of as a way to link interpretation and choice.52  It strives instead to better understand the context from which the action resulted and thus to facilitate the creation of better conditions for future decision making.53  The related phenomenon of hindsight bias54,55  plays a large role in judgments on the appropriateness of medical care.56  Once given information about the outcome, humans may interpret the preceding events in a different light, viewing outcomes as more predictable and actions as less correct. Although few studies currently link medical decision making and sensemaking, educational strategies may be used to disrupt the normal flow of sensemaking and more closely examine attentional focus, potential biases and heuristics that affect decision quality.

Many other cognitive, emotional, cultural, and environmental factors can also affect how anesthesiologists decide57,58  (table 1). Although an exhaustive description of known decision modifiers is beyond this review, we will focus on decision factors that have been studied in medical domains, including bias, overconfidence, memory error and hindsight, preferences for certainty, framing, loss aversion, and emotion (affect).

Table 1.

Nonrational Cognitive Factors that Influence Decision Making*

Nonrational Cognitive Factors that Influence Decision Making*
Nonrational Cognitive Factors that Influence Decision Making*

Bias and the “Bias Blind Spot”

A decision-making “bias” refers to a systematic preference to exclude certain perspectives on decision possibilities. Biases may result from subconscious influences from life experiences and individual preferences, and can affect medical decision making in both conscious and unconscious ways. An example of bias in anesthesia practice might include the following statement: “I usually do not perform awake fiberoptic intubation on teenagers because it is too traumatizing to the patient.” The perspective reflected in this statement excludes the possibility that awake fiberoptic intubation may not always be traumatizing, and may be tolerated by some teenagers, and also unintentionally prioritizes the avoidance of such trauma over the safety of the patient.

Confirmation bias, which describes the tendency to only seek or recognize information that supports a diagnosis or hypothesis, rather than information that refutes it has been extensively demonstrated in scientific fields such as forensic analysis59  and also in medicine.60  In one study, tanned skin was interpreted more often as jaundice when residents were biased toward liver cancer than stomach cancer.61  It is easy to see how confirmation bias may lead to diagnostic error in medicine. In anesthesia practice, examples of confirmation bias might include focusing on reassuring features of an easy airway, rather than exhaustively seeking evidence to indicate difficulty might be encountered.

A second bias potentially relevant to anesthesia practice is visceral (transference) bias. This bias describes the effect of negative or positive feelings about a patient on diagnostic or management decisions. Examples of visceral bias suggest that patients with unusual sociocultural attributes may get treated differently than “regular” patients.62,63  Examples of visceral bias in anesthesia may include any deviation from usual practice for a “very important person,” such as a different transfusion threshold, or an insistence that the attending personally perform all procedures.

A third bias, termed omission bias, describes a tendency toward inaction (and preserving the status quo) over action even when acting to change the current state is objectively better. For example, studies suggest that people given too much change are more likely to keep the extra money (inaction) than they are to steal an equivalent amount (action).64  Omission bias is well described in medicine65  and usually involves a psychological hurdle that must be overcome. This hurdle may be rooted in fear of harm, either of the patient in the case of a dramatic maneuver gone awry, or harm to the professional reputation if the anesthesiologist is wrong or appears incompetent. An example in anesthesia might be hesitating to initiate a surgical airway when routine airway management fails and the patient is deteriorating. Omission bias is also a possible contributor when team members fail to “speak up” about an important clinical or safety issue.

Although such biases can clearly affect decision behavior, consciously believing oneself to be impervious to bias can also affect decision behavior. Cognitive psychologists call this false sense of invulnerability from bias a “bias blind spot.”66  Interestingly, bias-induced effects on decision making appear correlated to cognitive ability, as the bias blind spot is more prominent among individuals with greater “cognitive sophistication.”67 

Overconfidence

An abundance of research in social science domains suggests that humans are prone to inaccurately high self-assessment with respect to desirable attributes such as ethics, productivity, rationality, intelligence, and health.68  This inappropriate tendency toward overconfidence occurs in both medical and nonmedical fields, in more than 90% of professors,69  and in physicians who are demonstrably inaccurate in self-assessment70  and accuracy of diagnosis.10,71 

Overconfidence may adversely affect decision making in two ways.7  First, unwarranted diagnostic confidence may prevent consideration of other diagnoses and lead to premature closure of the diagnostic process. In addition, overconfidence in technical or diagnostic abilities may delay physicians from calling for help, prevent them from recognizing a medical error, or cause them to choose less safe care strategies. For anesthesiologists, struggling with a difficult airway and being unwilling to call for help may be an example.

Finally, overconfidence may prevent individuals from adopting safety practices developed by others.72  An example in anesthesia might be a reluctance to adopt ultrasound guidance for central line insertion despite considerable evidentiary support and guideline recommendations. Paradoxically, overconfidence may be difficult to modify, as more than half of people believe themselves to be above average at accurate self-assessment,65  and may thus not be able to recognize overconfidence in themselves.

Memory Shifting (Reconstructed Memory Error)

Humans generally accept that memories are imperfect because most have experienced forgetting something altogether (such as a phone number). However, existing evidence suggests that the actual remembered content of events can also be flawed. Fuzzy trace theory suggests that humans store some verbatim information as discrete data values, but store experiences differently, encoding the meaning of the experience rather than the exact events.49,73,74  The act of “remembering” thus requires a process where details that were forgotten or never stored are “filled in” by partially true or even false information created to conform to the present narrative (fig. 3). Memories may therefore be inherently dynamic and vulnerable to disruption.73  An “everyday” illustration of this phenomenon is well described in the legal system, with the unreliability of eyewitness reports.75  An example in anesthesia practice might involve two physicians honestly recalling the details of an emergency event differently, or the certainty of recalling that a patient’s blood pressure remained above a certain threshold during a hemorrhagic event, when in fact it had fallen below that threshold.

Fig. 3.

Memory reconstruction error (also called “retrieval-induced distortion”). Neural processes are dynamic and vulnerable to disruption during reactivation. Information may be lost, and misinformation may be incorporated into memory, which is subsequently reconsolidated and stored in lieu of the original memory. This perpetuates with each memory reactivation, which is a necessary part of recollection. Note that this process is distinct from intentional distortion of the facts.

Fig. 3.

Memory reconstruction error (also called “retrieval-induced distortion”). Neural processes are dynamic and vulnerable to disruption during reactivation. Information may be lost, and misinformation may be incorporated into memory, which is subsequently reconsolidated and stored in lieu of the original memory. This perpetuates with each memory reactivation, which is a necessary part of recollection. Note that this process is distinct from intentional distortion of the facts.

Close modal

Preferences for Certainty

Among the first critiques of the EU model was the finding that EU decision analyses failed to predict human preferences for certainty. The French economist Maurice Allais76  created the first and most famous example:

Consider the following choices with equivalent EVs:

  • A: Eighty-nine percent chance of $1 million, 11% chance of 1 million ($1 million for sure);

  • B: Eighty-nine percent chance of $1 million, 10% chance of $5 million, and 1% chance of nothing.

Most would choose the guaranteed $1 million represented by A. But when given a related set of choices:

  • C: Eighty-nine percent chance of nothing, 11% chance of $1 million and

  • D: Ninety percent chance of nothing, 10% chance of $5 million,

most would choose D.

Allais argued that these preferences contradicted EU theory. He noted that in the first pair (A or B), most choose A even though choice B provides a higher EV. To show the decision was not due to the order of choices or wording, Allais created the second pair of choices (choices with identical EVs, differing only in the degree of certainty) and showed that when greater certainty was not a decision factor people chose the higher EV (D).

The human desire for certainty clearly violates EU-based choice algorithms but reflects deeply held preferences about risk avoidance. An example in anesthesia might be the preoperative potassium decision presented earlier. Statistically, the combination of potassium measurement error, intraoperative tolerance for abnormal potassium levels, and likelihood of a dangerously abnormal preoperative potassium level in a hemodialysis patient makes it unlikely that such a test will have therapeutic implications. Yet many practices routinely check a preoperative potassium level, in part because they value the certainty of knowing the level before inducing anesthesia.

Framing and Loss Aversion

One of the strongest demonstrable biases in human decision making is the preference to behave differently depending on whether the decision is viewed as a gain or a loss. This “framing” effect was first characterized by Tversky and Kahneman77  in 1981, with their now famous “Asian disease” problem. Asked to choose between two treatments for a disease expected to kill 600 people:

  • Treatment A: 200 people will be saved

  • Treatment B: there is a one third probability that 600 people will be saved, and two thirds probability that no people will be saved

a majority of subjects chose A. However, when a second group of subjects were presented with a different formulation of the two choices:

  • Treatment C: 400 people will die

  • Treatment D: There is a one third probability that nobody will die, and two thirds probability that 600 people will die

participants chose D. Note that both choice pairs are identical, but A and B are framed as gains (people saved) whereas C and D are framed as losses (people dead). Kahneman found that nearly all humans will choose the less risky option when the choice was framed as a gain, but the more risky option when the choice was framed as a loss.

Because many choices in medical care can be framed as gains or losses, this preference can easily affect patient and physician choices and contribute to decision diversity.78  Framing may play a role in anesthesia practice in decisions to delay or cancel cases where optimization is imperfect. For a patient with lung cancer and severe symptomatic hypothyroidism, for example, delaying the lobectomy allows the hypothyroid condition to be treated, reducing the risk of life-threatening complications, but increasing the risk of metastasis. Framing the case as a chance to cure cancer versus the chance of an adverse intraoperative cardiorespiratory event may result in different decisions. In addition to viewing choices as losses or gains (equivalence framing), framing may also be used to focus thinking on a subset of data or choices to match an explanation (emphasis framing), as with the obstetric pulmonary embolus versus hemorrhage vignette.

Another powerful framing effect, also first identified by Kahneman and Tversky, is a stronger preference to avoid a loss compared with the desire for a similarly sized gain. This preference, termed “loss aversion,” can cause negative associations from a loss to be twice as powerful as positive associations from an equivalent gain.79  In medicine, losses may take a variety of forms, including physical harm to a patient, perceived loss of reputation if a physician makes an error, and possibly even loss of licensure. In perioperative care, loss aversion may affect decision behavior by causing physicians to modify operative timing in subsequent patients80  or change thresholds for withdrawal of life support.81 

Affect (Emotion)

Most theories of human decision behavior focus on information processing. However, emotional responses to stimuli often occur before conscious analysis and can affect decisions in nonmedical contexts.82  Although the role of emotion in medical decision making is poorly studied, current evidence suggests that at least two emotions may significantly modify medical decisions.

The first of these is anger. Abundant evidence links anger to disruptive behavior by healthcare providers,83  and demonstrates that anger prevents effective communication between perioperative care providers.84  Caregiver perceptions also associate disruptive behavior with medical errors.85 

Regret is also a likely modifier of physician decisions. In a choice context, regret requires two distinct components: a wish that one had chosen differently and self-blame for not having done so.86,87  Clearly, regret-based decision making is nonrational. Not only does it not account for probabilities or payoffs, different people may regret different things, and to variable degrees. Moreover, regret may depend on events that occur after the decision. A person who buys a house, for example, may regret the purchase if a fire destroys the house several months later. Regret can affect medical decisions in two ways. First, physicians experiencing a bad outcome as a result of a specific decision may be less likely to make that same decision in the future.80  In addition, decisions may be influenced by anticipatory regret—the desire to avoid regret related to the consequence or outcome of a decision.88 

As with preferences for certainty, regret is a nonrational but a strong influence on human decision behavior in medicine. One possible reason for use of regret in clinical medicine may be that it allows a readily accessible decision tool for circumstances where adequate statistical information is unavailable. Under such circumstances, decision options may be ordered by their degree of anticipatory regret and the option that results in the least regret chosen.

Feedback Bias

Although incomplete feedback may disadvantage learners by failing to address underlying rationales, absent feedback may also distort decision behavior (fig. 4). Because many anesthesiologists lose contact with their patients after they leave the postanesthesia care unit, feedback on anesthesia-related outcomes such as nausea and vomiting, pain, recall, neuropathy or eye injury from inadequate protection, myocardial infarction, and death is often incomplete. Given the absence of routine feedback, anesthesia caregivers may assume everything went well, and thus fail to adequately calibrate themselves to true event incidences. This “feedback bias” may be one reason why survey-based estimates of recall under anesthesia are considerably lower than those based on directly measuring recall.89 

Fig. 4.

Feedback bias illustration. Timely and specific feedback about both outcome and processes is required for reflection and ongoing practice improvement. The mindset that “no news is good news” contributes to lack of reflection and practice improvement, and reinforces overconfidence.

Fig. 4.

Feedback bias illustration. Timely and specific feedback about both outcome and processes is required for reflection and ongoing practice improvement. The mindset that “no news is good news” contributes to lack of reflection and practice improvement, and reinforces overconfidence.

Close modal

Because so few decisions in medicine have rigorous statistical support, and because so many cognitive and emotional factors may affect human decision behavior, physicians’ choices are rarely purely statistically based. The following example will illustrate how many of the aforementioned factors can influence a familiar series of decisions in airway management. Note that because this vignette represents real-life decision making, the use of nonrational decision factors is not necessarily incorrect or inappropriate.

A sweet and shy 14-yr-old girl presents for a small inguinal hernia repair. She has a history of a severe sore throat and chipped tooth after an emergency appendectomy at an outside hospital, but has no external indicators of a difficult airway and was not told anything about her previous anesthetic. You perform the American Society of Anesthesiologists guideline–recommended 11-point airway exam,90  and in her case, all elements are normal except for a highly arched, narrow palate. Because no specific algorithm exists to convert her “1 out of 11” score into a meaningful prediction of difficult ventilation or intubation, you switch to an intuitive “gut” analysis (dual process). Is the airway truly difficult? It does not look like it to you (representativeness), and she has no medical alert bracelet or letter (confirmation bias). Moreover, you have always been able to intubate even when the laryngoscopic view is not good (overconfidence). However, you recall hearing about a patient just like this one several years ago who needed a tracheostomy when a colleague on call could not intubate or ventilate. It was a big deal at grand rounds (availability). You briefly consider an awake fiberoptic bronchoscopy, but the patient is so young, and you hate to subject her to the unpleasantness of that procedure (visceral bias). You again ask her to open her mouth and feel reassured that, aside from the palate, the rest of the airway exam is normal; (confirmation bias). You proceed with induction, feeling pretty sure you will be able to ventilate at least, and knowing that you can get a glidescope if you need it (overconfidence). As she says goodbye to her parents and you roll back to the operating room, you briefly question your decision, thinking about how bad you will feel if you have an airway catastrophe (regret, loss aversion) and wondering whether you should just use the fiberoptic approach to make sure you do not lose the airway (preference for certainty). The chain of thought is quickly lost as the surgeon reminds you that he has a long list of cases to do today and urges you to go ahead. “It will take you longer to get the airway than it will for me to do the surgery!” he jokes. Also, because you have not done many awake fiberoptic intubations recently, you worry you might appear incompetent if it is not smooth (loss aversion). You proceed to induce anesthesia and attempt mask ventilation. Even though you cannot ventilate easily, you figure that paralysis will improve mask ventilation as it often does (representativeness), so you give muscle relaxant. Your first laryngoscopy reveals a grade 4 view and you attempt external laryngeal manipulation to improve the view. After 10 to 15 s, the nurse asks if you need a hand, but you say nothing, persisting with your laryngoscopy and certain that if you can just move the larynx 1 mm more, you will be able to see (anchoring, overconfidence). The patient starts to desaturate, but you are focused on the task and just need a few more seconds (anchoring). The nurse points out that the patient looks dusky, which gets your attention, and you retry mask ventilation. However, it is now apparent that you cannot move air. You know you could ventilate before, and cannot believe you cannot ventilate now (anchoring, overconfidence). You place a laryngeal mask with some difficulty, but still cannot ventilate. After another failed laryngoscopy, you attempt a blind intubation with the bougie. You briefly consider asking for a tracheostomy tray but a surgical airway is so dire, and performing a tracheostomy on this young girl will surely get you sued, so the thought is quickly replaced with the hope that the glidescope will arrive soon (omission bias, framing a surgical airway as “bad” instead of life-saving). The heart rate now slows to 35 beats/min and you see wide QRS complexes on the electrocardiogram. Fortunately, you are able to place the bougie and intubate successfully. She stabilizes, the procedure is cancelled, and luckily, she wakes up without any neurologic deficit. You are devastated, relieved, and vow to have a lower threshold for performing awake fiberoptic intubations on patients in the future; this is a hard lesson you will never forget (availability bias). The surgeon reassures you, noting: “After all, all’s well that ends well!” (outcome bias). Your colleagues, however, are incredulous that you would have even considered an asleep intubation, given her history (hindsight bias).

For many decisions in medicine, an optimal choice is not clear. Individual preferences, bias, framing effects, cost, and other factors may all alter perceived decision quality, yet error persists. Self-awareness of human cognitive processes and their potential pitfalls may help improve deliberate thinking strategies (a process called metacognition). This section will describe educational or practice interventions that have been advocated for raising self-awareness of cognitive processes.

Educational Strategies

Targeting Rationale Instead of Behavior.

Developing expertise requires not only experience but also timely and specific feedback on decisions and behaviors.12,91  As Rudolph92  describes, effective feedback hinges not only on observing and correcting incorrect behaviors, but also on understanding how the learner perceives his or her decisions. In our view, all behaviors may be classified using a 2 × 2 table (table 2), where desired actions and undesired actions are crossed with appropriate or inappropriate rationales. Actions (behaviors) may thus be subclassified into the right action for the right reason, the right action for the wrong reason, the wrong action for the right reason, and the wrong action for the wrong reason. An example would be if a trainee administers oxygen to a patient before anesthetic induction (correct) but believes that the rationale is to increase the patient’s oxygen saturation (incorrect). Without specific probe into the rationale, no apparent error is detectable. Both overt performance gaps (actions) and incorrect rationales are important educational targets because an incorrect rationale may lead to future mistakes in related but different situations.

Table 2.

Behaviors, Rationales, and Educational Targets

Behaviors, Rationales, and Educational Targets
Behaviors, Rationales, and Educational Targets

Metacognition (Self-reflection and Reflective Practice).

Metacognition, or “thinking about thinking,” describes the process of reflecting on one’s own thought process and decision-making behavior.93–96  Although better studied in nonmedical domains, increasing insight into decision-making tendencies may improve awareness of decision processes at high risk for medical error.12,57,93,94  Reflective physicians may be more likely to recognize deviations from rational thought processes, and engage in strategies to clarify and/or support their thinking. Approaches such as explicitly testing assumptions, slowing down (to engage in deliberate cognitive processes), seeking alternative explanations, and accepting uncertainty are examples of metacognition and reflective practice.97,98 

Clinical Aids

Cognitive Self-monitoring Strategies.

Cognitive self-monitoring strategies (sometimes called “cognitive forcing strategies” or “debiasing strategies”) are attempts to minimize influences of nonrational decision preferences by creating rules to induce self-monitoring of decision-making processes.98,99 

Such strategies require both a reasoning process and a trigger that signals clinicians to recognize and rescue themselves from the error.100  One such strategy in anesthesiology is the “rule of three.”58  This strategy requires that the anesthesiologist consider at least three alternative explanations before a diagnosis may be accepted, and requires a reassessment of the diagnosis if the first three treatment maneuvers do not produce the expected response. Another is that of “prospective hindsight”—widely used in military strategy—in which the physician imagines a future in which his or her decision is wrong, and then answers the question “What did I miss?”101 

Two additional self-monitoring strategies exist to prevent a focus on obvious or statistically likely diagnoses from itself triggering an anchoring heuristic. The emergency room/trauma axiom “the most commonly missed injury in the emergency room is the second,” is an example intended to help physicians avoid fixation or tunnel vision error by conducting a thorough secondary survey, regardless of the primary or most obvious injury. Additionally, the “rule out worst case” approach to diagnosis is intended to ensure the consideration of statistically rare but very significant diagnoses.12 

Certain conditions may predispose to biased or nonrational decision processes. Clinicians may increase their self-monitoring of these vulnerabilities by asking questions such as (adapted from Graber, via personal communication§):

  • Was the diagnosis suggested to me by the patient, nurse, or a colleague?

  • Did I accept this patient as a “hand-off” from a previous caregiver?

  • Did I consider organ systems besides the obvious one?

  • Is this a patient I do not like, or like too much, for any reason?

  • Have I been interrupted or distracted while caring for this patient?

  • Am I cognitively overloaded right now?

  • Am I stereotyping the patient or the presentation?

Decision Support.

External decision support tools are also effective ways to reduce effects of nonrational cognitive factors. These tools are commonly used in aviation and include checklists, written algorithms, clinical decision aids built into electronic medical records, and guidelines. Intended to decrease omission of important steps in complex procedures, checklists have gained popularity in medicine and improved task performance in perioperative care settings.102  Checklists and similar algorithmic cognitive aids, particularly when managed by a caregiver (“reader”) whose explicit task is limited to ensuring the algorithm is followed, are increasingly popular as decision support tools for critical events in the operating room.103–105  Use of such aids may be effective in promoting better decisions and mitigating the influence of nonrational cognitive factors presented in this review and in table 1. Decision aids built into electronic medical records may also help physicians guard against nonrational cognitive influences. Existing studies, however, are mixed regarding their effectiveness at improving care.106,107  Among the possible explanations for this counterintuitive finding are the poor clinical validity of decision aids, the large amount of “copy and pasted” information in electronic medical records, alert fatigue, and poor physician compliance.

Clinical practice guidelines are another strategy to better align medical decision making with published evidence. Generated most commonly by expert panels convened by specialty societies, practice guidelines are intended to improve decision behavior both by providing an up-to-date literature review and by offering a “best practice” strategy. For anesthesiologists, the American Society of Anesthesiologists Webpage on standards and guidelines contains 11 guidelines, 5 standards, and 31 statements at the time of writing this article.

A recent focus on strategies to reduce medical error has led to greater academic and societal interest in medical decision making. Observations that physician decision behavior is highly variable and often statistically unsupported have raised the possibility that improving physician decision making may reduce medical errors and improve the quality of healthcare delivery.

In principle, a “correct” decision is logically consistent, statistically correct, and considers all available options. In medicine, however, a single correct decision is often difficult to identify. Patient and physician preferences, diagnostic and therapeutic uncertainty, and a wide variety of decision factors, may all complicate identification of the “best” decision. Even when viewed retrospectively, decisions may be difficult to evaluate, as hindsight bias may cause reasonable decisions to be judged harshly if they result in a poor outcome.

Although classic models of decision making in medicine are rooted in logic and probability, evidence shows that real-world medical decision making is frequently driven by the use of cognitive shortcuts, individual preferences, emotions, and an experience base distorted by imperfect recall and inaccurate estimates of likelihood. Although unproven, it seems reasonable that increasing awareness both of intuitive or autonomous decision processes and of statistically driven approaches where possible may improve both the accuracy and consistency of medical decisions.

No evidence exists to unequivocally support routine application of strategies for modifying or increasing awareness of nonrational decision factors. Nevertheless, without self-awareness of how humans make decisions, modifying decision behavior is likely to be difficult. Strategies for improving such self-awareness might begin with educating physicians with respect to the diversity of decision factors currently used in medicine. Explicit teaching about mechanisms of cognition and common errors that result could then be prioritized from the beginning of medical school. Strategies to rapidly recognize and recover from these errors could likewise be taught in medical school, throughout residency, and in continuing medical education. Understanding that framing effects alter the willingness to gamble, for example, may allow physicians to explore their own decision consistency by reconsidering the decision with a different frame. Other cognitive strategies that have been used effectively include counterbalancing heuristics or rules of thumb and forcing strategies where specific conditions trigger a “decision timeout” to make sure that relevant items have been considered. Finally, improving feedback, providing evidence-based guidelines, and increasing access to statistical tools and clinical decision support are other strategies that may raise awareness of decision factors and may improve decision behavior. More research is needed to evaluate the impact of these strategies on clinical outcomes.

*

For example, the recently formed The Society to Improve Diagnosis in Medicine. Available at: http://www.improvediagnosis.org. Accessed October 2, 2013.

The authors’ combined experience in presenting this vignette at multiple venues across the United States is that about 75% of anesthesiologists select PE over the statistically more likely diagnosis of postpartum bleeding.

Readers may note that in this obese patient who has refused heparin, the pretest risk of PE is much higher than baseline estimates. Obesity increases risk of PE (OR = 2.8), see BMJ 2009; 338:b388. But obesity also increases the risk of postpartum bleeding (OR = 2.25), (Am J Obstet Gynecol 2013;209:51.e1-6.) Use of heparin likewise alters risk of embolus. Guidelines for prophylaxis in surgical patients (Chest 2012; 141:e227S–77S) suggests a 40% reduction in PE—not nearly enough to overcome a 100:1 ratio in pretest probability.

§

Graber ML. via e-mail communication on September 30, 2013. Unpublished work related to: Checklists to reduce diagnostic error in emergency medicine. AHRQ ACTION 1 Contract Task Order #8 to RTI International: Develop Patient Safety Interventions to Reduce Diagnostic Error in the Ambulatory Setting. AHRQ Contract Number HHSA290200600001I, 2010.

1.
Kohn
LT
,
Corrigan
J
,
Donaldson
MS
:
To Err Is Human: Building a Safer Health System
.
Washington, D.C.
,
National Academy Press
,
2000
2.
Brook
RH
,
McGlynn
EA
,
Shekelle
PG
:
Defining and measuring quality of care: A perspective from US researchers.
Int J Qual Health Care
2000
;
12
:
281
95
3.
Reid
RO
,
Friedberg
MW
,
Adams
JL
,
McGlynn
EA
,
Mehrotra
A
:
Associations between physician characteristics and quality of care.
Arch Intern Med
2010
;
170
:
1442
9
4.
Schuster
MA
,
McGlynn
EA
,
Brook
RH
:
How good is the quality of health care in the United States?
Milbank Q
1998
;
76
:
517
63, 509
5.
McGlynn
EA
,
Asch
SM
,
Adams
J
,
Keesey
J
,
Hicks
J
,
DeCristofaro
A
,
Kerr
EA
:
The quality of health care delivered to adults in the United States.
N Engl J Med
2003
;
348
:
2635
45
6.
Driskell
OJ
,
Holland
D
,
Hanna
FW
,
Jones
PW
,
Pemberton
RJ
,
Tran
M
,
Fryer
AA
:
Inappropriate requesting of glycated hemoglobin (HbA1c) is widespread: Assessment of prevalence, impact of national guidance, and practice-to-practice variability.
Clin Chem
2012
;
58
:
906
15
7.
Berner
ES
,
Graber
ML
:
Overconfidence as a cause of diagnostic error in medicine.
Am J Med
2008
;
121
(
5 suppl
):
S2
23
8.
Morrow
DG
,
Durso
FT
:
Health care research that delivers: Introduction to the special issue on cognitive factors in health care.
J Exp Psychol Appl
2011
;
17
:
191
4
9.
Chellis
M
,
Olson
J
,
Augustine
J
,
Hamilton
G
:
Evaluation of missed diagnoses for patients admitted from the emergency department.
Acad Emerg Med
2001
;
8
:
125
30
10.
Podbregar
M
,
Voga
G
,
Krivec
B
,
Skale
R
,
Pareznik
R
,
Gabrscek
L
:
Should we confirm our clinical diagnostic certainty by autopsies?
Intensive Care Med
2001
;
27
:
1750
5
11.
Davies
JM
,
Posner
KL
,
Lee
LA
,
Cheney
FW
,
Domino
KB
:
Liability associated with obstetric anesthesia: A closed claims analysis.
Anesthesiology
2009
;
110
:
131
9
12.
Graber
ML
,
Kissam
S
,
Payne
VL
,
Meyer
AN
,
Sorensen
A
,
Lenfestey
N
,
Tant
E
,
Henriksen
K
,
Labresh
K
,
Singh
H
:
Cognitive interventions to reduce diagnostic error: A narrative review.
BMJ Qual Saf
2012
;
21
:
535
57
13.
Graber
ML
,
Franklin
N
,
Gordon
R
:
Diagnostic error in internal medicine.
Arch Intern Med
2005
;
165
:
1493
9
14.
van den Einden
LC
,
de Hullu
JA
,
Massuger
LF
,
Grefte
JM
,
Bult
P
,
Wiersma
A
,
van Engen-van Grunsven
AC
,
Sturm
B
,
Bosch
SL
,
Hollema
H
,
Bulten
J
:
Interobserver variability and the effect of education in the histopathological diagnosis of differentiated vulvar intraepithelial neoplasia.
Mod Pathol
2013
;
26
:
874
80
15.
Gobezie
R
,
Zurakowski
D
,
Lavery
K
,
Millett
PJ
,
Cole
BJ
,
Warner
JJ
:
Analysis of interobserver and intraobserver variability in the diagnosis and treatment of SLAP tears using the Snyder classification.
Am J Sports Med
2008
;
36
:
1373
9
16.
Lim
D
,
Alvarez
T
,
Nucci
MR
,
Gilks
B
,
Longacre
T
,
Soslow
RA
,
Oliva
E
:
Interobserver variability in the interpretation of tumor cell necrosis in uterine leiomyosarcoma.
Am J Surg Pathol
2013
;
37
:
650
8
17.
Aldrink
JH
,
Caniano
DA
,
Nwomeh
BC
:
Variability in gastroschisis management: A survey of North American pediatric surgery training programs.
J Surg Res
2012
;
176
:
159
63
18.
Buchan
CA
,
Bravi
A
,
Seely
AJ
:
Variability analysis and the diagnosis, management, and treatment of sepsis.
Curr Infect Dis Rep
2012
;
14
:
512
21
19.
Frank
SM
,
Savage
WJ
,
Rothschild
JA
,
Rivers
RJ
,
Ness
PM
,
Paul
SL
,
Ulatowski
JA
:
Variability in blood and blood component utilization as assessed by an anesthesia information management system.
A
2012
;
117
:
99
6
20.
Chong
PC
,
Greco
EF
,
Stothart
D
,
Maziak
DE
,
Sundaresan
S
,
Shamji
FM
,
Neilipovitz
D
,
McIntyre
L
,
Hébert
P
,
Seely
AJ
:
Substantial variation of both opinions and practice regarding perioperative fluid resuscitation.
Can J Surg
2009
;
52
:
207
14
21.
Vigoda
MM
,
Sweitzer
B
,
Miljkovic
N
,
Arheart
KL
,
Messinger
S
,
Candiotti
K
,
Lubarsky
D
:
2007 American College of Cardiology/American Heart Association (ACC/AHA) Guidelines on perioperative cardiac evaluation are usually incorrectly applied by anesthesiology residents evaluating simulated patients.
Anesth Analg
2011
;
112
:
940
9
22.
Vigoda
MM
,
Behrens
V
,
Miljkovic
N
,
Arheart
KL
,
Lubarsky
DA
,
Dutton
RP
:
Perioperative cardiac evaluation of simulated patients by practicing anesthesiologists is not consistent with 2007 ACC/AHA guidelines.
J Clin Anesth
2012
;
24
:
446
55
23.
Friedman
M
,
Savage
LJ
:
The expected-utility hypothesis and the measurability of utility.
J Political Econ
1952
;
60
:
463
74
24.
Schoemaker
PJH
:
The expected utility model—Its variants, purposes, evidence and limitations.
J Econ Lit
1982
;
20
:
529
63
25.
Leahey
TH
:
Herbert A. Simon: Nobel Prize in Economic Sciences, 1978.
Am Psychol
2003
;
58
:
753
5
26.
Cohen
BJ
:
Is expected utility theory normative for medical decision making?
Med Decis Making
1996
;
16
:
1
6
27.
Brandstatter
E
,
Gussmack
M
:
The cognitive processes underlying risky choice.
J Behav Decis Mak
2013
;
26
:
185
97
28.
Russell
LB
,
Schwartz
A
:
Looking at patients’ choices through the lens of expected utility: A critique and research agenda.
Med Decis Making
2012
;
32
:
527
31
29.
Ashby
D
:
Bayesian statistics in medicine: A 25 year review.
Stat Med
2006
;
25
:
3589
631
30.
Kadane
JB
:
Bayesian methods for health-related decision making.
Stat Med
2005
;
24
:
563
7
31.
Eddy
DM
,
Clanton
CH
:
The art of diagnosis: Solving the clinicopathological exercise.
N Engl J Med
1982
;
306
:
1263
8
32.
Mancuso
CA
,
Rose
DN
:
A model for physicians’ therapeutic decision making.
Arch Intern Med
1987
;
147
:
1281
5
33.
Kahneman
D
:
A perspective on judgment and choice: Mapping bounded rationality.
Am Psychol
2003
;
58
:
697
720
34.
Marewski
JN
,
Gigerenzer
G
:
Heuristic decision making in medicine.
Dialogues Clin Neurosci
2012
;
14
:
77
89
35.
Crowley
RS
,
Legowski
E
,
Medvedeva
O
,
Reitmeyer
K
,
Tseytlin
E
,
Castine
M
,
Jukic
D
,
Mello-Thoms
C
:
Automated detection of heuristics and biases among pathologists in a computer-based system.
Adv Health Sci Educ Theory Pract
2013
;
18
:
343
63
36.
Tversky
A
,
Kahneman
D
:
Judgment under uncertainty: Heuristics and biases.
Science
1974
;
185
:
1124
31
37.
Tversky
A
,
Kahneman
D
:
Extensional versus intuitive reasoning: The conjunction fallacy in probability judgment.
Psych Rev
1983
;
90
:
293
15
38.
Kahneman
D
,
Tversky
A
:
On the psychology of prediction.
Psych Rev
1973
;
80
:
237
51
39.
Ford
MK
,
Beattie
WS
,
Wijeysundera
DN
:
Systematic review: Prediction of perioperative cardiac complications and mortality by the revised cardiac risk index.
Ann Intern Med
2010
;
152
:
26
35
40.
Gupta
PK
,
Gupta
H
,
Sundaram
A
,
Kaushik
M
,
Fang
X
,
Miller
WJ
,
Esterbrooks
DJ
,
Hunter
CB
,
Pipinos
II
,
Johanning
JM
,
Lynch
TG
,
Forse
RA
,
Mohiuddin
SM
,
Mooss
AN
:
Development and validation of a risk calculator for prediction of cardiac risk after surgery.
Circulation
2011
;
124
:
381
7
41.
Morris
JM
,
Algert
CS
,
Roberts
CL
:
Incidence and risk factors for pulmonary embolism in the postpartum period.
J Thromb Haemost
2010
;
8
:
998
3
42.
Della Torre
M
,
Kilpatrick
SJ
,
Hibbard
JU
,
Simonson
L
,
Scott
S
,
Koch
A
,
Schy
D
,
Geller
SE
:
Assessing preventability for obstetric hemorrhage.
Am J Perinatol
2011
;
28
:
753
60
43.
Poses
RM
,
Anthony
M
:
Availability, wishful thinking, and physicians’ diagnostic judgments for patients with suspected bacteremia.
Med Decis Making
1991
;
11
:
159
68
44.
Richards
MS
,
Wierzbicki
M
:
Anchoring errors in clinical-like judgments.
J Clin Psychol
1990
;
46
:
358
65
45.
Senay
I
,
Kaphingst
KA
:
Anchoring-and-adjustment bias in communication of disease risk.
Med Decis Making
2009
;
29
:
193
1
46.
Croskerry
P
:
A universal model of diagnostic reasoning.
Acad Med
2009
;
84
:
1022
8
47.
Marcum
JA
:
An integrated model of clinical reasoning: Dual-process theory of cognition and metacognition.
J Eval Clin Pract
2012
;
18
:
954
61
48.
Croskerry
P
:
Clinical cognition and diagnostic error: Applications of a dual process model of reasoning.
Adv Health Sci Educ Theory Pract
2009
;
14
(
suppl 1
):
27
35
49.
Reyna
VF
:
A theory of medical decision making and health: Fuzzy trace theory.
Med Decis Making
2008
;
28
:
850
65
50.
Djulbegovic
B
,
Hozo
I
,
Beckstead
J
,
Tsalatsanis
A
,
Pauker
SG
:
Dual processing model of medical decision-making.
BMC Med Inform Decis Mak
2012
;
12
:
94
51.
Paget
MA
:
The Unity of Mistakes
.
Philadelphia
,
Temple University Press
,
1988
, pp
pp 96
7
52.
Rudolph
JW
,
Morrison
JB
,
Carroll
JS
:
The dynamics of action-oriented problem-solving: Linking interpretation and choice.
Acad Management Rev
2009
;
34
:
733
56
53.
Wieck
KE
,
Sutcliffe
KM
:
Organizing and the process of sensemaking.
Org Sci
2005
;
16
:
409
21
54.
Fischhoff
B
:
Hindsight not equal to foresight: The effect of outcome knowledge on judgment under uncertainty.
Qual Saf Health Care
2003
;
12
:
304
11; discussion 311
55.
Hawkins
SA
,
Hastie
R
:
Hindsight—Biased judgments of past events after the outcomes are known.
Psychol Bull
1990
;
107
:
311
27
56.
Caplan
RA
,
Posner
KL
,
Cheney
FW
:
Effect of outcome on physician judgments of appropriateness of care.
JAMA
1991
;
265
:
1957
60
57.
Stiegler
MP
,
Neelankavil
JP
,
Canales
C
,
Dhillon
A
:
Cognitive errors detected in anaesthesiology: A literature review and pilot study.
Br J Anaesth
2012
;
108
:
229
35
58.
Stiegler
MP
,
Ruskin
KJ
:
Decision-making and safety in anesthesiology.
Curr Opin Anaesthesiol
2012
;
25
:
724
9
59.
Dror
IE
,
Charlton
D
,
Péron
AE
:
Contextual information renders experts vulnerable to making erroneous identifications.
Forensic Sci Int
2006
;
156
:
74
8
60.
Leblanc
VR
,
Norman
GR
,
Brooks
LR
:
Effect of a diagnostic suggestion on diagnostic accuracy and identification of clinical features.
Acad Med
2001
;
76
(
10 suppl
):
S18
20
61.
Leblanc
VR
,
Brooks
LR
,
Norman
GR
:
Believing is seeing: The influence of a diagnostic hypothesis on the interpretation of clinical features.
Acad Med
2002
;
77
(
10 suppl
):
S67
9
62.
Weintraub
W
:
“The VIP syndrome”: A clinical study in hospital psychiatry.
J Nerv Ment Dis
1964
;
138
:
181
93
63.
Block
AJ
:
Beware of the VIP syndrome.
Chest
1993
;
104
:
989
64.
Baron
J
,
Ritov
I
:
Reference points and omission bias.
Organ Behav Hum Decis Process
1994
;
59
:
475
98
65.
Aberegg
SK
,
Haponik
EF
,
Terry
PB
:
Omission bias and decision making in pulmonary and critical care medicine.
Chest
2005
;
128
:
1497
505
66.
Pronin
E
,
Lin
DY
,
Ross
L
:
The bias blind spot: Perceptions of bias in self versus others.
Pers Soc Psych Bull
2002
;
28
:
369
81
67.
West
RF
,
Meserve
RJ
,
Stanovich
KE
:
Cognitive sophistication does not attenuate the bias blind spot.
J Pers Soc Psychol
2012
;
103
:
506
19
68.
Bazerman
MH
:
Motivational and Affective Influences on Decision Making. Judgment in Managerial Decision Making
, 6th edition.
Hoboken
,
John Wiley & sons
,
1986
, pp
pp 68
73
69.
Gilovich
T
:
How We Know What Isn’t So
.
New York
,
Simon & Schuster
,
1991
, pp
pp 77
70.
Davis
DA
,
Mazmanian
PE
,
Fordis
M
,
Van Harrison
R
,
Thorpe
KE
,
Perrier
L
:
Accuracy of physician self-assessment compared with observed measures of competence: A systematic review.
JAMA
2006
;
296
:
1094
102
71.
Landefeld
CS
,
Chren
MM
,
Myers
A
,
Geller
R
,
Robbins
S
,
Goldman
L
:
Diagnostic yield of the autopsy in a university hospital and a community hospital.
N Engl J Med
1988
;
318
:
1249
54
72.
Reason
JT
:
Managing the Risks of Organizational Accidents
.
Aldershot, Hants; Brookfield
,
Ashgate
,
1997
, pp
pp xvii 25270
73.
Lee
JL
:
Reconsolidation: Maintaining memory relevance.
Trends Neurosci
2009
;
32
:
413
20
74.
Schacter
DL
,
Guerin
SA
,
St Jacques
PL
:
Memory distortion: An adaptive perspective.
Trends Cogn Sci
2011
;
15
:
467
74
75.
Chan
JC
,
Lapaglia
JA
:
The dark side of testing memory: Repeated retrieval can enhance eyewitness suggestibility.
J Exp Psychol Appl
2011
;
17
:
418
32
76.
Allais
M
:
Le comportement de l’homme rationnel devant le risqué: Critique des postulats et axioms de l’ecole americaine.
Econometrica
1952
;
21
:
503
46
77.
Tversky
A
,
Kahneman
D
:
The framing of decisions and the psychology of choice.
Science
1981
;
211
:
453
8
78.
Gong
J
,
Zhang
Y
,
Yang
Z
,
Huang
Y
,
Feng
J
,
Zhang
W:
:
The framing effect in medical decision-making: A review of the literature.
Psychol Health Med
2013
;
18
:
645
53
79.
Tversky
A
,
Kahneman
D
:
Advances in prospect theory: Cumulative representation of uncertainty.
J Risk Uncertainty
1992
;
5
:
297
23
80.
Dale
W
,
Hemmerich
J
,
Moliski
E
,
Schwarze
ML
,
Tung
A
:
Effect of specialty and recent experience on perioperative decision-making for abdominal aortic aneurysm repair.
J Am Geriatr Soc
2012
;
60
:
1889
94
81.
Schwarze
ML
,
Redmann
AJ
,
Brasel
KJ
,
Alexander
GC
:
The role of surgeon error in withdrawal of postoperative life support.
Ann Surg
2012
;
256
:
10
5
82.
Zajonc
RB
:
Feeling and thinking: Preferences need no inferences.
Am Psych
1980
;
35
:
151
75
83.
Sansone
RA
,
Farukhi
S
,
Wiederman
MW
:
Disruptive behaviors in the medical setting and borderline personality.
Int J Psychiatry Med
2011
;
41
:
355
63
84.
Rosenstein
AH
,
O’Daniel
M
:
Impact and implications of disruptive behavior in the perioperative arena.
J Am Coll Surg
2006
;
203
:
96
5
85.
Rosenstein
AH
,
Naylor
B
:
Incidence and impact of physician and nurse disruptive behaviors in the emergency department.
J Emerg Med
2012
;
43
:
139
48
86.
Loomes
G
,
Sugden
R
:
Regret theory—An alternative theory of rational choice under uncertainty.
Economic J
1982
;
92
:
805
24
87.
Hozo
I
,
Djulbegovic
B
:
When is diagnostic testing inappropriate or irrational? Acceptable regret approach.
Med Decis Making
2008
;
28
:
540
53
88.
Sorum
PC
,
Mullet
E
,
Shim
J
,
Bonnin-Scaon
S
,
Chasseigne
G
,
Cogneau
J
:
Avoidance of anticipated regret: The ordering of prostate-specific antigen tests.
Med Decis Making
2004
;
241
:
49
59
89.
Pandit
JJ
,
Cook
TM
,
Jonker
WR
,
O’Sullivan
E
;
5th National Audit Project (NAP5) of the Royal College of Anaesthetists and the Association of Anaesthetists of Great Britain and Ireland
:
A national survey of anaesthetists (NAP5 Baseline) to estimate an annual incidence of accidental awareness during general anaesthesia in the UK.
Anaesthesia
2013
;
68
:
343
53
90.
American Society of Anesthesiologists Task Force on Management of the Difficult Airway
:
Practice guidelines for management of the difficult airway: An updated report by the American Society of Anesthesiologists Task Force on Management of the Difficult Airway.
A
2013
;
118
:
251
70
91.
Ericsson
KA
:
Deliberate practice and acquisition of expert performance: A general overview.
Acad Emerg Med
2008
;
15
:
988
94
92.
Rudolph
JW
,
Simon
R
,
Raemer
DB
,
Eppich
WJ
:
Debriefing as formative assessment: Closing performance gaps in medical education.
Acad Emerg Med
2008
;
15
:
1010
6
93.
Croskerry
P
:
The importance of cognitive errors in diagnosis and strategies to minimize them.
Acad Med
2003
;
78
:
775
80
94.
Croskerry
P
:
The cognitive imperative: Thinking about how we think.
Acad Emerg Med
2000
;
7
:
1223
31
95.
Mamede
S
,
Schmidt
HG
,
Rikers
R
:
Diagnostic errors and reflective practice in medicine.
J Eval Clin Pract
2007
;
13
:
138
45
96.
Mamede
S
,
van Gog
T
,
Moura
AS
,
de Faria
RM
,
Peixoto
JM
,
Rikers
RM
,
Schmidt
HG
:
Reflection as a strategy to foster medical students’ acquisition of diagnostic competence.
Med Educ
2012
;
46
:
464
72
97.
Moulton
CA
,
Regehr
G
,
Mylopoulos
M
,
MacRae
HM
:
Slowing down when you should: A new model of expert judgment.
Acad Med
2007
;
82
(
10 suppl
):
S109
16
98.
Trowbridge
RL
:
Twelve tips for teaching avoidance of diagnostic errors.
Med Teach
2008
;
30
:
496
500
99.
Croskerry
P
,
Singhal
G
,
Mamede
S
:
Cognitive debiasing 1: Origins of bias and theory of debiasing.
BMJ Qual Saf
2013
(
suppl 2
):
ii58–4
100.
Kassirer
JP
:
Teaching clinical reasoning: Case-based and coached.
Acad Med
2010
;
85
:
1118
24
101.
Singh
H
,
Graber
M
:
Reducing diagnostic error through medical home-based primary care reform.
JAMA
2010
;
304
:
463
4
102.
Haynes
AB
,
Weiser
TG
,
Berry
WR
,
Lipsitz
SR
,
Breizat
AH
,
Dellinger
EP
,
Herbosa
T
,
Joseph
S
,
Kibatala
PL
,
Lapitan
MC
,
Merry
AF
,
Moorthy
K
,
Reznick
RK
,
Taylor
B
,
Gawande
AA
;
Safe Surgery Saves Lives Study Group
:
A surgical safety checklist to reduce morbidity and mortality in a global population.
N Engl J Med
2009
;
360
:
491
9
103.
Moitra
VK
,
Gabrielli
A
,
Maccioli
GA
,
O’Connor
MF
:
Anesthesia advanced circulatory life support.
Can J Anaesth
2012
;
59
:
586
3
104.
Burden
AR
,
Carr
ZJ
,
Staman
GW
,
Littman
JJ
,
Torjman
MC
:
Does every code need a “reader?” improvement of rare event management with a cognitive aid “reader” during a simulated emergency: A pilot study.
Simul Healthc
2012
;
7
:
1
9
105.
Ziewacz
JE
,
Arriaga
AF
,
Bader
AM
,
Berry
WR
,
Edmondson
L
,
Wong
JM
,
Lipsitz
SR
,
Hepner
DL
,
Peyre
S
,
Nelson
S
,
Boorman
DJ
,
Smink
DS
,
Ashley
SW
,
Gawande
AA
:
Crisis checklists for the operating room: Development and pilot testing.
J Am Coll Surg
2011
;
213
:
212
217.e10
106.
Main
C
,
Moxham
T
,
Wyatt
JC
,
Kay
J
,
Anderson
R
,
Stein
K
:
Computerised decision support systems in order communication for diagnostic, screening or monitoring test ordering: Systematic reviews of the effects and cost-effectiveness of systems.
Health Technol Assess
2010
;
14
:
1
227
107.
Rollman
BL
,
Hanusa
BH
,
Lowe
HJ
,
Gilbert
T
,
Kapoor
WN
,
Schulberg
HC
:
A randomized trial using computerized decision support to improve treatment of major depression in primary care.
J Gen Intern Med
2002
;
17
:
493
3