Humans have always made decisions, but the systematic study of how decisions should be made, and how they actually are made, is surprisingly recent. For most of human history, decision-making was guided by custom, intuition, religious authority, or the counsel of experienced elders. The idea that there might be a formal science of decision-making, a set of mathematical principles governing how a rational agent should choose among alternatives under conditions of uncertainty, did not emerge until the seventeenth and eighteenth centuries, and did not achieve its mature form until the mid-twentieth century.
The intellectual journey from the first attempts to formalize reasoning about chance in the gambling houses of Renaissance Europe to the sophisticated decision-theoretic frameworks used today in artificial intelligence, military strategy, medical diagnosis, financial engineering, and public policy is a story of remarkable creativity, fierce intellectual debate, and enduring tension between how people should decide and how they actually do. This tension, between normative decision theory (what rationality requires) and descriptive decision theory (what people actually do), has driven the field's development and continues to generate its most productive controversies.
Understanding the origins of decision theory matters beyond academic curiosity. The frameworks that emerged from this intellectual history shape how governments evaluate policies, how doctors weigh treatment options, how investors price assets, how engineers design systems, and how artificial intelligence systems are built to make choices on our behalf. The assumptions embedded in these frameworks, assumptions about rationality, probability, utility, and information, have consequences that affect billions of lives. Knowing where these assumptions came from, what problems they were designed to solve, and what alternatives were considered and rejected provides essential context for evaluating whether those assumptions are appropriate for the decisions we face today.
The Prehistory: Probability and the Problem of Uncertainty
Decision theory could not exist without probability theory, because most interesting decisions involve uncertainty. And probability theory, despite seeming like an obvious mathematical concept, emerged remarkably late in the history of mathematics.
Gambling and the Birth of Probability
The mathematical study of probability began in 1654 with a famous correspondence between Blaise Pascal and Pierre de Fermat, prompted by a question from the gambler Antoine Gombaud, Chevalier de Meré. Gombaud wanted to know how to fairly divide the stakes of a game that was interrupted before completion. This seemingly trivial question about gambling forced Pascal and Fermat to develop the first rigorous mathematical treatment of uncertain outcomes.
"Probability theory is nothing but common sense reduced to calculation." -- Pierre-Simon Laplace
Pascal and Fermat's key insight was that uncertain future outcomes could be analyzed through systematic enumeration of all possible outcomes and their relative frequencies. If a fair coin is tossed twice, there are four equally likely outcomes (HH, HT, TH, TT). If you need to know the probability of getting at least one head, you count the favorable outcomes (3) and divide by the total (4), giving 3/4. This seems obvious now, but before Pascal and Fermat, even sophisticated thinkers struggled with problems involving chance because they lacked a mathematical framework for reasoning about uncertainty.
Pascal himself extended probabilistic reasoning beyond gambling in his famous "Pascal's Wager," perhaps the first explicit use of decision-theoretic reasoning. Pascal argued that a rational person should believe in God because the expected value of belief (infinite reward if God exists, minimal cost if God doesn't) exceeds the expected value of disbelief (no reward if God exists, no cost if God doesn't). The wager is philosophically problematic for many reasons, but its historical significance lies in applying mathematical reasoning about uncertain outcomes to a practical decision, combining probability (the uncertain existence of God) with utility (the value of the outcomes) to determine rational action.
Jacob Bernoulli and the Law of Large Numbers
Jacob Bernoulli's Ars Conjectandi (The Art of Conjecturing), published posthumously in 1713, established the first rigorous foundations of probability theory. Bernoulli proved the law of large numbers, which states that as the number of observations increases, the observed frequency of an event converges to its true probability. This theorem provided the mathematical link between the abstract concept of probability and the observable world of repeated events, giving probability theory its empirical grounding.
Bernoulli's work also began to consider the application of probability to practical decision-making beyond gambling. He discussed the evaluation of evidence in legal proceedings, the assessment of risk in commerce, and the general problem of making rational judgments under uncertainty. His nephew, Daniel Bernoulli, would take this practical orientation much further.
The St. Petersburg Paradox and the Birth of Utility Theory
The conceptual foundations of decision theory crystallized around a single paradox that revealed the inadequacy of the most obvious approach to decision under risk.
The Paradox
In 1713, Nicolas Bernoulli (another member of the mathematically prolific Bernoulli family) posed a problem that became known as the St. Petersburg paradox. Consider a game in which a fair coin is tossed repeatedly until it lands heads. If the first heads appears on the nth toss, you receive 2^n dollars. So if heads appears on the first toss, you get $2. If heads first appears on the second toss, you get $4. Third toss, $8. Fourth toss, $16. And so on.
The expected value of this game, the probability-weighted average of all possible payoffs, is infinite. The probability of heads on the first toss is 1/2, and the payoff is $2, so the expected value from this outcome is $1. The probability of first heads on the second toss is 1/4, and the payoff is $4, so the expected value from this outcome is also $1. Each possible outcome contributes $1 to the expected value, and there are infinitely many possible outcomes, so the total expected value is 1 + 1 + 1 + ... = infinity.
If rational decision-making simply means maximizing expected value, then a rational agent should be willing to pay any finite amount to play this game. But no reasonable person would pay more than a modest amount, perhaps $10 or $20. The paradox is that the mathematically "rational" decision (pay any price) conflicts dramatically with the intuitively sensible decision (pay only a small amount).
Daniel Bernoulli's Solution: Diminishing Marginal Utility
Daniel Bernoulli proposed a solution in 1738 that introduced one of the most important concepts in economics and decision theory: diminishing marginal utility. Bernoulli argued that the value of money to a person is not proportional to the amount of money but to the logarithm of the amount. The first $1,000 you receive has much greater value to you than the additional $1,000 received after you already have $100,000. By replacing monetary value with utility (subjective value), the expected utility of the St. Petersburg game becomes finite, and the paradox dissolves.
Bernoulli's insight had implications far beyond the paradox. It provided a psychological explanation for risk aversion: because utility is a concave function of wealth (each additional dollar provides less additional utility than the previous one), a certain outcome is preferred to a gamble with the same expected value. Receiving $50 for certain provides more utility than a 50/50 chance of receiving $100 or $0, because the utility lost by going from $50 to $0 is greater than the utility gained by going from $50 to $100. This explained why people buy insurance (accepting a small certain loss to avoid a large uncertain loss) and why investors demand higher returns for riskier investments.
"It is not the man who has too little, but the man who craves more, that is poor." -- Seneca
The Twentieth-Century Formalization
The informal insights of Pascal, the Bernoullis, and their successors were transformed into a rigorous mathematical framework in the twentieth century through three landmark contributions that, together, defined the structure of modern decision theory.
Von Neumann and Morgenstern: Axiomatic Utility Theory
In 1944, mathematician John von Neumann and economist Oskar Morgenstern published Theory of Games and Economic Behavior, which laid the foundations for both game theory and modern utility theory. Their contribution to decision theory was the axiomatic derivation of expected utility.
Von Neumann and Morgenstern showed that if an agent's preferences over gambles satisfy four axioms, then the agent behaves as if assigning numerical utilities to outcomes and choosing the option with the highest expected utility. The four axioms are:
Completeness: For any two options A and B, the agent either prefers A, prefers B, or is indifferent between them. No options are incomparable.
Transitivity: If the agent prefers A to B and B to C, then the agent prefers A to C. Preferences are consistent.
Continuity: If the agent prefers A to B to C, then there exists some probability p such that the agent is indifferent between B for certain and a gamble giving A with probability p and C with probability 1-p. Preferences vary smoothly, without abrupt jumps.
Independence: If the agent prefers A to B, then the agent also prefers a gamble that gives A with probability p and C with probability 1-p to a gamble that gives B with probability p and C with probability 1-p. Preferences between two options are not affected by what other options are mixed in.
The brilliance of this approach was that it derived expected utility maximization as a theorem rather than assuming it as a postulate. If your preferences satisfy these reasonable-seeming axioms, then you are mathematically guaranteed to be behaving as if maximizing expected utility. This gave the rational actor model an axiomatic foundation that seemed nearly irrefutable: who would want to have incomplete, intransitive, discontinuous, or non-independent preferences?
"The best we can do is size up the chances, calculate the risks involved, estimate our ability to deal with them, and then make our plans with confidence." -- Henry Ford
Savage: Subjective Probability and Decision Theory
Leonard Savage extended von Neumann and Morgenstern's framework in his 1954 book The Foundations of Statistics. While von Neumann and Morgenstern assumed that probabilities were objectively given (as in a fair coin toss or a roulette wheel), Savage showed that both utilities and probabilities could be derived from an agent's preferences.
Savage's framework dealt with decisions where the probabilities of different states of the world are not objectively known, which describes most real-world decisions. When deciding whether to carry an umbrella, you do not know the objective probability of rain; you have a subjective degree of belief that it will rain. Savage showed that if an agent's preferences over actions satisfy a set of axioms (similar to but more elaborate than von Neumann and Morgenstern's), then the agent behaves as if assigning subjective probabilities to states and expected utility to actions.
This was a profound result because it unified probability and utility into a single decision-theoretic framework. A rational agent, in Savage's sense, simultaneously has consistent beliefs (represented by subjective probabilities) and consistent values (represented by utilities), and makes choices that maximize expected utility given those beliefs and values. The framework provided the theoretical foundation for Bayesian decision theory, which became the dominant normative framework for decision-making under uncertainty.
The Role of World War II: Operations Research
While theorists were developing axiomatic foundations, World War II generated enormous practical demand for systematic decision-making methods. Operations research (OR), which emerged from military efforts to optimize resource allocation, logistics, and strategy, pushed theoretical developments toward practical application.
British and American military scientists developed quantitative methods for deciding how to deploy radar installations, how to configure convoy formations to minimize submarine losses, how to allocate bombing resources among multiple targets, and how to schedule aircraft maintenance to maximize fleet availability. These were genuine decision problems involving uncertainty, multiple objectives, and scarce resources, exactly the problems that decision theory was designed to address.
The success of wartime operations research led to its rapid adoption in civilian contexts after the war. Companies like the RAND Corporation, founded in 1948 as a military think tank, employed mathematicians, economists, and engineers to apply decision-theoretic methods to problems ranging from nuclear strategy to urban planning. The Ford Foundation funded the creation of management science programs at leading business schools. Linear programming, dynamic programming, queuing theory, and other optimization methods developed during and after the war became standard tools of business management and government planning.
The Split: Normative Versus Descriptive Decision Theory
By the 1960s, decision theory had achieved an impressive mathematical structure. But a fundamental tension had been building within the field from the beginning: the tension between how rational agents should decide and how actual human beings do decide. This tension eventually split the field into two distinct branches.
The Difference Between Normative and Descriptive
Normative decision theory prescribes how decisions should be made if the goal is to satisfy the axioms of rationality. It says: if you want to be consistent, coherent, and immune to exploitation by money pumps (schemes that extract money from someone with intransitive preferences), then you should maximize expected utility. Normative theory does not describe what people do; it describes what rationality requires.
Descriptive decision theory studies how people actually make decisions. It uses experiments, surveys, and observation to discover the patterns, regularities, and systematic deviations that characterize real human decision-making. Descriptive theory does not prescribe what people should do; it describes what they actually do.
The gap between normative and descriptive theory is the gap between the rational actor model and real human behavior. And that gap, as decades of research have shown, is enormous.
The Allais Paradox: The First Major Challenge
The first major empirical challenge to expected utility theory came from French economist Maurice Allais in 1953. Allais presented a pair of choice problems to a group of sophisticated decision-makers at a conference in Paris and showed that the majority violated the independence axiom, one of the cornerstones of expected utility theory.
In the first choice problem, people chose between:
- Option A: $1 million with certainty
- Option B: 89% chance of $1 million, 10% chance of $5 million, 1% chance of nothing
Most people chose A, preferring the certain $1 million to the gamble.
In the second choice problem, people chose between:
- Option C: 11% chance of $1 million, 89% chance of nothing
- Option D: 10% chance of $5 million, 90% chance of nothing
Most people chose D, preferring the slightly lower probability but much larger potential payoff.
The combination of choosing A and D violates the independence axiom. Allais demonstrated this to several of the assembled decision theorists, including Savage himself, who initially made the "irrational" choice pattern and was disturbed to discover that he had violated his own axioms. The Allais paradox showed that even highly sophisticated, mathematically trained decision-makers systematically violated the axioms of expected utility theory, not through ignorance or carelessness, but because something about the axioms failed to capture a genuine feature of human preferences.
The feature that Allais identified was what he called the "certainty effect": people place special value on outcomes that are certain versus merely probable, a value that is not captured by the linear probability weighting assumed by expected utility theory. This insight would later be incorporated into Kahneman and Tversky's prospect theory, which explicitly modeled the nonlinear weighting of probabilities.
The Ellsberg Paradox: Ambiguity Aversion
A second major challenge came from Daniel Ellsberg (later famous for leaking the Pentagon Papers) in 1961. Ellsberg showed that people's preferences violate Savage's axioms when they face ambiguity, meaning uncertainty about the probabilities themselves.
In Ellsberg's experiment, subjects are presented with an urn containing 90 balls. They know that 30 balls are red, and the remaining 60 are some unknown mixture of black and yellow. They are asked to choose between bets, and their choices consistently reveal a preference for known over unknown probabilities, even when this preference violates the axioms of subjective expected utility theory.
Ellsberg's paradox demonstrated ambiguity aversion: people dislike situations where they don't even know the probabilities, above and beyond their dislike of risk (known probabilities of loss). This distinction between risk and ambiguity (sometimes called "Knightian uncertainty" after economist Frank Knight, who distinguished them in 1921) has proved important in understanding behavior in domains from financial markets (investors demand higher returns for ambiguous investments) to medical decision-making (patients prefer treatments with well-documented outcomes over treatments with uncertain evidence).
Prospect Theory: The Descriptive Revolution
The most influential alternative to expected utility theory emerged in 1979 when Daniel Kahneman and Amos Tversky published "Prospect Theory: An Analysis of Decision Under Risk" in Econometrica. Prospect theory did not merely identify anomalies in expected utility theory; it proposed a comprehensive alternative model of how people actually make risky decisions.
"The confidence people have in their beliefs is not a measure of the quality of evidence but of the coherence of the story that the mind has managed to construct." -- Daniel Kahneman
How Prospect Theory Challenged Classical Decision Theory
Prospect theory departed from expected utility theory in three fundamental ways.
First, reference dependence: prospect theory evaluated outcomes as gains and losses relative to a reference point, not as absolute levels of wealth. Expected utility theory evaluated outcomes in terms of total wealth: a decision that results in a final wealth of $100,000 has the same utility regardless of whether the person started with $50,000 (a gain of $50,000) or $200,000 (a loss of $100,000). Prospect theory recognized that these situations feel completely different psychologically and produce different choices.
Second, loss aversion: the value function in prospect theory was steeper for losses than for gains, reflecting the empirical finding that people feel the pain of a loss approximately twice as intensely as the pleasure of an equivalent gain. This asymmetry explained the endowment effect (people value things they own more than identical things they don't own), the status quo bias (people prefer to keep what they have), and the disposition effect in investing (people hold losing investments too long and sell winners too early).
Third, probability weighting: prospect theory replaced the linear probability weighting of expected utility theory with a nonlinear weighting function that overweights small probabilities and underweights large probabilities. This explained both the certainty effect identified by Allais (the discontinuous jump in value between a certain outcome and an almost-certain outcome) and the simultaneous purchase of insurance (overweighting the small probability of catastrophic loss) and lottery tickets (overweighting the small probability of a large gain).
Prospect theory's success was not merely empirical. It was also explanatory. The theory did not merely describe departures from expected utility theory; it explained them in terms of psychologically plausible mechanisms. Reference dependence reflects the brain's orientation toward detecting changes rather than absolute levels. Loss aversion reflects the evolutionary advantage of reacting more strongly to threats than to opportunities. Probability weighting reflects the psychological salience of certainty and the difficulty of discriminating between similar probability levels.
Decision Theory and Artificial Intelligence
The relationship between decision theory and artificial intelligence has been deeply intertwined since both fields emerged in the mid-twentieth century. Early AI pioneers like John McCarthy, Marvin Minsky, and Herbert Simon drew explicitly on decision-theoretic frameworks, and modern AI systems are built on foundations that trace directly to von Neumann, Savage, and their successors.
Utility Functions and Rational Agents
Modern AI systems, particularly those based on the rational agent paradigm described in Stuart Russell and Peter Norvig's influential textbook Artificial Intelligence: A Modern Approach, are explicitly designed as expected utility maximizers. An AI agent perceives the state of its environment, evaluates possible actions according to a utility function, and selects the action that maximizes expected utility given its beliefs about the current state and the consequences of its actions.
This design choice makes AI agents normatively rational in Savage's sense: they have consistent beliefs (represented by probability distributions), consistent values (represented by utility functions), and make choices that maximize expected utility. The question of whether this is a good design for AI systems depends on whether the normative theory is appropriate for the domains in which AI operates, a question that the descriptive critiques of decision theory make more urgent.
Bayesian Reasoning in AI
Bayesian decision theory, which combines Savage's subjective probability framework with von Neumann and Morgenstern's utility theory, provides the default reasoning framework for many AI systems. Bayesian networks, Bayesian classifiers, and Bayesian reinforcement learning all use probability theory to represent uncertainty and utility theory to evaluate outcomes.
The Bayesian framework's strength is its mathematical elegance and theoretical coherence. Its weakness is computational: in complex real-world environments, exact Bayesian reasoning is often computationally intractable, requiring approximations that may sacrifice some of the framework's theoretical guarantees. This computational challenge echoes Simon's bounded rationality argument: even an artificial agent with perfect information may lack the computational resources to fully optimize, and may need to satisfice using heuristics that approximate rational behavior.
AI Decision-Making and the Alignment Problem
As AI systems become more capable and are deployed in higher-stakes domains (autonomous vehicles, medical diagnosis, financial trading, military operations), the foundations of decision theory become directly relevant to safety and ethics. The alignment problem, ensuring that AI systems pursue goals that are aligned with human values, is fundamentally a decision-theoretic problem: how do you specify a utility function that captures what humans actually want?
The descriptive findings of behavioral economics suggest that this is harder than it might seem. Human preferences are reference-dependent, loss-averse, context-sensitive, and sometimes inconsistent. Translating these messy human preferences into the clean utility functions that AI systems require is an unsolved problem that lies at the intersection of decision theory, philosophy, and artificial intelligence.
| Era | Development | Key Figures | Significance |
|---|---|---|---|
| 1654 | Probability correspondence | Pascal, Fermat | Mathematical reasoning about chance |
| 1738 | Diminishing marginal utility | Daniel Bernoulli | Resolved St. Petersburg paradox; utility vs. money |
| 1921 | Risk vs. uncertainty distinction | Frank Knight | Not all uncertainty involves known probabilities |
| 1944 | Axiomatic utility theory | Von Neumann, Morgenstern | Mathematical foundation for rational choice |
| 1953 | Allais paradox | Maurice Allais | First systematic violation of expected utility |
| 1954 | Subjective expected utility | Leonard Savage | Unified probability and utility from preferences |
| 1961 | Ellsberg paradox | Daniel Ellsberg | Ambiguity aversion violates subjective expected utility |
| 1979 | Prospect theory | Kahneman, Tversky | Comprehensive descriptive alternative to expected utility |
| 1990s-present | AI decision systems | Russell, Norvig, et al. | Decision theory embedded in artificial agents |
The Enduring Questions
Decision theory continues to grapple with several fundamental questions that have resisted resolution since the field's earliest days.
Is rationality a normative ideal or a practical requirement? The axioms of expected utility theory define a concept of rationality that many thoughtful people violate in systematic ways. Does this mean that the axioms need revision, or that people need correction? The debate between Kahneman's view (people are systematically irrational and need help) and Gigerenzer's view (people are ecologically rational, using heuristics well-adapted to their environments) reflects a deep disagreement about the nature and standards of rational decision-making.
How should decision theory handle uncertainty about probabilities? Savage's framework assumes that agents can assign subjective probabilities to all relevant events. But in many real-world situations, we face deep uncertainty where even subjective probability assignment seems arbitrary. Climate change scenarios, pandemic risks, financial system collapses, and existential risks from artificial intelligence all involve uncertainties that resist quantification. Alternative frameworks, including maximin (choose the action whose worst-case outcome is least bad), minimax regret (minimize the maximum difference between your outcome and the best outcome you could have achieved), and robust decision-making (find strategies that perform adequately across a wide range of scenarios), have been proposed for these situations.
How should decision theory incorporate ethical considerations? Classical decision theory is agnostic about values: it tells you how to efficiently pursue whatever goals you have, but it says nothing about what goals you should have. The integration of ethics with decision theory, particularly in contexts like medical allocation, criminal sentencing, and AI governance, remains a fundamental challenge.
What is the relationship between individual and collective decision-making? Decision theory was primarily developed for individual agents, but many of the most important decisions in the world are made by groups, organizations, and institutions. Arrow's impossibility theorem (1951) showed that no voting system can satisfy all reasonable fairness criteria simultaneously, revealing fundamental tensions in collective decision-making that individual decision theory does not address.
"In any moment of decision, the best thing you can do is the right thing, the next best thing is the wrong thing, and the worst thing you can do is nothing." -- Theodore Roosevelt
These questions ensure that decision theory remains a living, evolving field rather than a settled body of doctrine. The mathematical frameworks developed over three centuries provide indispensable tools for analyzing decisions, but the limitations of those frameworks, revealed through paradoxes, experiments, and practical failures, ensure that the search for better theories continues.
Key Researchers and Their Contributions
The formal structure of decision theory was built by a small number of mathematicians, economists, and statisticians working in the mid-twentieth century, often in close intellectual proximity and sometimes in direct response to each other's work.
John von Neumann (1903-1957) was born in Budapest and completed his doctorate in mathematics from Budapest and Zurich simultaneously at age 22. By the time he joined the Institute for Advanced Study in Princeton in 1933, he had already made fundamental contributions to set theory, quantum mechanics, and ergodic theory. His collaboration with economist Oskar Morgenstern on Theory of Games and Economic Behavior (1944) emerged from a chance meeting at Princeton and grew into a three-year project. Von Neumann contributed the mathematical framework while Morgenstern provided the economic applications. Von Neumann's axiomatic derivation of expected utility was one of several contributions in the book; his work on zero-sum games and the minimax theorem proved equally influential. After the war, he became a central figure in the development of the modern computer at the Institute for Advanced Study and served on the Atomic Energy Commission, where he applied decision-theoretic reasoning to nuclear strategy. He died of cancer in 1957, almost certainly contracted from his presence at atomic bomb tests.
Oskar Morgenstern (1902-1977) was an Austrian economist who had directed the Vienna Institute for Business Cycle Research before emigrating to the United States in 1938. His contribution to the von Neumann collaboration was primarily economic and conceptual: he recognized that economics lacked a rigorous mathematical foundation and that game theory could provide one. After the collaboration, Morgenstern spent decades at Princeton applying game theory to economic forecasting and military strategy. He was notably skeptical of economic data quality, arguing in his 1963 book On the Accuracy of Economic Observations that most economic statistics were far too imprecise to support the conclusions economists drew from them.
Leonard Jimmie Savage (1917-1971) completed his PhD in mathematics at the University of Michigan in 1941 and worked at several institutions before settling at the University of Chicago from 1954 to 1960 and then Yale. His 1954 book The Foundations of Statistics extended von Neumann and Morgenstern's framework to situations of genuine uncertainty, where the probabilities of outcomes are not objectively known. Savage's seven postulates for rational preference under uncertainty, which he called "personal probability" theory, provided the mathematical foundation for Bayesian statistics as applied to decision-making. He was reportedly shaken when Maurice Allais demonstrated at the 1952 Paris conference that Savage himself violated his own axioms in a concrete choice problem; Savage initially admitted the violation was genuine before later arguing it represented a preference error he would correct upon reflection.
Maurice Allais (1911-2010) was a French economist who trained as an engineer at the Ecole Polytechnique and spent his career at the Ecole Nationale Superieure des Mines in Paris. His 1952 paper presenting what became the Allais paradox was initially dismissed by American economists as confused, partly because Allais wrote primarily in French and partly because American economics was then dominated by the Chicago School perspective. Allais received the Nobel Prize in Economics in 1988, but his work on the paradox of expected utility was not widely recognized by Anglophone economists until Kahneman and Tversky explicitly acknowledged it as a predecessor to prospect theory.
Daniel Ellsberg (born 1931) completed his Harvard dissertation in economics in 1962, the same year his paper presenting the ambiguity aversion paradox appeared in the Quarterly Journal of Economics. Ellsberg worked for the RAND Corporation applying decision theory to nuclear strategy before joining the Department of Defense. His intellectual trajectory shifted dramatically in 1971 when he leaked the Pentagon Papers to the New York Times, a decision that itself exemplifies the gap between normative decision theory (the expected consequences were severe) and actual human decision-making (Ellsberg acted on moral conviction regardless of personal cost). He was tried under the Espionage Act; the case was dismissed in 1973 due to government misconduct.
Kenneth Arrow (1921-2017) made two foundational contributions to decision theory. His 1951 impossibility theorem, published in Social Choice and Individual Values, proved that no voting system satisfying a small set of reasonable fairness conditions can aggregate individual preferences into a consistent collective preference. This result, which Arrow proved while still in his twenties, established fundamental limits on collective rationality that complement the limits on individual rationality explored by Kahneman and Tversky. Arrow also contributed to general equilibrium theory and the economics of information. He received the Nobel Prize in 1972, the youngest economist to do so at the time.
Historical Case Studies That Changed the Field
Several specific incidents, experiments, and publications serve as landmarks in the intellectual development of decision theory.
The St. Petersburg Problem Correspondence (1713-1738). The chain of intellectual events beginning with Nicolas Bernoulli's 1713 letter to Pierre Raymond de Montmort and culminating in Daniel Bernoulli's 1738 paper in the Commentarii Academiae Scientiarum Imperialis Petropolitanae (the journal of the Imperial Academy of Sciences in St. Petersburg, giving the paradox its name) established the basic tension between expected value and subjective utility that would animate decision theory for three centuries. Daniel Bernoulli's proposal that utility grows as the logarithm of wealth was not merely a solution to a puzzle; it introduced the idea that mathematical models of decision-making should incorporate psychological realism, a program that Kahneman and Tversky would complete two centuries later.
The 1952 Paris Conference. The gathering that preceded the Allais paradox's formal publication was a 1952 colloquium in Paris organized by the Centre National de la Recherche Scientifique. Allais presented choice problems to several leading decision theorists, including Savage and Paul Samuelson, who were attending. When Allais revealed that several participants, including Savage himself, had chosen option combinations violating the independence axiom, the room went quiet. Savage reportedly said he must have made a mistake. This moment is often cited as the first public demonstration that sophisticated, mathematically trained economists could systematically violate their own normative theory in concrete choice situations.
The RAND Corporation and Nuclear Strategy (1948-1960s). RAND, founded in 1948 as a nonprofit research organization primarily funded by the U.S. Air Force, became the crucible in which decision theory was applied to the most consequential choices of the Cold War era. RAND economists including Thomas Schelling, Herman Kahn, and Daniel Ellsberg used game theory and decision theory to analyze nuclear deterrence, arms control negotiations, and crisis management. Schelling's 1960 book The Strategy of Conflict, which applied game theory to international relations, won the Nobel Prize in Economics in 2005. RAND's work demonstrated both the power and the limits of formal decision theory: the framework provided rigorous language for analyzing strategic interaction, but the assumption of rational adversaries was itself questionable in high-stakes, emotionally charged international crises.
Kahneman and Tversky's 1979 Econometrica Publication. The decision to publish prospect theory in Econometrica rather than a psychology journal was deliberate and consequential. Kahneman and Tversky had observed that psychological critiques of economic assumptions were routinely ignored by economists who dismissed them as failing to understand the field. By publishing in the premier economics journal, using the mathematical notation of economics, and directly comparing their model to expected utility theory using the same formal apparatus, they made it impossible for economists to dismiss their work as psychological hand-waving. The paper was accepted after significant revision and immediately recognized as important; it has since become the most cited paper in the history of Econometrica.
The Behavioral Decision Theory Program at the Oregon Research Institute (1970s-1980s). While Kahneman and Tversky worked at Hebrew University, a parallel research program at the Oregon Research Institute in Eugene, led by Paul Slovic and Baruch Fischhoff (who completed his doctorate with Tversky), extended the heuristics and biases program to applied domains including nuclear power risk perception, medical decision-making, and legal judgment. Slovic's work on risk perception demonstrated that lay people and technical experts systematically disagree about which risks are most serious, and that these disagreements follow predictable patterns related to dread, controllability, and catastrophic potential rather than statistical probability. This work directly influenced regulatory policy and environmental law.
How These Ideas Are Applied Today
Decision theory has moved from philosophy and academic economics into engineering systems, medical practice, financial regulation, and artificial intelligence design.
Medical Decision-Making. The Society for Medical Decision Making, founded in 1979, applies formal decision theory to clinical medicine. Researchers including David Eddy at Kaiser Permanente and Gerd Gigerenzer at the Max Planck Institute for Human Development have developed decision aids, expected utility calculations for treatment choices, and analyses of how doctors and patients misunderstand probability and risk. The Center for Informed Medical Decisions (now the Informed Medical Decisions Foundation) has produced over 500 decision aids used in clinical practice to help patients make value-consistent choices about surgery, cancer screening, and chronic disease management. Research by Barry and colleagues published in JAMA Internal Medicine in 1995 showed that men who used a decision aid for prostate cancer screening chose differently from those who received standard physician counseling, demonstrating that the process of decision support changes outcomes.
Financial Regulation. The 2010 Dodd-Frank Act in the United States and the EU's MiFID II directive (2018) incorporated behavioral decision theory insights into financial regulation. Requirements for standardized risk disclosure, cooling-off periods for financial products, and restrictions on complex financial instruments all reflect the regulatory conclusion that consumers cannot make fully rational decisions about complex financial products. The U.K. Financial Conduct Authority established a Behavioural Economics and Data Science unit that uses prospect theory and related frameworks to analyze how financial product design exploits cognitive biases and to design regulatory interventions accordingly.
Artificial Intelligence System Design. The rational agent architecture described in Stuart Russell and Peter Norvig's textbook Artificial Intelligence: A Modern Approach, used in virtually every AI course worldwide, directly embeds von Neumann-Morgenstern expected utility theory. Planning algorithms, reinforcement learning systems, and game-playing AI all use some form of expected utility maximization. The emerging field of AI alignment, which addresses how to ensure AI systems pursue goals consistent with human values, is fundamentally a decision theory problem: how do you specify a utility function that captures what humans actually want, given that human preferences are reference-dependent, inconsistent, and context-sensitive in ways that prospect theory describes? Organizations including the Machine Intelligence Research Institute, the Center for Human-Compatible Artificial Intelligence (Stuart Russell's group at Berkeley), and DeepMind's safety team are working on this problem, drawing directly on the literature on normative and descriptive decision theory.
Behavioral Insights in Public Policy. The application of decision theory to policy design has moved beyond nudging into structural policy reform. The U.S. Pension Protection Act of 2006 embedded the decision-theoretic insight about default power into law. The UK's Automatic Enrolment pension reform (2012), designed with input from the Behavioural Insights Team, automatically enrolled 10 million workers into workplace pensions by 2018, increasing pension participation from 55% to 78% among eligible workers. The OECD's Behavioural Insights and Public Policy project, launched in 2014, coordinates the application of decision theory insights across member governments, documenting experiments in domains from tax compliance to energy efficiency.
References and Further Reading
Von Neumann, J. & Morgenstern, O. (1944). Theory of Games and Economic Behavior. Princeton University Press. https://press.princeton.edu/books/paperback/9780691130613/theory-of-games-and-economic-behavior
Savage, L. J. (1954). The Foundations of Statistics. John Wiley & Sons. https://doi.org/10.1002/nav.3800010316
Kahneman, D. & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econometrica, 47(2), 263-291. https://doi.org/10.2307/1914185
Bernoulli, D. (1738/1954). Exposition of a new theory on the measurement of risk. Econometrica, 22(1), 23-36. https://doi.org/10.2307/1909829
Allais, M. (1953). Le comportement de l'homme rationnel devant le risque. Econometrica, 21(4), 503-546. https://doi.org/10.2307/1907921
Ellsberg, D. (1961). Risk, ambiguity, and the Savage axioms. Quarterly Journal of Economics, 75(4), 643-669. https://doi.org/10.2307/1884324
Knight, F. H. (1921). Risk, Uncertainty, and Profit. Houghton Mifflin. https://oll.libertyfund.org/titles/knight-risk-uncertainty-and-profit
Russell, S. & Norvig, P. (2020). Artificial Intelligence: A Modern Approach (4th ed.). Pearson. https://aima.cs.berkeley.edu/
Arrow, K. J. (1951). Social Choice and Individual Values. John Wiley & Sons. https://cowles.yale.edu/arrow-social-choice-and-individual-values
Simon, H. A. (1955). A behavioral model of rational choice. Quarterly Journal of Economics, 69(1), 99-118. https://doi.org/10.2307/1884852
Hacking, I. (2006). The Emergence of Probability: A Philosophical Study of Early Ideas About Probability, Induction and Statistical Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511817557
Peterson, M. (2017). An Introduction to Decision Theory (2nd ed.). Cambridge University Press. https://doi.org/10.1017/9781316585061
Gilboa, I. (2009). Theory of Decision Under Uncertainty. Cambridge University Press. https://doi.org/10.1017/CBO9780511840203
Resnik, M. D. (1987). Choices: An Introduction to Decision Theory. University of Minnesota Press. https://www.upress.umn.edu/book-division/books/choices
Gigerenzer, G. & Selten, R. (Eds.). (2001). Bounded Rationality: The Adaptive Toolbox. MIT Press.
Thaler, R. H. & Sunstein, C. R. (2008). Nudge: Improving Decisions About Health, Wealth, and Happiness. Yale University Press.
Laplace, P.-S. (1814/1951). A Philosophical Essay on Probabilities. Dover Publications.
Kahneman, D. (2011). Thinking, Fast and Slow. Farrar, Straus and Giroux.
Empirical Studies That Forced Decision Theory to Revise Its Foundations
Decision theory's normative framework was repeatedly confronted by experimental findings that could not be explained away as mere error or confusion. The most important of these findings became permanent fixtures in the field's intellectual landscape.
Paul Slovic's Research on Risk Perception and Expert Disagreement (1979-1990). Paul Slovic at Decision Research in Eugene, Oregon, conducted a systematic research program examining how laypeople and technical experts perceive risk, producing findings that challenged the decision theory assumption that risk assessments are driven primarily by probability times magnitude calculations. In a 1979 paper in Science, "Rating the Risks," co-authored with Baruch Fischhoff and Sarah Lichtenstein, Slovic's team asked 76 college students, 40 League of Women Voters members, and 15 business and professional club members to rank 30 hazards by riskiness. The groups consistently ranked nuclear power as the highest risk -- higher than motor vehicles (which cause far more deaths annually) and handguns -- while a comparison group of technical experts (risk specialists) ranked nuclear power 20th. The psychometric analysis revealed that lay risk judgments were primarily driven by two factors: "dread" (the degree to which a hazard is uncontrollable, catastrophic, and fatal) and "unknown risk" (the degree to which a hazard is new, unfamiliar, and delayed in its effects). Nuclear power scored extremely high on both dimensions; motor vehicle accidents scored low on both despite causing far more statistical deaths. This finding was practically consequential for regulatory policy: the Nuclear Regulatory Commission, which at the time used expected-death calculations to prioritize safety investments, commissioned follow-up research and eventually incorporated public risk perception alongside technical risk assessment in its regulatory framework. Slovic's subsequent work, including a 1987 Science paper "Perception of Risk," documented the same pattern across 15 countries and across professional groups including physicians, judges, and members of Congress, establishing that the gap between expert and lay risk assessment was not a matter of ignorance but of different weighting of risk dimensions that were both legitimate and persistent.
Kahneman, Knetsch, and Thaler's Endowment Effect Field Experiments (1990-1991). The endowment effect -- the tendency to value objects more once you own them -- was demonstrated in laboratory settings by Kahneman and Thaler in the 1980s, but its practical significance became clearest in a series of field experiments. Jack Knetsch at Simon Fraser University conducted a pivotal study in 1989, published in the Journal of Economic Behavior and Organization, using three experimental groups: one group was given a coffee mug and offered the opportunity to trade it for a chocolate bar; a second group was given a chocolate bar and offered the opportunity to trade it for a mug; a third group was simply asked to choose between a mug and a chocolate bar. Standard economic theory predicted similar preferences across all three groups (approximately 50% choosing each in the choice group, with mug recipients trading at roughly the same rate). Instead, approximately 89% of mug recipients kept the mug, and 90% of chocolate bar recipients kept the chocolate bar -- the item received was retained regardless of which item subjects would have chosen absent ownership. Kahneman, Knetsch, and Thaler's 1990 Journal of Political Economy paper extended these findings using real monetary exchanges in classroom settings, showing that minimum willingness to accept (WTA) for surrendering a mug averaged $5.78 while maximum willingness to pay (WTP) for acquiring the same mug averaged only $2.21 -- a WTA/WTP ratio of 2.6, consistent across seven experiments. This WTA/WTP gap has since been documented in over 200 studies and across multiple domains including environmental goods, health outcomes, and financial assets. A 2005 meta-analysis by Horowitz and McConnell in the Journal of Economic Behavior and Organization, reviewing 201 studies, found a mean WTA/WTP ratio of 7.2, with the ratio being larger for public goods and non-market commodities than for private market goods -- suggesting that the endowment effect is strongest precisely where decision theory is most often applied to policy analysis.
Amos Tversky and Derek Koehler's Support Theory and Probability Judgment (1994). Amos Tversky, working with Derek Koehler in the final years before his death in 1996, developed support theory to explain a systematic pattern in probability judgment that had emerged across dozens of studies: people consistently assign higher probability to an event when it is described in detail than when it is described in summary form, even when the descriptions are logically equivalent. In a study published in Psychological Review in 1994, Tversky and Koehler asked subjects to estimate the probability that a randomly selected Stanford student had died from a particular cause in the previous year. When "death by natural causes" was broken down into specific categories (heart attack, cancer, stroke, other natural causes), the probabilities assigned to each specific category summed to 73% of all expected deaths -- while when "death by natural causes" was presented as a single category, it received an assignment of only 42% of all expected deaths. Unpacking a category into components reliably increased its judged probability, a finding that violated classical probability theory's requirement that the probability of a category equal the sum of probabilities of its mutually exclusive components. The practical implications were significant: prosecutors and lawyers who presented detailed narratives of one interpretation of events would predictably increase jurors' assessment of that interpretation's probability relative to opposing interpretations described in more summary form, not because the evidence was stronger but because the description was more explicit. Support theory influenced subsequent work on narrative persuasion by Jonathan Haidt at NYU and on legal decision-making by Norbert Kerr and colleagues, becoming a bridge between abstract decision theory and applied judgment research.
Decision Theory Applied to Catastrophic and Irreversible Choices
The most consequential applications of decision theory -- nuclear deterrence, climate policy, pandemic preparedness -- involve decisions under radical uncertainty where standard expected utility calculations face serious objections.
Thomas Schelling's Analysis of Nuclear Deterrence (1960-1980). Thomas Schelling at RAND and Harvard applied game-theoretic decision theory to nuclear deterrence in a research program that won him the Nobel Prize in Economics in 2005. His 1960 book The Strategy of Conflict demonstrated that the effectiveness of nuclear deterrence depended not on the capability to destroy an adversary but on the credibility of the threat to do so -- a distinction that classical decision theory treated as irrelevant but that was central to the strategic interaction. Schelling's analysis of "commitment devices" showed how nations could make threats credible precisely by limiting their own freedom of action: deploying conventional troops in a forward position (the "tripwire" strategy in West Germany) committed the United States to respond to a Soviet invasion because failure to respond would mean the death of American soldiers, not because American decision-makers calculated that nuclear war was the utility-maximizing response. This insight -- that decision-making effectiveness sometimes requires precommitting to actions that would not be optimal at the time they are executed -- directly applied to the design of nuclear postures, arms control agreements, and crisis management protocols. Schelling's work on focal points (the "Schelling point" concept, where people who cannot communicate still converge on the same solution because it is culturally salient) challenged the assumption that rational agents with common knowledge would always achieve optimal coordination, showing instead that coordination relied on shared cultural context that formal decision theory could not represent.
Howard Raiffa's Decision Analysis at the Harvard Kennedy School (1968-1985). Howard Raiffa, Schelling's Harvard colleague, developed "decision analysis" as a practical methodology for applying decision theory to organizational and policy decisions, creating an applied tradition that translated abstract theory into consulting practice. His 1968 textbook Decision Analysis: Introductory Lectures on Choices Under Uncertainty, developed from a Harvard course Raiffa taught to MBA students, provided the first systematic framework for eliciting subjective probabilities, constructing decision trees, and calculating expected utilities in complex multi-stage decisions. Raiffa applied these methods to a series of high-stakes cases: advising the US Army Corps of Engineers on dam siting decisions that involved environmental, economic, and safety trade-offs; analyzing decisions about offshore oil drilling; and providing analytical support for international environmental negotiations. His most significant applied contribution was the Program on Negotiation at Harvard Law School, which he co-founded in 1983 with Roger Fisher and William Ury. The negotiation research program developed a decision-analytic approach to bargaining -- identifying parties' interests and reservation prices through decision tree analysis -- that influenced international negotiations including the 1978 Camp David Accords (where Jimmy Carter's team used interest-based analysis) and the Montreal Protocol negotiations on ozone depletion (1987). Raiffa's student David Bell at Harvard Business School subsequently documented that negotiators who used formal decision-analytic preparation tools reached agreements that were significantly closer to the Pareto frontier (the boundary of outcomes where no party can be made better off without making another worse off) than negotiators using informal judgment, in a series of experimental studies published in Management Science from 1988 to 1995.
The IPCC's Handling of Deep Uncertainty in Climate Projections (1990-2021). The Intergovernmental Panel on Climate Change, established in 1988, faced a decision theory problem of unprecedented scale: how to communicate probabilistic uncertainty in climate projections to policymakers who needed to make irreversible infrastructure investment decisions with 50-100 year time horizons. The IPCC's first Assessment Report (1990) described uncertainties informally and inconsistently, with different contributing scientists using "likely," "probable," and "uncertain" to mean different things. The resulting miscommunication contributed to political disputes about what the science actually said. The IPCC's Fourth Assessment Report (2007) introduced a formal calibrated uncertainty language, specifying that "very likely" meant greater than 90% probability, "likely" meant greater than 66%, and "more likely than not" meant greater than 50%. A study by Budescu, Broomell, and Por at Cornell, published in Climatic Change in 2009, surveyed 223 respondents on their interpretation of these terms and found systematic misinterpretation: readers consistently interpreted "very likely" as meaning approximately 71% probability rather than the intended 95%, systematically underestimating the confidence the scientific community was expressing. The IPCC revised its communication standards for the Fifth Assessment Report (2013) in response, adding explicit numeric probability ranges alongside verbal descriptors. The episode illustrated a general challenge for decision theory applications to public policy: probability formats that are technically precise (subjective expected utility theory requires numerical probabilities) may communicate less accurately to real audiences than qualitative descriptions, creating a tension between decision theory's normative requirements and the practical demands of democratic communication.
Frequently Asked Questions
When did decision theory emerge as a formal discipline?
Decision theory crystallized in the 1940s-50s with von Neumann and Morgenstern's game theory and Savage's foundations of statistics, though earlier work by Bernoulli and others laid groundwork.
What is the difference between normative and descriptive decision theory?
Normative decision theory prescribes how rational agents should make decisions, while descriptive theory studies how people actually make choices. This gap led to behavioral economics.
How did prospect theory challenge classical decision theory?
Kahneman and Tversky's prospect theory showed people systematically violate expected utility theory through loss aversion, reference dependence, and probability weighting—decisions are often predictably irrational.
What role did World War II operations research play?
Military operations research drove practical decision analysis for resource allocation, logistics, and strategy. This pragmatic context pushed theoretical developments toward applied decision-making frameworks.
How does decision theory relate to artificial intelligence?
AI builds on decision theory through utility functions, Bayesian reasoning, and game-theoretic models. Modern AI systems use decision-theoretic frameworks for planning and action selection.
What is the St. Petersburg paradox and why does it matter?
This 18th-century paradox showed expected value doesn't fully capture rational choice, leading to utility theory. It revealed that people consider diminishing marginal value, not just mathematical expectation.