Hostname: page-component-78c5997874-s2hrs Total loading time: 0 Render date: 2024-11-11T18:03:38.411Z Has data issue: false hasContentIssue false

How Should Risk and Ambiguity Affect Our Charitable Giving?

Published online by Cambridge University Press:  16 May 2023

Lara Buchak*
Affiliation:
Princeton University, Princeton, NJ, USA
Rights & Permissions [Opens in a new window]

Abstract

Suppose we want to do the most good we can with a particular sum of money, but we cannot be certain of the consequences of different ways of making use of it. This article explores how our attitudes towards risk and ambiguity bear on what we should do. It shows that risk-avoidance and ambiguity-aversion can each provide good reason to divide our money between various charitable organizations rather than to give it all to the most promising one. It also shows how different attitudes towards risk and ambiguity affect whether we should give to an organization which does a small amount of good for certain or to one which does a large amount of good with some small or unknown probability.

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press

1. Introduction

Suppose you have 200 dollars, and want to use it to do the most good you can. Some ways you could use it are nearly certain to result in more total good: for example, you could donate it to your local food bank, which will almost certainly result in fewer people going hungry. Other ways you could use it may result in more good, or may result in almost no improvement at all: for example, if your area has a 30% chance of a major earthquake in the next 50 years, you could donate it to an organization that distributes earthquake preparedness kits; if an earthquake happens, donating to this organization will do substantially more good than donating to the food bank will, but if not, doing so will do substantially less good. Still other ways to use the money may or may not result in a change to the overall good, but you may not be able to put a precise probability on this: for example, you could donate it to a highly experimental start-up, which may succeed in reducing poverty, but may fail, where it's difficult to determine the probability of success.

The same question reappears at the level of national and international aid. We must choose between funding various types of causes, each of which may do various amounts of good with various probabilities; in addition, we may not always be able to determine the probabilities precisely. Even when we make choices among programs that address the same problem, we may not know which of two programs or organizations is apt to be more successful.

This article concerns how to make choices whose sole purpose is to do the most good, when you cannot be certain of the consequences of the alternatives. When you are uncertain of the consequences of various alternatives, you cannot choose precisely to bring about the most good, but instead must choose to bring about the best gamble among possibilities with varying levels of goodness. It is typically assumed that to choose the best gamble is to maximize expected utility; but recent work has challenged the idea that this is the correct framework for rational and moral decision-making. In particular, some have argued contra this framework that we can be risk-avoidant or risk-inclined and we can be ambiguity-averse or ambiguity-seeking. The goal of this article is to determine how these attitudes make a difference to which causes we should give to.

We will focus on two questions. First, when we give money purely with the intent of bringing about the best gamble, to what extent should we diversify among the programs that we give to? Second, when we're choosing between giving to a program that has a known, high probability of doing good (for example, a health amelioration program) and another program that has a very small or perhaps unknown probability of doing much more good (for example, a program that makes the already-small and perhaps unknown probability of nuclear annihilation even smaller), which should we choose? We will see how our attitudes towards risk and ambiguity affect the answers to each of these questions.

2. Expected utility and extensions

Readers who are already familiar with expected utility, risk-weighted expected utility, and α-maximin can skip to section 3.

2.1 Expected utility

The most well-known formal theories of decision-making under uncertainty tell us how to evaluate acts, where an act is an assignment of consequences to states of the world.Footnote 1 For example, the act of going to the beach assigns consequence <I have an unpleasant time> to the state in which the weather is cold, <I have an okay time> to the state in which it is warm, and <I have a great time> to the state in which it is hot. We can represent this act as {COLD, I have an unpleasant time; WARM, I have an okay time; HOT, I have a great time}. Each consequence has a utility value; we can thus write these acts as assignments of utility values to states of the world, for example {COLD, 2; WARM, 3; HOT, 6}.Footnote 2

In the classical theory, we can additionally assign sharp probabilities to states of the world; for example, we might assign probability 0.2 to the weather being cold, 0.4 to it being warm, and 0.4 to it being hot, so that we can represent the act as {2, 0.2; 3, 0.4; 6, 0.4}. It is rational to choose or prefer acts with the highest expected utility, where expected utility is a weighted average of the utilities of each consequence, each utility value weighted by the probability of the state in which it obtains.Footnote 3 Graphically, if the height of each bar represents utility, and its width probability, expected utility is the area under the curve in Figure 1.Footnote 4 In this case, the value of the gamble ‘go to the beach’ will be (0.2)(2) + (0.4)(3) + (0.4)(6) = 4.

Figure 1. Expected utility.

Formally, let ${\rm \succeq }$ represent some individual's preference relation, with ≻ and ≈ defined in the usual way. Then let g = {E1, x1; …; En, xn} be a gamble that yields outcome xi in event Ei. Let p(Ei) be the subjective probability of Ei, and let u(xi) be the utility of xi. The expected utility of g is:

(1)$$EU( g ) = \mathop \sum \limits_{i = 1}^n p( {E_i} ) u( x_i) $$

One way to think about expected utility conceptually is this. An act has possible consequences, and the amount each consequence matters in determining the total value of the act – the weight each consequence gets in evaluating the act – is its probability. For example, the act above has three possible utility consequences (2, 3, 6). The possibility of utility 2 gets weight 0.2 in the overall evaluation of the act, the possibility of utility 3 gets weight 0.4, and the possibility of utility 6 gets weight 0.4.

However, instead of thinking of the act as having three possible utility consequences, we can instead think of it as having three possible utility levels one might achieve: one might get at least utility 2; one might get at least utility 3 (an additional utility of 1 beyond the prior level); or one might get utility 6 (an additional utility of 3 beyond the prior level). In Figure 2, the height of each bar represents the utility achieved beyond the prior level, and the width represents the probability of achieving at least that utility.

Figure 2. Expected utility, in terms of utility levels.

Conceptualized in this way, the act has three features to take into account: with probability 1, you will get at least utility 2; with probability 0.8, you will get at least 1 additional utility; and with probability 0.4, you will get 3 additional utility. Again, the weight of each of these considerations is its probability, so again, the value of ‘go to the beach’ will be (2) + (0.8)(3 − 2) + (0.4)(6 − 3) = 4.

Thus, the expected utility of g is:

(2)$$EU( g ) = \mathop \sum \limits_{i = 1}^n \left[{\left({\mathop \sum \limits_{\,j = i}^n p( {E_j} ) } \right)( {u( {x_i} ) -\;u( {x_{i-1}} ) } ) } \right]$$

where we stipulate that u(x0) = 0. This is mathematically equivalent to Equation 1, but conceptualized in this second way.

This is the classical theory. Many think, however, that this theory presents too limited a picture of rational decision-making, because it requires a particular attitude towards risk,Footnote 5 or because it requires sharp probability values. I will take each of these complaints, and the extensions designed to address them, in turn.

2.2 Attitudes towards risk: risk-weighted expected utility

Begin with risk, and return to the idea that there are three features one has to take into account when evaluating the example act – the possibility of getting at least utility 2, the possibility of getting at least 1 additional utility above that, and the possibility of getting at least 3 additional utility above that. I have argued elsewhere that there isn't a uniquely rational way to weight each of these considerations.Footnote 6 If I care a lot about making sure things go as well as they can if things turn out disfavorably, and not as much about what happens if things turn out favorably, then the fact that I have probability 0.8 of getting an additional 1 utility beyond the minimum might not get a weight of 0.8; instead, it might get a lower weight, say, 0.64. And the fact that I have a probability 0.4 of getting an additional 3 utility more might not get a weight of 0.4; instead, it might get a weight of, say, 0.16 (Figure 3).

Figure 3. Risk-weighted expected utility, with risk-avoidance.

Thus, the value of ‘go to the beach’ will be (2) + (0.64)(3 − 2) + (0.16)(6 − 3) = 3.12.

Conversely, if I care a lot about what happens if things turn out favorably, but not as much about what happens if they don't, then each utility increment might get more weight than its probability.

The basic idea is to introduce a function of probabilities – a risk function – that measures how much you take the top p-portion of consequences into account in evaluating an act. (This is an example of a rank-dependent model: a model in which the weight of a consequence depends on its position or ‘rank’ relative to other possible consequences of the gamble; it is also a generalization of Quiggin's anticipated utility model to subjective probabilities.Footnote 7) Graphically, the risk function ‘stretches’ or ‘shrinks’ the rectangles, and we call the new area under the curve risk-weighted expected utility (REU). As long as your risk function obeys a few basic constraints,Footnote 8 then rationality requires only that you maximize risk-weighted expected utility, relative to your risk function. In contrast, recall that according to the classical theory, the sizes of the rectangles are fixed – any stretching or shrinking would be an instance of irrationality.

Formally, let g’ = {E1, x1; …; En, xn} be an ordered gamble that yields outcome xi in event Ei, where$\;x_1{\rm \preceq } \ldots {\rm \preceq }x_n$. (We use g’ to denote the fact that every gamble g can be re-ordered in this way.) The risk-weighted expected utility of g’ is then:

(3)$$REU( {g^{\prime}} ) = \mathop \sum \limits_{i = 1}^n \left[{r\left({\mathop \sum \limits_{\,j = i}^n p( {E_j} ) } \right)( {u( {x_i} ) -\;u( {x_{i-1}} ) } ) } \right]$$

where r is a “risk function” from [0, 1] to [0, 1], with r(0) = 0, r(1) = 1, and r non-decreasing (or possibly required to be increasing), and we stipulate that u(x0) = 0. This is just Equation 2, with the risk function applied to probabilities. Notice that there may be multiple ordered representations of the same unordered gamble, due to ties. All ordered representations of a given unordered gamble will yield the same risk-weighted expected utility.

There will thus be several types of rational decision-makers. For some decision-makers, utility levels that are attained only in smaller and smaller portions of states matter proportionally less and less. These decision-makers have a strictly convex risk function – rectangles get shrunk more and more as their width gets smaller and smaller – and are thus strictly risk-avoidant. Other decision-makers are the opposite: utility levels that are attained only in smaller and smaller portions of states matter proportionally more and more; they have a strictly concave risk function – rectangles get stretched more and more as their width gets smaller and smaller – and are thus strictly risk-inclined. (When I use the term ‘risk-avoidant’ and ‘risk-inclined’ in what follows, I will mean strictly risk-avoidant and strictly risk-inclined unless otherwise specified.) Others are classical expected utility maximizers: they have a linear risk function, and utility increments matter in proportion to their probabilities; these decision-makers are thus globally neutral. One could, of course, have a risk function that fits none of these profiles, such as one according to which utility levels that are attained in a very large and very small portion of states both matter proportionally more than those attained in a middling portion of states.

Let us relate this back to the general idea of being risk-averse. To be risk-averse in a good (e.g., money) is to prefer gambles in which the amounts of that good you might obtain are not ‘spread out’: you'd rather have 50 units of that good than, say, a 50/50 chance of 25 units or 75 units. More generally, let a mean-preserving spread be a transformation that takes some probability mass from the ‘center’ of the gamble and moves it toward both extremes while keeping the mean quantity of the good the same, and a mean-preserving contraction to be the reverse. To be (strictly) risk-averse is to (strictly) prefer mean-preserving contractions: for all lotteries f and g where f can be transformed into g by a series of mean-preserving spreads, f is (strictly) preferred to g.Footnote 9 A special case of this is preferring a sure thing to a lottery whose mean value is equal to the value of the sure thing (for example, preferring 50 units of a good to the 50/50 chance of 25 units or 75 units).

The classical theory recognizes that many people are risk-averse in money, and holds that all rational risk-aversion can be captured by the shape of the utility function. If an individual's utility function diminishes marginally (her utility function is concave) – if each dollar adds less utility than the previous dollar – then if she is an expected utility maximizer, she will be risk-averse. And vice versa: if she is risk-averse and an expected utility maximizer, then her utility function will diminish marginally.Footnote 10 But there is an alternative reason that an individual might prefer acts that are less spread out in possible amounts of a good, namely that she cares more about worse consequences of a gamble than better consequences: she is risk-avoidant (her risk function is convex).Footnote 11 REU maximization allows for an agent to be risk-averse in either way or both ways. In addition, REU maximization allows that one can be risk-averse in utility – one can prefer mean-preserving contractions in utility. (In the classical theory, a rational individual by definition is indifferent among all gambles with the same mean utility value.) Indeed, if an individual is risk-avoidant then she will be risk-averse in utility, and if she is risk-averse in utility then she will be risk-avoidant.Footnote 12

2.3 Attitudes Towards Ambiguity: α-maximin

A different phenomenon has also given rise to extensions of the classical theory: ambiguity. So far we've been talking about cases in which an individual can assign ‘sharp’ probabilities to the various possibilities, but many philosophers believe that being able to assign sharp probabilities is not a requirement of rationality, or is even contrary to rationality.Footnote 13 Furthermore, many people do in fact fail to assign sharp probabilities to states, and choose differently when they do not assign sharp probabilities from when they do.Footnote 14

For example, I might be unsure about the probability of cold, warm, and hot weather – perhaps the forecaster can only give me a range, or different forecasters give different numbers and I don't simply average them.Footnote 15 Instead of assigning 0.2 to cold weather, 0.4 to warm weather, and 0.4 to hot weather, I might assign a set of probability functions, for example a set that includes p1 = {0.1, COLD; 0.3, WARM; 0.6, HOT}, p2 = {0.3, COLD; 0.5, WARM; 0.2, HOT}, and mixtures between them such as p3 = {0.2, COLD; 0.4, WARM; 0.4, HOT}. Alternatively, I might assign interval ranges to each event: for example, a range of [0.1, 0.3] to cold weather, a range of [0.3, 0.5] to warm weather, and a range of [0.2, 0.6] to hot weather.Footnote 16

What should someone choose when she cannot assign sharp probabilities but she instead assigns sets or intervals?

Three types of decision theories have been particularly popular in the literature.Footnote 17 The first type – which includes Γ-maximin, Γ-maximax, α-maximin, and ‘best-estimate’ models – operates on sets of probabilities. These models calculate expected utility for each probability function in the set, and then hedge between some of these EU values in some way. The second type – which includes Choquet expected utility – operates on probability intervals: it first selects a probability from each interval, and then uses this probability to calculate EU. (It is a rank-dependent model, and an ancestor of REU.) The third type – which includes maximality, E-admissibility, and interval dominance – provides criteria for when an act may be selected from a given menu of acts, but doesn't characterize one's attitude towards ambiguity. I will concentrate on α-maximin – the most popular model of the first type – because this framework appears to be most well-known among philosophers, but this shouldn't be taken to indicate a preference for that theory or for theories of its type.Footnote 18

We can apply α-maximin when uncertainty is represented by a set of probability functions.Footnote 19 Each probability function gives rise to an EU value for each gamble. For example, the EU of ‘go to the beach’ according to p1 is (0.1)(2) + (0.3)(3) + (0.6)(6) = 4.7; and the EU according to p2 is (0.3)(2) + (0.5)(3) + (0.2)(6) = 3.3; and the EU according to p 3 is (0.2)(2) + (0.4)(3) + (0.4)(6) = 4. A person who is maximally ambiguity averse will value the act at its minimum expected utility (3.3), and a person who is maximally ambiguity seeking will value the act at its maximum expected utility (4.7). More generally, a person with attitude α will hedge between these values, assigning weight α to the maximum EU and weight (1 − α) to the minimum EU.

Formally, let ${\cal P}$ be a set of probability functions. The α-maximin of g is:

(4)$$\alpha \hbox{-}{\rm maximin}( g ) = ( {1-\alpha } ) [ {{\min }_{{\rm p}\in \cal P}{\rm E}{\rm U}_{\rm p}( g ) } ] + \alpha [ {{\max }_{{\rm p}\in \cal P}\;{\rm E}{\rm U}_{\rm p}( {\rm g} ) } ] $$

If α < 0.5 then the decision-maker is ambiguity averse, if α > 0.5 then she is ambiguity seeking, and if α = 0.5 then she is ambiguity neutral. For example, if α = 0.25, then the value of ‘go to the beach’ will be (0.25)(4.7) + (0.75)(3.3) = 3.65.

Pictorially, we can think of α-maximin as considering two separate ‘standard’ EU graphs and putting more or less weight on the deliverance of each (Figure 4).

Figure 4. α-maximin.

Attitudes towards risk and attitudes towards ambiguity are conceptually independent. Your attitude towards risk is about how much to weight a consequence whose probability is given. Your attitude towards ambiguity is about what you should do when you are not sure what the true probability is. Thus, it is possible to have any combination of risk-attitudes and ambiguity-attitudes; and we can account for both simultaneously. To combine REU and α-maximin, we can calculate the REU value according to each probability distribution, and then hedge between the minimum and maximum REU, giving the maximum REU weight α, and giving the minimum REU weight 1 − α. We can similarly combine REU with other models of resolving ambiguity by using REU in place of EU.

3. Diversification and hedging

In ethics, we are interested in the good or in moral value, rather than in preferences. But we can make use of the same formal frameworks to rank gambles about the good, since the question of what is morally better under uncertainty has the same structure as the question of what is preferable under uncertainty: what is the relationship between gambles about moral value and their constituent moral values? There are two ways to make use of the formal frameworks about preference to think about moral value, the first of which treats the moral value of outcomes as something to which we have independent access, and the second of which doesn't. The first way is to assume that in moral decisions, the content of a person's preferences is the good: her utility function is equivalent to the moral value function, and she should maximize moral value according to her attitudes towards risk and ambiguity. The second way is to assume that the moral value function is the utility function that represents the preferences of a perfectly moral decision-maker, where these preferences obey a set of axioms that also represent this decision-maker's attitude towards risk and ambiguity.Footnote 20 On either picture, we can use the terms ‘utility’ and ‘moral value’ interchangeably, and ≽ becomes the moral value relation.

We can now apply the idea that there are non-EU attitudes towards both risk and ambiguity to two important questions about charitable giving. The first question: should someone who is concerned only with choosing the best gamble give all of her money to one charity, or should she instead diversify over a number of charities?Footnote 21

We will start by assuming that the status quo is ‘flat’ – contains no risk or ambiguity – in order to see how risk attitudes and ambiguity attitudes make a difference in this simple case. We will then examine the upshots for more realistic cases.

To make things concrete, let us assume you have $200 to distribute to two charities, A and B, and your options are to give $200 to Charity A, to give $200 to Charity B, or to give $100 to each charity. You are interested only in doing what results in the most total good. Let us assume further that the good (represented by utility) is linear in money – each $100 increases the good just as much as the previous $100 – but you have some uncertainty about the effectiveness of each charity: either every $100 donation will increase the good by 1 or every $100 will increase the good by 2. (The uncertainty could come about for a number of reasons, including both uncertainty about the empirical facts and normative uncertainty.) We can represent your choice as among the following three gambles, where ‘SQ’ stands for the status quo, ‘E’ for the event in which each $100 to Charity A will increase the total utility by 2, and ‘F’ for the event in which each $100 to Charity B will increase the total utility by 2:

Give $200 to A (‘2A’): {, u(SQ) + 2; E, u(SQ) + 4}

Give $200 to B (‘2B’): {, u(SQ) + 2; F, u(SQ) + 4}

Give $100 to each (‘AB’): {E̅F̅, u(SQ) + 2; E̅F v EF̅, u(SQ) + 3; EF, u(SQ) + 4}

Since the status quo is flat, it has a single value, u(SQ). If the status quo is itself a gamble, then we can represent it as a single value ‘EU(SQ)’ in expected utility calculations without error, but we cannot do so in the calculations for other theories.

Let us start with the expected utility maximizer who assigns sharp probabilities to E and F. There are only three possibilities here: if p(E) > p(F), then she should strictly prefer 2A to the other options; if p(E) < p(F), then she should strictly prefer 2B; and if p(E) = p(F), then she should be indifferent between all three options:

EU(2A) = u(SQ) + 2 + 2p(E)

EU(2B) = u(SQ) + 2 + 2p(F)

EU(AB) = u(SQ) + 2 + p(E) + p(F)

Notice that for an EU maximizer, the utility value of two gambles – when each gamble is formulated in terms of utility increases to the status quo – is just the sum of their utility values. So whichever gamble is better by itself will still be better in conjunction with another copy of itself. Thus, an EU-maximizer can only prefer diversification (AB) to both other options if taking one copy changes the expected utility value of the two options: if (contrary to the setup here) the good for each dollar spent decreases marginally or if (again contrary to the setup here) the ‘probabilistic good’ of each dollar spent decreases marginally, for example, if the project is an all-or-nothing one for which each additional bit of money increases the probability of success less than the last bit, or if (again contrary to the setup here), as Snowden (Reference Snowden, Greaves and Plumer2019) points out, donations to one charity affect the amount of good that the other charity can do.Footnote 22

What about the risk-avoidant person? Let's take a simple case in which the probability of success is the same for both charities (p(E) = p(F)). If E and F are perfectly correlated, that is, they overlap entirely, then all three acts will be equivalent lotteries – they will yield the same utilities with the same probabilities – so the risk-avoidant person will be indifferent among them. But if they are not perfectly correlated, then the risk-avoidant person will prefer AB to both alternatives;Footnote 23 and she will prefer AB by more the less E and F overlap. She will prefer AB by the most when E and F are anti-correlated, that is, when E obtains only if F does not.Footnote 24 To see this, notice that when p(E̅F) = p(EF̅), AB is a mean-preserving contraction of both 2A and 2B, and AB with less correlation between E and F (i.e., higher values of p(E̅F) and p(EF̅)) is a mean-preserving contraction of AB with more correlation between E and F. (To put the point graphically: 2A and 2B concentrate higher utility in states that are less likely, and AB concentrates not-as-high utility in states that are more likely; the former states will get proportionally less weight – the horizontal rectangle will be shrunk more – than the latter.) The risk-avoidant person would rather hedge her bets on only one of E and F holding than go all-in on either E or F.Footnote 25

It's worth pointing out that for most charities, the conditions for their success will not be perfectly correlated. And indeed, some will be anti-correlated, or close to it. For example, the success of Charity A and the success of Charity B might depend on empirical claims known to be opposed. Consider research for a treatment for Alzheimer's or cancer, where the success of various research programs depends on the causal mechanism behind these diseases. Or the success of Charity A and Charity B may depend on opposing normative claims: one might extend life without improving its quality while the other might improve the quality of life without extending it, where we are unsure whether a longer life or a higher quality of life makes the greater difference to a person's well-being.

Diversification can be rational for the risk-avoidant person even if the two charities aren't exactly equal in terms of their mean utility. If Charity B has a slightly lower expected utility, either because B has a lower probability of being successful or will under its success condition be less successful than Charity A under its success condition, then the risk-avoidant person may still prefer the mixture of A and B (and more so to the extent that E and F are anti-correlated).Footnote 26 In other words, the reasoning that leads to diversification doesn't depend on the two charities being exactly equal from an expected utility perspective.

These points about diversification follow from a more general point: risk-avoidant individuals prefer to hedge their bets. And in the case in which the status quo is flat, hedging one's bets requires diversifying.Footnote 27

We assumed that the status quo is flat; but in typical cases we are uncertain of facts about the world other than which charity will be better. Thus, a more realistic case is one in which the status quo is itself a gamble, namely, one in which the status quo does not have a single utility value u(SQ). Can the results be generalized to this case? Not completely, because the status quo can be such that hedging does not always require diversification. There are three ways in which this could happen, and thus three situations in which a risk-avoidant individual might prefer not to diversify.

First, one of the charities might serve as a kind of insurance against a risk currently present in the status quo. For example, the status quo might be flat except for a 3 utility ‘dip’ in event E. (Maybe both charities involve distributing vitamins, and Charity A's vitamins are effective against a particular disease that may or may not be present in the population, whereas Charity B's vitamins increase the quality of life of an already-healthy population if they are effective.) In this case, giving money to Charity A serves as insurance against E, and so the risk-avoidant individual will prefer to give $200 to Charity A.

Second, if one of the charities has higher expected utility than the other and if the status quo has a particular character – if it has ‘large tails’ and is sufficiently ‘dispersed' in technical senses discussed in Tarsney (Reference Tarsney2020) – then giving $200 to the charity with higher expected utility will stochastically dominate giving $100 to each charity, and thus will have higher risk-weighted expected utility for even the risk-avoidant individual.Footnote 28

Third, if in the status quo you have strong reason to believe that everybody else is giving their money to Charity A, or, more precisely, that the total of everyone else's donations to Charity A will be at least $200 greater than the total of their donations to Charity B, then if you are risk-avoidant you will prefer to give $200 to Charity B, as a way of hedging ‘our’ bet, or of buying insurance against the event in which Charity A is less effective ().

Thus, the set of cases in which diversifying among charities will be required for the risk-avoidant, while wider than the set in which diversifying will be required for EU-maximizers, is still somewhat constrained. Still, the risk-avoidant reason for diversifying is important, both for explanatory and for action-guiding purposes.Footnote 29 Even if the status quo is not flat, many people might assume it is. Alternatively, some people might care about doing the most good with their money in particular, that is, making the most difference, which by definition assumes a flat status quo (no difference made). Thus, we have a candidate explanation for why many people in fact diversify even with respect to small sums of money, and a candidate reason for doing so.

More importantly, for those risk-avoidant individuals or groups who are accounting for all the facts and care about bringing about the best gamble, diversification will still be required when doing so is genuinely the best way to hedge. It will be required in cases that don't meet the above three conditions. And even if we are in a case of a dispersed status quo overall, if someone focuses on helping with respect to a particular issue, cause, or place, the status quo will be less dispersed relative to these. Finally, if we consider interventions large enough to swamp the status quo – such as money donated by all of us collectively – then diversification will still be required.

Let us move next to ambiguity. We will again assume a flat status quo, and consider the choice between 2A, 2B, and AB, as above. We will assume that the individual cannot set a precise value for p(E), p(F), and p(EF), but instead assigns sets of probabilities; and we will assume that she uses α-maximin, which (recall) takes an average of the minimum EU and the maximum EU. To determine the effects of ambiguity aversion on the choice among these options, we want to know how the spread of possible values of EU(AB) is related to the spread of possible values of EU(2A) and EU(2B).

Taking AB rather than 2A or 2B amounts to betting some on E and some on F rather than betting all on one of these events. Thus, the value of AB for the person who is sensitive to ambiguity depends on the effects of betting some on each event (rather than more on one event) on the spread of possible EU values of the gamble, which in turn depend on how ‘combining’ the probabilities of E and F affects the spread of probabilities of winning at least one of the bets on E or F.

Let Mp(E) stand for the midpoint of the interval representing p(E), and let Sp(E) stand for how spread out the interval representing p(E) is, so that the ambiguity in p(E) is represented by the interval [Mp(E ) − Sp(E), Mp(E) + Sp(E)].Footnote 30 Taking 2A amounts to getting, above the minimum (above u(SQ) + 2): 2 utility if E comes to pass, that is, to an expectation of 2p(E). Taking 2B amounts to getting, above the minimum (above u(SQ) + 2): 2 utility if F comes to pass, that is, to an expectation of 2p(F). Taking AB amounts to getting, above the minimum (above u(SQ) + 2): if E comes to pass, 1 utility; and also, if F comes to pass, 1 utility – that is, to an expectation of p(E) + p(F).

Thus, the value of AB to someone who cares about ambiguity depends on the various values that p(E) + p(F) could take, that is, on the interval [Mp(E)+p(F) − Sp(E)+p(F), Mp(E)+p(F) + Sp(E)+p(F)]. The size of this interval in turn depends on how much the value of p(E) is constrained by the value of p(F), and vice versa, for each probability in the set. If p(E) and p(F) constrain each other a lot, then this interval is small; for example, if p(E) + p(F) = 1 for all p ∊ ${\cal P}$, then Sp(E)+p(F) = 0, and this interval is simply [1]. If p(E) and p(F) don't constrain each other at all – so that p(E) and p(F) can simultaneously take their lowest (or highest) valuesFootnote 31 – then this interval is as wide as it can be, with Sp(E)+p(F) = Sp(E) + Sp(F). We will assume that Mp(E)+p(F) = Mp(E) + Mp(F); in effect this means that p(E) and p(F) constrain each other in symmetrical ways.Footnote 32

An ambiguity-neutral individual will not strictly prefer diversifying: either she will strictly prefer 2A to all of the other options, or she will strictly prefer 2B to all of the other options, or she will be indifferent among all three options.Footnote 33

What about ambiguity-averse and ambiguity-seeking individuals?

If p(E) and p(F) do not constrain each other's extreme values – if Sp(E)+p(F) = Sp(E) + Sp(F) – then diversification cannot be preferred, for any value of α.Footnote 34 An intuitive way to think about this is that if p(E) and p(F) do not constrain each other, then the ambiguity present in the gambles together is just their combined ambiguity. Thus, whichever gamble the ambiguity-averse individual prefers when taken by itself, that gamble will also be preferred no matter which other gamble she already holds (and similarly for the ambiguity-seeking individual). Diversifying doesn't reduce overall ambiguity.

However, if p(E) and p(F) do constrain each other – if Sp(E)+p(F) < Sp(E) + Sp(F) – then pairing A with B does reduce ambiguity, and so diversifying can be preferred to putting all of one's money towards one gamble.Footnote 35 To take an extreme example, while I might not know whether p(E) is 0.25 or 0.75, and I might not know whether p(F) is 0.25 or 0.75, I might know that if p(E) is 0.25 then p(F) is 0.75, and vice versa, that is, it might be that all the probability functions in my set have p(E) + p(F) = 1. If so, then when I already hold A, taking B provides a kind of insurance against the possibility of A having a bad probability distribution. And as in the risk case, diversifying can be preferred even if one of the charities is by itself slightly preferred to the other.

Thus, the results for ambiguity-aversion and risk-avoidance are similar, but whereas diversification is more apt to be preferred in the case of risk-avoidance when the possible events constrain each other (when their probabilities are not independent), diversification is more apt to be preferred in the case of ambiguity-avoidance when the possible probabilities constrain each other.

In cases of charitable giving, it is not uncommon for the probabilities to constrain each other. Consider the examples of anti-correlation discussed in the case of risk-avoidance, in which the success of Charity A and the success of Charity B depend on opposing empirical or normative claims. If claims are anti-correlated, then their probabilities will often constrain each other (for example, if p(E v F) is fixed, then if p(E & F) = 0, when p(E) is high, p(F) will be low). Alternatively, probabilities may constrain each other if, for example, the highest probability of E and the highest probability of F are based on opposing scientific models. Diversifying in the case of ambiguity is similar to diversifying in the case of risk: whereas in that case, it was rational for the risk-avoider to hedge between different ways the world could turn out, in this case, it is rational for the ambiguity-averter to hedge between different ways the probabilities could be.

When the status quo is flat, hedging between ambiguous gambles makes sense if the individual is ambiguity-averse and if the probabilities constrain each other. However, as in the case of hedging between risky gambles, this fact only holds if the status quo is not itself ambiguous in a particular way. If, for example, donating to Charity A reduces the ambiguity in the world, by hedging against already existing ambiguity, then one might prefer to donate $200 to Charity A rather than to diversify. Or, again, if one knows that many people are donating to Charity A, then one should, if one is ambiguity-averse, donate $200 to Charity B, to reduce the overall ambiguity. So, as in the case of risk-avoidance, ambiguity-aversion won't explain all cases of diversifying; nor will diversifying always be rationally required for ambiguity-averse individuals. However, we have again hit upon a new explanation for diversification, as well as a set of circumstances in which ambiguity-averse individuals ought to diversify. Insofar as diversification looks like buying insurance against the probability distribution turning out a certain way, it's going to be a good thing for the ambiguity-averter. And insofar as it doesn't, it won't be.

4. Existential risk mitigation vs. health amelioration

We have seen how risk-avoidance and ambiguity-aversion can lead to diversification in charitable giving. Let us turn next to the choice between different types of programs to give money to. On the one hand, there are programs which have well-known, high probability, positive effects in exchange for relatively small amounts of money. Consider, for example, the Against Malaria Foundation, which provides insecticidal nets, or Give Directly, which gives money directly to people in need.Footnote 36 Let us define a steady increase program as follows, simplifying ‘high probability’ to ‘probability 1’ for ease of exposition:

Steady Increase Program (SIP): A given small dollar amount raises the utility of the status quo by a fixed, small utility value in every state of the world.

On the other hand, there are programs which have a small probability of doing a massive amount of good, but for which: (1) donating a small amount of money serves primarily to slightly increase the probability of some massive difference in how good the outcome is; and (2) that probability is ambiguous. For example, consider the International Campaign to Abolish Nuclear Weapons (ICAN) or a program advocating for increased precautions against pandemics, both of which have a small probability of causing us to avoid a massively bad outcome; or consider SpaceX's program to colonize Mars, which has a small probability of bringing about a massively good outcome.

There are two ways to slightly increase the probability of some massive utility difference: we could decrease the probability of a massive amount of bad or increase the probability of a massive amount of good. (Here, ‘bad’ just means a possible state that is relatively much worse than most other possible states, and ‘good’ means a possible state that is relatively much better than most other possible states.) Thus, we can distinguish two types of “small probability of large utility” programs. One program we can call existential insurance: in the status quo, some massively bad event might happen, but its probability is small. Existential insurance lowers the chance of that massively bad event, or makes things not go poorly if that event comes to pass; ICAN and the program advocating for pandemic precautions are examples.

Let us define existential insurance as follows:

Existential Insurance (EI): Whereas the status quo currently contains as a possibility a massively bad event (an event with much lower utility than the rest of the status quo) with a small probability, a given dollar amount lowers the probability of getting that utility value.

One special case of existential insurance is the case of lowering the probability of the bad event to 0.

There are also programs that we might call an existential lottery ticket. An existential lottery ticket raises the probability of some currently improbable massively good event. Examples include SpaceX or the transhumanist program dedicated to extending individual lives indefinitely.Footnote 37 We can define an existential lottery ticket as follows:

Existential Lottery (EL): Whereas the status quo currently contains as a possibility a massively good event (an event with much higher utility than the rest of the status quo) with a small probability, a given dollar amount raises the probability of getting that utility value.

One special case of the existential lottery is the case of raising the probability of the “possible” good event from 0.

In this section, we are interested in how different attitudes towards risk and ambiguity should affect our giving to these three types of causes. To isolate the effects of risk and ambiguity from other effects, we will consider a case in which an example of each type has the same expected utility. (This implicitly assumes that there are no infinite values involved.) We will then ask what happens if we take a non-EU attitude towards risk or ambiguity or both.

Let B refer to the bad event with massively low utility and low probability (e.g., nuclear war), G refer to the good event with massively high utility and low probability (e.g., colonizing other galaxies), and N refer to the neutral event in which neither happens. To give some definite values, we will say that in the status quo, the bad event has utility −100, the neutral event has utility 0, and the good event has utility 100. Furthermore, the steady increase program adds just over 1 utility in the neutral event (we are assuming that it doesn't add anything in the good or bad event). The existential insurance lowers the probability of the bad event by e and the existential lottery ticket raises the probability of the good event by t. So we have:

SQ = {−100, p(B); 0, p(N); 100, p(G)}

SIP = {−100, p(B); 1/p(N), p(N); 100, p(G)}

EI = {−100, p(B) − e; 0, p(N) + e; 100, p(G)}

EL = {−100, p(B); 0, p(N) − t; 100, p(G) + t}

We will begin by assuming that we can give e and t definite values; we will assume e = t = 0.01. When one is comparing actual Steady Increase Programs to Existential Insurance to Existential Lottery programs, the utility numbers may be much greater in magnitude and probability numbers much smaller in magnitude than the ones in our example; however, very large and very small numbers can be difficult to conceptualize, so I'm using this toy example to make the effects of risk and ambiguity easier to understand.

Under these assumptions, the expected utility maximizer is indifferent between SIP, EI, and EL. (It is easy to see that EU(SIP) = EU(EI) = EU(EL) = EU(SQ) + 1.) All three gambles involve adding 1 expected utility to the status quo, spread out in different ways: SIP adds approximately 1 utility nearly everywhere; EI adds 100 utility to 0.01 of the states in the bottom part of the gamble; and EL adds 100 utility to 0.01 of the states in the top part of the gamble. (If p(B) = e, then EI eliminates the worst possibilities, and if p(G) = 0, then EL introduces new best possibilities.)

Now let us consider risk-attitudes. The strictly risk-avoidant person puts more weight on relatively worse scenarios and less weight on relatively better scenarios. Given 1 utility to distribute, she prefers putting as much as possible in the worst possible states. Thus, the strictly risk-avoidant person will strictly prefer the existential insurance to the steady increase program, and strictly prefer the steady increase program to the existential lottery.Footnote 38 The strictly risk-inclined person will have the opposite preferences, since she cares more about what happens in better states than in worse states (Table 1).

Table 1. Attitudes towards risk

To consider ambiguity-attitudes, recall that both existential programs can involve uncertainty about the effects of one's contribution. To capture this, instead of fixing e and t at 0.01, they will be represented by an interval centered on 0.01 with spread 0 < s ≤ 0.01: e ∈ [0.01 − s, 0.01 + s], t ∈ [0.01 − s, 0.01 + s]. This captures the idea that although these programs might change the probability of the massively bad or massively good thing by 0.01, they might instead (optimistically) change it by 0.01 + s, or they might instead (pessimistically) change it by 0.01 − s.

For the ambiguity-avoidant person, the ambiguity involved in both the existential insurance and the existential lottery makes these programs worse than the steady increase program. However, since EU is insensitive to where utility is added, a theory of ambiguity-preferences based on averaging the minimum and maximum EU will treat ambiguity in worse states equivalently to ambiguity in better states.Footnote 39 Therefore, if the ambiguity in both gambles is the same (the spread is equally wide, and concerns the same amount of utility), the ambiguity-averse individual will be indifferent between the existential insurance and the existential lottery. Thus, the ambiguity-averse person strictly prefers the steady increase program to both the existential insurance and the existential lottery, and is indifferent between the latter two; and by analogous reasoning, the ambiguity-seeking person strictly disprefers the steady increase program to the existential insurance and the existential lottery, which she is indifferent between (Table 2). (It might be, contra the toy example, that there is more ambiguity involved in the existential lottery than in the existential insurance, or vice versa, in which case the ambiguity-averse person will prefer the gamble with less ambiguity; the point is that the effects of ambiguity do not depend on whether a gamble is insurance or a lottery.)

Table 2. Attitudes towards ambiguity.

We can also consider what happens when we combine attitudes to risk with attitudes to ambiguity.Footnote 40 The possibilities are listed in Table 3.Footnote 41 Notice that for many combinations of attitudes, there are multiple possibilities. For example, if you're risk-avoidant and ambiguity-averse, then the existential lottery is made worse than the other choices by risk-avoidance, and made not better by ambiguity-aversion; but since risk-avoidance makes the existential insurance preferable to the steady increase program and ambiguity-aversion does the opposite, which is preferable overall will depend on which attitude is more severe.

Table 3. Preferences with different attitudes

It would be unrealistic, of course, to assume that Steady Increase Programs, Existential Insurance Programs, and Existential Lotteries in general have the same average utilities, and the orderings in Figure 5 are only guaranteed to hold if they do. When the programs do not have the same average utilities, “≻” and “≺” and “≈” instead represent the direction of change in value based on risk and ambiguity attitudes; for example, for an ambiguity-averse and risk-avoidant individual, Steady Increase Programs will go up in value relative to Existential Lottery Programs, in contrast to how an EU-maximizer evaluates these.

One surprising fact to note is that the Steady Increase Program can only look better than both Existential Programs if one is ambiguity-averse; and even then, it may not, if one is sufficiently risk-avoidant or risk-inclined. In other words, caring about risk (in either direction) won't provide reason to pick the option that makes a difference for certain, since the person who avoids risk will rather have insurance, and the person who seeks risk will rather have a lottery. This surprising fact, of course, rests on the assumption that the programs are equally good on average. Furthermore, while I have provided canonical examples of Existential Programs, the claim that the relevant good or bad events in these examples are genuinely much better or worse than the status quo (which includes people in poverty or with unmet health needs or prematurely dying) rests on a particular aggregative theory of the good, among other assumptions. So application to real-world choices – deciding to give money to increase pandemic precautions or to reduce cases of malaria, for example – rests on carefully considering whether these features hold.

5. What attitudes should we adopt in our charitable giving?

We've explored how we should apportion our giving when we adopt different attitudes towards risk and ambiguity. But which attitudes should we adopt? There are several possibilities for answering this question, and this final section will explain where the decision points lie.Footnote 42

The first decision point is whether charitable giving should be thought of as an individual or collective enterprise: am I trying to do the most (risk-weighted and ambiguity-resolved expected) good possible with my money, or are we trying to do the most (risk-weighted and ambiguity-resolved expected) good possible with our money? If it should be thought of as an individual enterprise, then each individual giver should hold fixed what others are doing – treat others’ behavior as part of the status quo – and concentrate on the difference her donations make to the good. According to this view, we should then ask: which risk-attitudes and ambiguity-attitudes should an individual adopt when making decisions whose purpose is to bring about good for others? There are three possible answers. First, that she should adopt her own attitudes. Second, that she should adopt the others’ attitudes, however ‘their attitudes’ is to be defined for the collective she is trying to help.Footnote 43 And third, that she should adopt some particular, objective risk-attitude and ambiguity-attitude. Uncertainty in the status quo will be, for the typical giver, more pronounced under the view that charitable giving is an individual enterprise, since the amount of money any particular giver is giving is dwarfed by others’ contributions.

If charitable giving should be thought of as a collective enterprise, then we should treat the status quo as including only contingencies that exist apart from our collective intervention, and ask about the difference that our collective charitable giving will make to the (risk-weighted and ambiguity-resolved expected) good. There are again three different ways to think about which attitudes we should adopt when making decisions whose purpose is to bring about good for others, echoing those above. First, that we should adopt our own attitudes, where ‘our attitudes’ is determined by aggregating in some way, just as we might aggregate if we disagree about the probabilities. Second, that we should adopt the attitudes of the others we are trying to help. Finally, that we should adopt some particular, objective attitudes. This last option will yield a different recommendation than under the ‘individual’ view, since the choices will be about total giving rather than individual giving holding fixed what others give, and thus both the characterization of the status quo and the amount of money under consideration will be very different.

Many ways of resolving these choice points will lead to risk-avoidance, ambiguity-aversion, or both. Since many people in fact have these attitudes, if we use the attitudes of either the giver or the ones being helped, we are apt to land on risk-avoidant and ambiguity-averse choices. (Though I note that the common ‘S-shaped’ risk profile will also lead to risk-inclination in small probabilities, such as in the case of the existential lottery.) Likewise, if we think there is an objective requirement on how to set these attitudes when making choices that affect others, then there may be a good case to be made that that requirement is risk-avoidance or ambiguity-aversion.Footnote 44

More will have to be done to say which attitudes we should adopt in our charitable giving, and exactly which charities we should give to. But my hope is that this article has shed light on exactly why and how these attitudes matter when figuring out how to do the most good.

Acknowledgments

This article benefited greatly from comments by Jacob Barrett, Hilary Greaves, Andreas Mogensen, Christian Tarsney, Teru Thomas, David Thorstad, Hayden Wilkinsen, and Timothy Williamson; from comments by the editor of this journal and two anonymous referees; and from discussions at the Global Priorities Institute in Oxford in 2019, where it was first presented as the Derek Parfit Memorial Lecture in 2019.

Competing interests

The author(s) declare none.

Footnotes

1 This characterization of what formal theories of decision-making tell us is Savage's (Reference Savage1954). The extensions I discuss are all based on Savage's theory. Other theories have slightly different objects of evaluation; for example, Jeffrey's (Reference Jeffrey1965) theory tells us how to evaluate “news items.” It is an interesting question whether and how we can extend Jeffrey's theory to accommodate risk and ambiguity in the senses discussed in this article, but it is beyond the scope of this article. For an account of risk and ambiguity using Jeffrey's theory, see Bradley (Reference Bradley2015), Stefánsson and Bradley (Reference Stefánsson and Bradley2015), Bradley and Stefánsson (Reference Bradley and Orri Stefánsson2019). Their characterization of risk and ambiguity is notably different from that of the theories discussed here, in that they see risk and ambiguity as part of the consequences themselves – as objects of final rather than instrumental value.

2 Similar examples appear in Buchak (Reference Buchak2023).

3 Preferring acts that maximize expected utility (EU) is equivalent to having preferences that accord with certain axioms, such as those of Savage (Reference Savage1954). I take no view on whether the utility function or satisfaction of the axioms is more basic for any of the theories discussed here. However, the reader should note that when considering the non-EU theories here, all of which are strictly more general than EU, we cannot think of the utility function as issuing from the axioms of EU theory; instead, we must think of the utility function as issuing from the more general axioms of the theory in question, which agrees with EU if and only if the agent maximizes EU according to her preferences.

4 Figures 1 to 4 are similar to those in Buchak (Reference Buchak2023).

5 There is, of course, a sense in which expected utility maximization allows for different attitudes towards risk, if we think of one's attitude towards risk as a property of the utility function (i.e., its concavity). The complaint is that this device doesn't adequately capture the full range of attitudes towards risk; see Hansson (Reference Hansson, Gärdenfors and Sahlin1988), Rabin (Reference Rabin2000), and Buchak (Reference Buchak2013), among others.

7 Quiggin (Reference Quiggin1982). An example of a rank-dependent model for ambiguity is Choquet expected utility (Schmeidler Reference Schmeidler1982, Gilboa Reference Gilboa1987). See also the rank-dependent theories of Yaari (Reference Yaari1987), Kahneman and Tversky (Reference Kahneman and Tversky1979), and Tversky and Kahneman (Reference Tversky and Kahneman1992); and the general analysis of rank-dependent theories in Wakker (Reference Wakker2010).

8 r(0) = 0; r(1) = 1; r is non-decreasing. We might additionally require that r is increasing or that r is continuous.

9 For example, {3, 0.1; 4, 0.8; 5, 0.1} is a mean-preserving spread of {4, 1}, because 0.1 of the probability mass is moved ‘outward’ by 1 unit on both sides. For another example, {2, 0.2; 4, 0.4; 5, 0.4} is also a mean-preserving spread of {4, 1}, because 0.2 of the probability mass is moved outward by 2 units to the left, and a counterbalancing 0.4 of the probability mass is moved outward by 1 unit to the right. See definition and results in Rothschild and Stiglitz (Reference Rothschild and Stiglitz1972).

10 Rothschild and Stiglitz (Reference Rothschild and Stiglitz1972).

11 See Buchak (Reference Buchak2013: 63 n. 16), which cites Chew et al. (Reference Chew, Karni and Safra1987: 375, Corollary 2). An individual displays strict risk-aversion iff (1) the utility function is weakly concave and the risk function is weakly convex and (2) the utility function is strictly concave or the risk function is strictly convex. Note that these are not just notational variants: someone who maximizes expected utility with a diminishing marginal utility function will have different preferences than someone who maximizes risk-weighted expected utility with a convex risk function; for example, the well-known Allais preferences are an example of the preferences of the latter type but not the former.

12 See Buchak (Reference Buchak2013: 63). The shape of u and r are independent, so that you could, for example, have a concave u but be risk-inclined, and so forth.

14 See especially examples in Ellsberg (Reference Ellsberg1961, Reference Ellsberg2001).

15 Some support for each of these ideas comes from the way that scientists report their data. For example, if there are three climate models, each of which is plausible, that give three different probability values for a drastic sea level rise, then climate scientists in general won't report a number that is simply the average of the three probabilities; instead, they will report an interval spanning the probabilities or a set of the three possible probabilities.

16 A representation of the first type can determine a representation of the second, assuming that the set of probability functions is convex (Wakker (Reference Wakker2010: 324)), where convexity means that the set includes any mixture of its probability functions, for example p3 = 0.5p1 + 0.5p2.

17 See Buchak (Reference Buchak and Zalta2022) for a survey.

18 α-maximin and Choquet expected utility have many similar upshots; indeed, they are equivalent when evaluating two-outcome prospects (Wakker (Reference Wakker2010: 327, Observation 11.5.1)).

20 Both ways are compatible with views according to which moral value is not equivalent to individual welfare, such as prioritarianism or egalitarianism. There is one complication, which is that the moral value of a gamble might depend not just on the risks involved, but on how those risks are distributed among individuals’ welfare; this would prevent us from assigning a moral value to outcomes which is stable when those outcomes compose a gamble. We will ignore this complication, since the point is to see how risk- and ambiguity-attitudes affect what we should do when all else is equal. (Those who think we can use EU-maximization to capture attitudes towards risk can hold that we have access to the moral value function, but that in moral decisions, the utility function is a non-linear function of the moral value function.)

21 See also the discussion of diversification in charitable giving in Bovens (Reference Bovens and White2019) and Snowden (Reference Snowden, Greaves and Plumer2019), the latter of whom argues against my view.

22 An example of the first situation might be if we are dealing with a charity with a very small, specific purpose. An example of the second might be if we are trying to find a cure for a specific disease and two charities are funding work on different sorts of cures and are already fairly well-funded. An example of the third might be if one charity helps diagnose a disease and the other helps cure it (Snowden Reference Snowden, Greaves and Plumer2019: 3.2). See additional examples in Snowden (Reference Snowden, Greaves and Plumer2019). I note additionally that in order for diminishing marginal utility to recommend diversification between two charities, the two charities would have to be close enough in diminishment at their margins for the money given to one to be the tipping point to make marginal donations to the other have a higher EU.

23 Snowden (Reference Snowden, Greaves and Plumer2019) provides an example of this form.

24 See Buchak (Reference Buchak2013: 65 n. 26) and the definition of general risk aversion in Buchak (Reference Buchak2013: 22–23).

25 Note that a more general point is true: anyone who is averse to mean-preserving spreads in utility or in the good will rather hedge her bets, not just individuals whose preferences conform to REU.

26 We can see this by noticing that in the setup where A and B have the same mean, AB is strictly better than both 2A and 2B; thus you'd be willing to lose a bit of money ɛ to have AB rather than 2A. And so if B is made worse in a way that makes the resulting AB worse by δ < ɛ, the resulting AB is still better than 2A.

27 As long as the bets’ distributions over utility aren't wildly different. For example, if one charity has a far greater utility spread, then a risk-avoidant individual might, like the EU-maximizer, prefer to give her money to only one charity – the one with the much smaller spread – rather than to divide her money between the charities.

28 Tarsney's general result implies that for any two prospects, one of which has higher EU, there is some status quo such that the prospect with higher EU will stochastically dominate the prospect with lower EU.

29 The idea that people in fact diversify in their charitable giving, even when the marginal effects of their donations are linear, is a point of discussion in effective altruist circles; however, I cannot find sources that back up this empirical claim.

30 Let $M_{p( E ) } = {{max_pp( E ) + min_pp( E ) } \over 2}$ and$\;S_{p( E ) } = {{max_pp( E ) -min_pp( E ) } \over 2}$, so that min pp(E) = M p(E) − S p(E) and max pp(E) = M p(E) + S p(E).

31 That is, minp[p(E) + p(F)] = minp[p(E)] + minp[p(F)], and maxp[p(E) + p(F)] = maxp[p(E)] + maxp[p(F)].

32 For example, it can't be the case that p(E) and p(F) can simultaneously take their lowest value but cannot simultaneously take their highest value.

33 Proof:

α-maximin(2A) = u(SQ) + 2 + 2[(1 − α)minp[p(E)] + (α)maxp[p(E)]]

α-maximin(2B) = u(SQ) + 2 + 2[(1 − α)minp[p(F)] + (α)maxp[p(F)]]

α-maximin(AB) = u(SQ) + 2 + [(1 − α)minp[p(E) + p(F)] + (α)maxp[p(E) + p(F)]]

In terms of midpoint and spread, we have:

α-maximin(2A) = u(SQ) + 2 + 2[Mp(E) + (2α − 1)Sp(E)]

α-maximin(2B) = u(SQ) + 2 + 2[Mp(F) + (2α − 1)Sp(F)]

α-maximin(AB) = u(SQ) + 2 + Mp(E)+p(F) + (2α − 1)Sp(E)+p(F)

We assume that Mp(E)+p(F) = Mp(E) + Mp(F).

If α = 0.5, then

α-maximin(2A) = u(SQ) + 2 + 2Mp(E)

α-maximin(2B) = u(SQ) + 2 + 2Mp(F)

α-maximin(AB) = u(SQ) + 2 + Mp(E)+p(F) = u(SQ) + 2 + Mp(E) + Mp(F)

So either 2AAB2B (if Mp(E) > Mp(F)) or 2BAB2A (if Mp(E) < Mp(F)) or 2A2BAB (if Mp(E) = Mp(F)). Diversifying can't be strictly preferred to all other options.

34 We know that:

α-maximin(A) = u(SQ) + 1 + [Mp(E) + (2α − 1)Sp(E)]

α-maximin(B) = u(SQ) + 1 + [Mp(F) + (2α − 1)Sp(F)]

Assume that AB. Then [Mp(E) + (2α − 1)Sp(E)] ≥ [Mp(F) + (2α − 1)Sp(F)]. Therefore, 2[Mp(E) + (2α − 1)Sp(E)] ≥ [Mp(E) + (2α − 1)Sp(E)] + [Mp(F ) + (2α − 1)Sp(F)] ≥ 2[Mp(F) + (2α − 1)Sp(F)]. If we assume that Mp(E)+p(F) = Mp(E) + Mp(F) and Sp(E)+p(F) = Sp(E) + Sp(F), then these inequalities are equivalent to α-maximin(2A) ≥ α-maximin(AB) ≥ α-maximin(2B); therefore 2AAB2B. These reverse if we assume that BA. Again, diversifying can't be strictly preferred to all other options.

35 We assume that Sp(E)+p(F) < Sp(E) + Sp(F) .

α-maximin(AB) − α-maximin(2A) = [Mp(E) + Mp(F)2Mp(E)] + (2α − 1)[Sp(E) + p(F) − 2Sp(E)] = [Mp(F) − Mp(E)] + (2α − 1)[Sp(E) + p(F) − (Sp(F) + Sp(E)) + (Sp(F) − Sp(E))]

(i) If Mp(E) = Mp(F) and Sp(E) = Sp(F), then α-maximin(AB) – α-maximin(2A) = (2α − 1)[Sp(E) + p(F) – (Sp(F) + Sp(E))]. Since Sp(E)+p(F) < Sp(E) + Sp(F) , this value is positive for α < 0.5 and negative for α > 0.5. Mutatis mutandis for α-maximin(AB) – α-maximin(2B). Thus, diversifying is strictly preferred for α < 0.5 and strictly dispreferred for α > 0.5.

(ii) In the case in which Mp(E) > Mp(F) and Sp(F) > Sp(E), for α < 0.5, α-maximin(AB) − α-maximin(2A) will still be positive for small enough values of Mp(E) − Mp(F) and Sp(F) − Sp(E).

36 According to GiveWell (2022), AMF protects an additional person from malaria for each $2.50–$3.06 spent, and prevents an additional death for each $2,676–$7,458. There are reasons to question whether the positive impact of these programs approaches anything close to certainty – some of the data relies on best guesses or averages, and these programs may have indirect effects as well – but we will set these worries aside.

37 For a different kind of example, consider Pascal's Wager. A related non-theological example is Dyson's Wager (Wilkinson Reference Wilkinson2022).

38 EL is a mean-preserving spread of SIP, and SIP is a mean-preserving spread of EI. Thus, the risk-avoidant individual has EISIPEL, with inequalities strict for strict risk-avoidance (i.e., the risk function is strictly convex); and the risk-inclined individual has the opposite preferences. Note that, as above, a more general point is true: anyone who is averse to mean-preserving spreads in utility or in the good will rather hedge her bets, not just individuals whose preferences conform to REU. One might wonder about the comparison between EI and the ‘gamble’ that adds exactly 1 utility in every state, including the good and bad states, that is, {−99, p(B); 1, p(N); 101, p(G)}. This gamble will be very close in value to SIP; however, with very extreme risk-avoidance, it could be preferred to EI. This is because EI reduces the possibility of a massively bad event, but (unless p(B) = e) it does not eliminate this possibility, so its worse case is still worse than that of SIP. For example, in the limit case of maximin, SIP will be preferred for this reason (“If we might die in a nuclear war, at least we'll all be healthier!”).

39 α-maximin(SQ) = EU(SQ)

α-maximin(SIP) = EU(SQ) + 1

EU(EI) = EU(SQ) + 100e; therefore α-maximin(EI) = EU(SQ) + 1 + (2α – 1)(100s)

EU(EL) = EU(SQ) + 100t; therefore α-maximin(EL) = EU(SQ) + 1 + (2α – 1)(100s)

For α < 0.5, SIPEIEL. For α > 0.5, EI’ELSIP. For α = 0.5, SIPEIEL

40 Instead of hedging between the minimum and maximum EU value, we hedge between the minimum and maximum REU value:

α-REU(g) = (1 − α)[minp∊$_{\cal P }$ REUp(g)] + α[maxp∊$_{\cal P }$ REUp(g)]

Define SIPx = {−100, p(B); 100x/p(N), p(N); 100, p(G)}, with special case SIP0 = SQ. This is the EU-equivalent for EI with e = x and EL with t = x.

We know: α-REU(SIP) = REU(SIP) = 0.5(REU(SIP0.01-s)) + 0.5(REU(SIP0.01+s)).

For non-zero x, EL with t = x is a mean-preserving spread of SIPx, and SIPx is a mean-preserving spread of EI with e = x. Therefore, for a risk-avoidant agent, minp∊$_{\cal P}$ REUp(EI) ≥ REU(SIP0.01-s) ≥ minp∊$_{\cal P}$ REUp(EL) and maxp∊$_{\cal P}$ REUp(EI) ≥ REU(SIP0.01+s) ≥ maxp∊$_{\cal P}$ REUp(EL); with the first set of inequalities equal in the degenerate case in which s = 0.01; and both sets of inequalities strict if s < 0.01 and risk-avoidance is strict. (Inequalities reversed for risk-inclination, and inequalities equal for global neutrality.)

So we have:

α-REU(EI) = (1 − α)[minp∊$_{\cal P}$ REUp(EI)] + α[maxp∊$_{\cal P}$ REUp(EI)]

α-REU(SIP) = (1 − 0.5)[REU(SIP 0.01−s)] + (0.5)[REU(SIP 0.01+s)]

α-REU(EL) = (1 − α)[minp∊$_{\cal P}$ REUp(EL)] + α[maxp∊$_{\cal P}$ REUp(EL)]

Under α = 0.5 and global neutrality, all three values are equal.

In each calculation, we mix a higher value with a lower value. Risk-attitudes affect the size of the two REU expressions in each calculation, and ambiguity attitudes affect the weight each of the two expressions gets for EI and EL. This shows how we get the possible inequalities listed in Table 3.

41 We here omit the degenerate case in which s = 0.01 and α = 1; this would allow for SIPEIEL under ambiguity-averse/risk-avoidant attitudes and under ambiguity-averse/risk-inclined attitudes.

42 See also the overview of views in Thoma (Reference Thoma2023).

43 Snowden (Reference Snowden, Greaves and Plumer2019) points out that different acts might help different people, so if we adopt the attitudes of the people we are trying to help, we also have to decide whether to aggregate across people first or across states (taking into account risk and ambiguity) first.

44 For the claim that risk-avoidance is the required attitude, see Buchak (Reference Buchak2017); for the claim that risk-neutrality is the required attitude, see Zhao (Reference Zhao2023); for the claim that the required attitude is different when one is making a risky decision for someone than when one is executing a decision on someone's behalf, see Bovens (Reference Bovens and White2019).

On ambiguity-aversion as the required attitude, see Rowe and Voorhoeve (Reference Rowe and Voorhoeve2019), as well as discussions of the precautionary principle by Hansson (Reference Hansson1997), Gardiner (Reference Gardiner2006, Reference Gardiner, Gardiner, Caney, Jamieson and Shue2010), McKinnon (Reference McKinnon2009), and Moellendorf (Reference Moellendorf2014).

References

Bovens, Luc. 2019. The Ethics of Making Risky Decisions for Others. In Oxford Handbook of Ethics and Economics, ed. White, Mark D. (Oxford: Oxford University Press), pp. 446–73.10.1093/oxfordhb/9780198793991.013.21CrossRefGoogle Scholar
Bradley, Richard. 2015. Ellsberg's Paradox and the Value of Chances. Economics and Philosophy 32(2): 231–48.CrossRefGoogle Scholar
Bradley, Richard, and Orri Stefánsson, H.. 2019. What is risk aversion? British Journal for the Philosophy of Science 70(1): 77102.Google Scholar
Buchak, Lara. 2013. Risk and Rationality (Oxford: Oxford University Press).CrossRefGoogle Scholar
Buchak, Lara. 2015a. Replies to Commentators. Philosophical Studies 174(9): 23972414.CrossRefGoogle Scholar
Buchak, Lara. 2015b. Revisiting Risk and Rationality: A reply to Pettigrew and Briggs. Canadian Journal of Philosophy 45(5–6): 841–62.CrossRefGoogle Scholar
Buchak, Lara. 2017. Taking Risks Behind the Veil of Ignorance. Ethics 127(3): 610–44.CrossRefGoogle Scholar
Buchak, Lara. 2022. Normative Theories of Rational Choice: Rivals to Expected Utility Theory. The Stanford Encyclopedia of Philosophy, ed. Zalta, Edward N..Google Scholar
Buchak, Lara. 2023. Philosophical Foundations for Worst-Case Arguments. Philosophy, Politics, and Economics.CrossRefGoogle Scholar
Chandler, Jake. 2014. Subjective Probabilities Need Not Be Sharp. Erkenntnis 79(6): 1273–86.CrossRefGoogle Scholar
Chew, Soo Hong, Karni, Edi, and Safra, Zvi. 1987. Risk-Aversion in the Theory of Expected Utility with Rank Dependent Probabilities. Journal of Economic Theory 42: 370–81.Google Scholar
Elga, Adam. 2010. Subjective Probabilities Should be Sharp. Philosophers Imprint 10(5): 111.Google Scholar
Ellsberg, Daniel. 1961. Risk, Ambiguity, and the Savage Axioms. Quarterly Journal of Economics 75: 643–69.10.2307/1884324CrossRefGoogle Scholar
Ellsberg, Daniel. 2001. Risk, Ambiguity, and Decision (New York, NY: Garland).Google Scholar
Gardiner, Stephen M. 2006. A Core Precautionary Principle. Journal of Political Philosophy 14(1): 3360.CrossRefGoogle Scholar
Gardiner, Stephen M. 2010. Ethics and Global Climate Change. In Climate Ethics: Essential Readings, ed. Gardiner, Stephen M., Caney, Simon, Jamieson, Dale, Shue, Henry (New York, NY: Oxford University Press), pp. 135.CrossRefGoogle Scholar
Ghirardato, Paolo, Maccheroni, Fabio, Marinacci, Massimo, and Siniscalchi, Marciano. 2003. A Subjective Spin on Roulette Wheels. Econometrica 71: 18971908.CrossRefGoogle Scholar
Gilboa, Itzhak. 1987. Expected Utility with Purely Subjective Non-Additive Probabilities. Journal of Mathematical Economics 16: 6588.CrossRefGoogle Scholar
Hájek, Alan, and Smithson, Michael. 2012. Rationality and Indeterminate Probabilities. Synthese 187: 3348.CrossRefGoogle Scholar
Hansson, Bengt. 1988. Risk-Aversion as a Problem of Conjoint Measurement. In Decision, Probability, and Utility, ed. Gärdenfors, Peter and Sahlin, Nils-Eric (Cambridge, UK: Cambridge University Press), pp. 136–58.CrossRefGoogle Scholar
Hansson, Sven Ove. 1997. The Limits of Precaution. Foundations of Science 2: 293306.CrossRefGoogle Scholar
Hurwicz, Leonid. 1951. The Generalized Bayes-Minimax Principle. Cowles Commission Discussion Paper: Statistics No. 355.Google Scholar
Jeffrey, Richard. 1965. The Logic of Decision (New York, NY: University of Chicago Press).Google Scholar
Joyce, James M. 2011. A Defense of Imprecise Credences in Inference and Decision Making. Philosophical Perspectives 24: 281323.CrossRefGoogle Scholar
Kahneman, Daniel, and Tversky, Amos. 1979. Prospect Theory: An Analysis of Decision under Risk. Econometrica 47: 263–91.CrossRefGoogle Scholar
Konek, Jason. ms. Epistemic Conservativity and Imprecise Credence.Google Scholar
Luce, R. Duncan, and Raiffa, Howard. 1957. Games and Decisions (New York, NY: John Wiley).Google Scholar
Mayo-Wilson, Conor, and Wheeler, Gregory. 2016. Scoring Imprecise Credences: A Mildly Immodest Proposal. Philosophy and Phenomenological Research 93(1): 5587.CrossRefGoogle Scholar
McKinnon, Catriona. 2009. Runaway Climate Change: A Justice-Based Case for Precautions. Journal of Social Philosophy 40(2): 187203.CrossRefGoogle Scholar
Moellendorf, Darrel. 2014. The Moral Challenge of Dangerous Climate Change: Values, Poverty, and Policy (Cambridge, UK: Cambridge University Press).CrossRefGoogle Scholar
Quiggin, John. 1982. A Theory of Anticipated Utility. Journal of Economic Behavior and Organization 3: 323–43.CrossRefGoogle Scholar
Rabin, Matthew. 2000. Risk-Aversion and Expected-Utility Theory: A Calibration Theorem. Econometrica 68(5): 1281–92.CrossRefGoogle Scholar
Rothschild, Michael, and Stiglitz, Joseph E.. 1972. Addendum to “Increasing Risk: I. A Definition”. Journal of Economic Theory 5: 306.CrossRefGoogle Scholar
Rowe, Thomas, and Voorhoeve, Alex. 2019. Egalitarianism under Severe Uncertainty. Philosophy and Public Affairs 46(3): 239–68.CrossRefGoogle Scholar
Savage, Leonard. 1954. The Foundations of Statistics (New York, NY: John Wiley & Sons).Google Scholar
Schmeidler, David. 1982. Subjective Probability and Expected Utility without Additivity. Econometrica 57(3): 571–87.CrossRefGoogle Scholar
Schoenfield, Miriam. 2017. The Accuracy and Rationality of Imprecise Credences. Noûs 5(4): 667–85.CrossRefGoogle Scholar
Schoenfield, Miriam. 2020. Can Imprecise Probabilities Be Practically Motivated? A Challenge to the Desirability of Ambiguity Aversion. Philosophers’ Imprint 20(30): 121.Google Scholar
Seidenfeld, Teddy. 1988. Decision Theory without “Independence” or without “Ordering”. Economics and Philosophy 4: 267–90.CrossRefGoogle Scholar
Seidenfeld, Teddy, Schervish, Mark J., and Kadane, Joseph B.. 1990. Decisions Without Ordering. In Acting and Reflecting: The Interdisciplinary Turn in Philosophy, ed. Seig, E. (Dordrecht: Kluwer Academic Publishers), pp. 143–70.CrossRefGoogle Scholar
Seidenfeld, Teddy, Schervish, Mark J., and Kadane, Joseph B.. 2012. Forecasting with imprecise probabilities. International Journal of Approximate Reasoning 53: 1248–61.CrossRefGoogle Scholar
Shackle, G. L. S. 1952. Expectations in Economics (Cambridge, UK: Cambridge University Press).Google Scholar
Snowden, James. 2019. Should We Give to More Than One Charity? In Effective Altruism: Philosophical Issues, ed. Greaves, Hilary and Plumer, Theron (Oxford: Oxford University Press), pp. 6979.CrossRefGoogle Scholar
Stefánsson, H. Orri, and Bradley, Richard. 2015. How Valuable are Chances? Philosophy of Science 82(4): 602–25.CrossRefGoogle Scholar
Tarsney, Christian. 2020. Exceeding expectations: stochastic dominance as a general decision theory. GPI Working Paper No.32020.Google Scholar
Thoma, Johanna. 2023. Taking Risks on Behalf of Another. Philosophy Compass 18(3): 113.CrossRefGoogle Scholar
Tversky, Amos, and Kahneman, Daniel. 1992. Advances in Prospect Theory: Cumulative Representation of Uncertainty. Journal of Risk and Uncertainty 5: 297323.CrossRefGoogle Scholar
Vallinder, Aron. 2018. Imprecise Bayesianism and Global Belief Inertia. British Journal for the Philosophy of Science 69(4): 1205–30.CrossRefGoogle Scholar
Wakker, Peter P. 2010. Prospect Theory for Risk and Ambiguity (Cambridge, UK: Cambridge University Press).CrossRefGoogle Scholar
White, Roger. 2009. Evidential Symmetry and Mushy Credence. In Oxford Studies in Epistemology, ed. Szabo Gendler, T. and Hawthorne, J. (Oxford: Oxford University Press), pp. 161–86.Google Scholar
Wilkinson, Hayden. 2022. In Defense of Fanaticism. Ethics 122(2): 445–77.CrossRefGoogle Scholar
Yaari, Menahem E. 1987. The Dual Theory of Choice under Risk. Econometrica 55(1): 95115.CrossRefGoogle Scholar
Zhao, Michael. 2023. Ignore Risk; Maximize Expected Value. Noûs 57(1): 144–61.CrossRefGoogle Scholar
Figure 0

Figure 1. Expected utility.

Figure 1

Figure 2. Expected utility, in terms of utility levels.

Figure 2

Figure 3. Risk-weighted expected utility, with risk-avoidance.

Figure 3

Figure 4. α-maximin.

Figure 4

Table 1. Attitudes towards risk

Figure 5

Table 2. Attitudes towards ambiguity.

Figure 6

Table 3. Preferences with different attitudes