Utilitarianism is the view that a social planner should choose options which maximise the social utility of the resulting social outcome. The central object in utilitarianism is the social utility function u:S→R which assigns a real value u(s)∈R to each social outcome s∈S. This function typically involves variables such as the well-being, preferences, and mental states of individuals, distributional factors like inequality, and other relevant factors such as justice, social cohesion, and freedoms. Utilitarianism is a broad class of social choice principles, one corresponding to each function u:S→R.
In my previous article, I introduced aggregative principles, which state that a social planner should make decisions as if they will face the aggregated personal outcomes of every individual in the population. The central object in aggregativism is the function ζ:S→P, represented with the the greek letter zeta, which assigns a personal outcome ζ(s)∈P to each social outcome s∈S. This function typically aggregates the collection of personal outcomes facing the entire population into a single personal outcome. Aggregativism is a broad class of social choice principles, one corresponding to each function ζ:S→P.
We examined three well-known aggregative principles:
Live Every Life Once (LELO), where ζ(s) is the concatenation of every individual's life.
Harsanyi's Lottery (HL), where ζ(s) is a uniform lottery over every individual's life.
Rawls' Original Position (ROI), where ζ(s) is Knightian uncertainty over every individual's life.
I'm interested in aggregative principles because they avoid many theoretical pitfalls of utilitarian principles. Unlike utilitarianism, aggregativism doesn't require specifying a social welfare function, which is notoriously intractable. Moreover, it seems less prone to counterintuitive conclusions such as the repugnant conclusion or the violation of moral side constraints.[1] In this article, I will show that, under natural conditions of human rationality, aggregative principles approximate utilitarian principles. Therefore, even though aggregativism avoids these theoretical pitfalls, we should nonetheless expect aggregativism to generate roughly-utilitarian recommendations in practical social contexts, and thereby retain the most appealing insights from utilitarianism.
The rest of the article is organized as follows. Section 2 formalises social choice principles as functions of type (X→S)→P(X). Section 3 demonstrates the structural similarity between two strategies to specifying such principles, namely the aggregative and utilitarian strategies. Section 4 proves that under natural conditions about human rationality, the aggregative and utilitarian principles are mathematically equivalent. This theorem is the key contribution of the article. Sections 5, 6, and 7 applies the theorem to LELO, HL, and ROI respectively.
2. Social choice principles
Suppose you are a social planner choosing from a set of options X={x1,…,xn}. The set X might be the set of available tax rates, environmental policies, military actions, political strategies, neural network parameters, or whatever else is being chosen by the social planner. Now, your choice will presumably depend on the social consequences of the options, even if you also consider non-consequentialist factors. We can model the social consequences with a function f:X→S, where S is the set of social outcomes. In particular, if you choose an option x∈X, then the resulting social outcome would be f(x)∈S.
We call f:X→S the "social context". As a concrete example, suppose the options are different tax rates (say 10%, 20%, and 30%), and the social outcomes are characterized by variables like total tax revenue, income inequality, and unemployment rate. Then the social context is the function f:[0,1]→S which maps each tax rate x∈[0,1] to the resulting values of these social outcome variables.
A social choice principle should say, for each social context, which options are acceptable. Formally, a social choice principle is characterised by some function Ψ:(X→S)→P(X), which takes a social context f:X→S as input and returns a subset of the options Ψ(f)⊆X as output. Specifically, Ψ(f)⊆X consists of exactly those options which satisfy the principle in the social context f:X→S.
Note that (X→S) denotes the set of all functions from X to S, so Ψ is a higher-order function, meaning it receives another function as input. Additionally P(X) denotes the powerset of X, i.e. the set of subsets of X. We use the powerset P to allow for the fact that multiple options may satisfy a principle: if a principle Ψ permits only options x1 and x2 in a context f:X→S then Ψ(f)={x1,x2}. Finally, the powerset P(X) includes the empty set ∅, which allows for the case Ψ(f)=∅. Informally, Ψ(f)=∅ means that the social planner, following principle Ψ and faced with context f:X→S, has no acceptable options, which allows for principles that aren't universally satisfiable.
Here are some examples of social choice principle:
Context-independence Let X0⊆X be any fixed subset of the options, and consider the principle Ψ:f↦X0, which returns X0 regardless of the input f. Whether an option x∈X satisfies this principle depends only on whether x∈X0, and is independent of the social context. At one extreme, there's a trivial principle Ψ:f↦X which never constrains the social planner, and at the other extreme, there's a principle Ψ:f↦∅ which is always unsatisfiable. When X0={x0} consists of a single option, the principle Ψ states that the social planner must choose x0∈X regardless of the context.[2]
Targets Let Starget⊆S be any fixed subset of the social outcomes, whose elements we'll call targets. There is a principle which says that the social planner should choose an option which achieves a target. This principle is characterised by the function Ψ:f↦f−1(Starget), where f−1(Starget)⊆X denotes the preimage of Starget, i.e. f−1(Starget):={x∈X:f(x)∈Starget}. Note that if f(x)∉Starget for all x∈X then Ψ(f)=∅, i.e. if no option would achieve a target then the principle says all options are unacceptable.
Impact minimisation We can also characterise more unusual principles as functions (X→S)→P(X). For example, consider the principle that says a social planner should choose an option if most other options would've led to the same social outcome. Intuitively, this captures some notion of impact minimisation. Formally, this principle is characterised by the function Ψ(f):={x∈X∣#[x]f>#X2}, where #S denote the cardinality of a set and [x]f⊆X denotes the f-equivalency class of x∈X, i.e. [x]f={y∈X∣f(x)=f(y)}.
These examples illustrate the diversity of conceivable social choice principles. The key point is that they can all be represented by functions Ψ:(X→S)→P(X). I've found this a productive way to think about principles of decision-making, and agency more generally.[3] Finding compelling social choice principle is the central problem in social ethics, and different normative frameworks will propose different principles.
3. Two strategies for specifying principles
3.1. Utilitarian principles
Utilitarianism and aggregativism are two strategies for specifying a social choice principle Ψ:(X→S)→P(X). The utilitarian strategy specifies a social choice principle using two components:
A social utility function u:S→R that assigns a real-valued utility u(s)∈R to each social outcome s∈S.
The argmaxX operator, which maps a real-valued function r:X→R to the set of points that maximize it. Formally, argmaxX(r):={x∈X∣∀x′∈X:r(x′)≤r(x)}. Note that argmaxX(r) is a subset of X, possibly containing multiple points in case of ties, or no points in the case of unbounded functions.
Given the social utility function u:S→R and the operator argmaxX:(X→R)→P(X), the utilitarian principle is defined by Ψ(f):=argmaxX(u∘f). Note that if f:X→S is the social context, then the composition u∘f:X→R calculates the social utility resulting from each option, thereby providing a real-valued function r:X→R. The utilitarian principle f↦argmaxX(u∘f) says that the social planner should choose an option that maximizes this function.
As a simplistic example, consider a social utility function u:S→R that measures the gross world product of a social outcome. The resulting utilitarian principle f↦argmaxX(u∘f) would oblige maximizing gross world product. In practice, utilitarians typically endorse more nuanced utility functions that account for factors like individual well-being, fairness, and existential risk.
3.2. Aggregative principles
Aggregativism offers an alternative strategy to specifying social choice principle. Like utilitarianism, it defines the principle Ψ:(X→S)→P(X) using two components:
A function ζ:S→P that assigns a personal outcome ζ(s)∈P to each social outcome s∈S. We call ζ the social zeta function.
A model of a self-interested human, characterised by a function Π:(X→P)→P(X), explained below.
The function Π:(X→P)→P(X) should model a self-interested human in the following sense: for each personal context g:X→P the subset Π(f)⊆X should contain the options that the hypothetical human might choose in that context. A personal context g:X→P is an assignment of a personal outcome to each of the options, analogously to a social context. For example, if g:X→P maps some options to finding a dollar and the remaining options to drowning in a swamp, then presumably Π(g) contains only the former options.
Given the social zeta function ζ:S→P and a model of self-interested human Π:(X→P)→P(X), the aggregationist principle is defined by Ψ(f):=Π(ζ∘f). Note that if f:X→S is the social context, then the composition ζ∘f:X→P calculates the hypothetical personal outcome resulting from each option, thereby providing a personal context g:X→P. The aggregative principle f↦Π(ζ∘f) says that the social planner should choose an option a self-interested human might choose in this personal context.
For example, consider a social zeta function ζ:S→P that maps each social outcome s to the personal outcome of living every individual's life in sequence, starting with the earliest-born humans. The resulting aggregative principle f↦Π(ζ∘f) obliges affecting society such that living the concatenated lives is personally desirable.
3.3. Structural similarity between the two strategies
This comparison reveals the structural similarity between utilitarianism and aggregativism. Both strategies specify the principle Ψ using a two components:
A function mapping social outcomes to a different space, either R (in the case of utilitarianism) or P (in the case of aggregativism).
A choice principle in that different space, either maximization (in the case of utilitarianism) or a model of a self-interested human (in the case of aggregativism).
Both Π, the model of a self-interested human, and the argmaxX operator are choice principles: Π is a personal choice principle, it 'chooses' one the options based on their associated personal outcomes, and argmaxX is a real choice principle, it 'chooses' one of the options based on their associated real value. (Of course, argmaxX doesn’t literally choose anything, it’s simply a mathematical operator, but so too is Π.)
In general, for any space R, let's say an R-context is any function with type-signature X→R, and an R-choice principle is any function with type-signature (X→R)→P(X). That is, an R-choice principle Φ, when provided with an R-context r:X→R, identifies some subset Φ(f)⊆X of the options which are 'acceptable'.
How might one use an R-choice principle Φ to specify a social choice principle Ψ? Well, what's needed is some function σ:S→R from social outcomes to elements of R. This function σ will extends any social context f:X→S to an R-context σ∘f:X→R, which can then be provided to the R-choice principle to identify the acceptable options. Formally, Ψ:f↦Φ(σ∘f). This is how utilitarianism and aggregativism succeed in defining social choice principles. The key difference is that utilitarianism uses real numbers while aggregativism uses personal outcomes.
4. Equivalence between aggregativism and utilitarianism
4.1. Three conditions for equivalence
Despite their differences, there are natural conditions under which the utilitarian and aggregative principles are equivalent, in the sense that a social planner is permitted to choice an option, under the utilitarian principle, if and only if they are permitted to choice the same option under the aggregative principle.
Formally, let Ψu denote the utilitarian principle Ψu:f↦argmaxX(u∘f) and let Ψa denote the aggregative principle Ψa:f↦Π(ζ∘f); under what conditions does Ψu(f)=Ψa(f) for all social contexts f:X→S?
In the previous article, we showed that LELO, HL, and ROI each employ social zeta functions which aggregates the personal outcomes across all individuals in the population. Formally, ζ(s):=α(γ(−,s)M(π)) where I is a fixed set of individuals; γ:I×S→P is a fixed function mapping a social outcome s∈S and an individual i∈I to the personal outcome γ(i,s)∈P that i faces when s obtains; M is the monad capturing a notion of 'collection'; π∈M(I) be a fixed collection of individuals impartially representing the population; and α:M(P)→P is an M-algebra specifying how to aggregate collections of personal outcomes into a single personal outcome.
Supposing ζ has the general form above, and the three conditions below are satisfied, then the utilitarian principle Ψu and the aggregative principle Ψa are mathematically equivalent:
A self-interested human maximises personal utility.
Formally, the first condition states that the function Π:(X→P)→P(X) has the form Π(f)=argmaxX(v∘f) for some personal utility function v:P→R which assigns a real-valued utility v(p)∈R to each personal outcome p∈P. Even by itself, this condition is quite strong. It implies that if, for some personal context f:X→P, two options x1 and x2 result in the same personal outcome, i.e. f(x1)=f(x2), then the human might choose x1 if and only if they might choose x2. Hence, this condition precludes nonconsequential considerations.
Let's call this condition "Humans Maximise Personal Utility" (HMPU).
Personal utility is 'rational', in a technical sense defined below.
Let α:M(P)→P denote an M-algebra on personal outcomes, describing how to aggregate a collection of personal outcomes into a single personal outcome. Let β:M(R)→R denote an M-algebra on real numbers, describing how to aggregate a collection of real numbers into a single real number. The second condition states that v∘α=β∘vM. Informally, this condition means that the personal utility of an aggregate of personal outcomes is the aggregate of the personal utilities of each personal outcome being aggregated. In mathematical jargon, the personal utility function v:P→R must be a homomorphism between the M-algebras (P,α) and (R,β), which means it preserves the algebraic structure on P and R.
Let's call this condition "Rationality of Personal Utility" (RUP).
Social utility is the aggregate of personal utilities across all individual in the population.
Formally, the third condition states that u(s)=β((v∘γ)(−,s)M(π)), where v:P→R is the personal utility function introduced in HMPU, γ:I×S→P is the function assigning personal outcomes to each individual in each social outcome, π∈M(I) is the distinguished collection of individuals representing the population, and β:M(R)→R is the M-algebra describing how to aggregate a collection of real numbers into a single real number. Informally, this condition states that the social utility of a social outcome is the aggregate of the personal utilities of the personal outcomes faced by all individuals in the population.
Let's call this condition "Social Utility Aggregates Personal Utilities" (SUAPU).
The aggregative principle (when our model of a self-interested human is a rational personal utility maximiser) is equivalent to the utilitarian principle (when social utility is the impartial aggregation of personal utility over each individual). The full proof is elementary and uninsightful.[4]
Now, these three conditions are only approximately true, and they fail in systematic ways. However, the theorem will help elucidate exactly the extent to which the aggregative principle approximates the corresponding utilitarian principle. Namely, the aggregative principle will approximate the utilitarian principle to the degree that these conditions hold.
Because RPU and SUAPU depend on the specific monad M under discussion, I will spell out the details for three paradigm examples: the list monad List (representing finite sequences), the distribution monad Δ (representing probability distributions), and the nonempty finite powerset monad P+f (representing nonempty finite sets).
5. Equivalence between LELO and longtermist total utilitarianism
The previous section proved an equivalence, under certain conditions, between aggregative principles and utilitarian principles. This section will apply that theorem to the monad List, which is used to formalise Live Every Life Once (LELO). We will see that LELO is equivalent to longtermist total utilitarianism.
5.2. Monoidal rationality of personal utility?
The real numbers admit a concatenation operator in the obvious way, i.e., there exists a function sum:List(R)→R defined by sum([r1,…,rk]):=0+r1+⋯+rk. This is simply the well-known summation operator, which sends a list of real values to their sum.
Let's unpack RPU, which formally states that v∘conc=sum∘vList. In other words, for any list of personal outcomes [p1,…,pn], we have equality between v∘conc([p1,…,pn]) and 0+v(p1)+⋯+v(pn). Informally, the personal utility of a concatenated outcome equals the sum of the personal utilities of the outcomes being concatenated. This 'monoidal' rationality condition constrains how humans must value the concatenation of different personal outcomes.
In the previous article, we saw that the concatenation operator conc:List(P)→P can be equivalently presented by a binary operator ▹ and a constant ϵ∈P, with the intended interpretation p▹p′:=conc([p1,p2]) and ϵ:=conc([]). We can restate the RPU condition in terms of ▹ and ϵ with two equations: v(ϵ)=0 and v(p▹p′)=v(p)+v(p′) for all p,p′∈P.
How realistic is the RPU condition? That is, supposing humans do maximise a personal utility function, how monoidally rational is it? I think this condition is approximately true, but unrealistic in several ways. I'll assume here that p▹p′ is interpreted as facing p and then facing p′ in sequence, rather than some exotic notion of concatenation.
Firstly, RPU rules out permutation-dependent values. It precludes a personal utility function v:P→R such that v(p1▹p2)≠v(p2▹p1). Informally, RPU assumes human values must be invariant to the ordering of experiences: they cannot value saving the best till last, nor saving the worse till last. In particular, RPU assumes that humans values are time-symmetric, which seems unrealistic, as illustrated by the following examples. Compare the process of learning, i.e. ending with better beliefs than one started with, with the process of unlearning, i.e. ending with worse beliefs than one started with. Humans seem to value learning above unlearning, but such time-asymmetric values are precluded by RPU. Similarly, humans seem to value a history of improvement over a history of degradation, even if both histories are different permutations of the same list of moments, but such values are precluded by RPU.
Secondly, RPU rules out time-discounted values. Under exponential time-discounting, a common assumption in economics, the personal utility function v:P→R obeys the equation v(p1▹p2)=v(p1)+(1+δ)−duration(p1)⋅v(p2). Here duration:P→R≥0 gives the duration of each outcome and δ>0 is the discount rate. This discounting formula weights the first outcome p1 more than the second outcome p2, with the difference growing exponentially with the duration of p1. For instance, let p1 and p′1 be equally valuable experiences lasting different durations, like a minute of ecstasy and a week of contentment respectively. Time-discounting implies that v(p1▹p2) depends more on v(p2) than v(p′1▹p2) does. However, RPU precludes this possibility, as it requires that δ=0, i.e. that humans are equally concerned with all life stages, not discounting future rewards relative to present ones
Thirdly, RPU rules out path-dependent values. Informally, whether I value a future p more than a future q must be independent of my past experiences. But this is an unrealistic assumption about human values, as illustrated in the following examples. If p denotes reading Moby Dick and q denotes reading Oliver Twist, then humans seem to value p▹p less than p▹q but value q▹p more than q▹q. This is because humans value reading a book higher if they haven't already read it, due to an inherent value for novelty in reading material. Alternatively, if p and q denote being married to two different people, then humans seem to value p▹p more than p▹q but value q▹p less than q▹q. This is because humans value being married to someone for a decade higher if they've already been married to them, due to an inherent value for consistency in relationships.[5] But RPU would precludes such path-dependent values.
5.2. Social utility sums personal utility?
Now let's unpack SUAPU, which formally states that u(s)=sum((v∘γ(−,
1. Introduction
Utilitarianism is the view that a social planner should choose options which maximise the social utility of the resulting social outcome. The central object in utilitarianism is the social utility function u:S→R which assigns a real value u(s)∈R to each social outcome s∈S. This function typically involves variables such as the well-being, preferences, and mental states of individuals, distributional factors like inequality, and other relevant factors such as justice, social cohesion, and freedoms. Utilitarianism is a broad class of social choice principles, one corresponding to each function u:S→R.
In my previous article, I introduced aggregative principles, which state that a social planner should make decisions as if they will face the aggregated personal outcomes of every individual in the population. The central object in aggregativism is the function ζ:S→P, represented with the the greek letter zeta, which assigns a personal outcome ζ(s)∈P to each social outcome s∈S. This function typically aggregates the collection of personal outcomes facing the entire population into a single personal outcome. Aggregativism is a broad class of social choice principles, one corresponding to each function ζ:S→P.
We examined three well-known aggregative principles:
I'm interested in aggregative principles because they avoid many theoretical pitfalls of utilitarian principles. Unlike utilitarianism, aggregativism doesn't require specifying a social welfare function, which is notoriously intractable. Moreover, it seems less prone to counterintuitive conclusions such as the repugnant conclusion or the violation of moral side constraints.[1] In this article, I will show that, under natural conditions of human rationality, aggregative principles approximate utilitarian principles. Therefore, even though aggregativism avoids these theoretical pitfalls, we should nonetheless expect aggregativism to generate roughly-utilitarian recommendations in practical social contexts, and thereby retain the most appealing insights from utilitarianism.
The rest of the article is organized as follows. Section 2 formalises social choice principles as functions of type (X→S)→P(X). Section 3 demonstrates the structural similarity between two strategies to specifying such principles, namely the aggregative and utilitarian strategies. Section 4 proves that under natural conditions about human rationality, the aggregative and utilitarian principles are mathematically equivalent. This theorem is the key contribution of the article. Sections 5, 6, and 7 applies the theorem to LELO, HL, and ROI respectively.
2. Social choice principles
Suppose you are a social planner choosing from a set of options X={x1,…,xn}. The set X might be the set of available tax rates, environmental policies, military actions, political strategies, neural network parameters, or whatever else is being chosen by the social planner. Now, your choice will presumably depend on the social consequences of the options, even if you also consider non-consequentialist factors. We can model the social consequences with a function f:X→S, where S is the set of social outcomes. In particular, if you choose an option x∈X, then the resulting social outcome would be f(x)∈S.
We call f:X→S the "social context". As a concrete example, suppose the options are different tax rates (say 10%, 20%, and 30%), and the social outcomes are characterized by variables like total tax revenue, income inequality, and unemployment rate. Then the social context is the function f:[0,1]→S which maps each tax rate x∈[0,1] to the resulting values of these social outcome variables.
A social choice principle should say, for each social context, which options are acceptable. Formally, a social choice principle is characterised by some function Ψ:(X→S)→P(X), which takes a social context f:X→S as input and returns a subset of the options Ψ(f)⊆X as output. Specifically, Ψ(f)⊆X consists of exactly those options which satisfy the principle in the social context f:X→S.
Note that (X→S) denotes the set of all functions from X to S, so Ψ is a higher-order function, meaning it receives another function as input. Additionally P(X) denotes the powerset of X, i.e. the set of subsets of X. We use the powerset P to allow for the fact that multiple options may satisfy a principle: if a principle Ψ permits only options x1 and x2 in a context f:X→S then Ψ(f)={x1,x2}. Finally, the powerset P(X) includes the empty set ∅, which allows for the case Ψ(f)=∅. Informally, Ψ(f)=∅ means that the social planner, following principle Ψ and faced with context f:X→S, has no acceptable options, which allows for principles that aren't universally satisfiable.
Here are some examples of social choice principle:
Let X0⊆X be any fixed subset of the options, and consider the principle Ψ:f↦X0, which returns X0 regardless of the input f. Whether an option x∈X satisfies this principle depends only on whether x∈X0, and is independent of the social context. At one extreme, there's a trivial principle Ψ:f↦X which never constrains the social planner, and at the other extreme, there's a principle Ψ:f↦∅ which is always unsatisfiable. When X0={x0} consists of a single option, the principle Ψ states that the social planner must choose x0∈X regardless of the context.[2]
Let Starget⊆S be any fixed subset of the social outcomes, whose elements we'll call targets. There is a principle which says that the social planner should choose an option which achieves a target. This principle is characterised by the function Ψ:f↦f−1(Starget), where f−1(Starget)⊆X denotes the preimage of Starget, i.e. f−1(Starget):={x∈X:f(x)∈Starget}. Note that if f(x)∉Starget for all x∈X then Ψ(f)=∅, i.e. if no option would achieve a target then the principle says all options are unacceptable.
We can also characterise more unusual principles as functions (X→S)→P(X). For example, consider the principle that says a social planner should choose an option if most other options would've led to the same social outcome. Intuitively, this captures some notion of impact minimisation. Formally, this principle is characterised by the function Ψ(f):={x∈X∣#[x]f>#X2}, where #S denote the cardinality of a set and [x]f⊆X denotes the f-equivalency class of x∈X, i.e. [x]f={y∈X∣f(x)=f(y)}.
These examples illustrate the diversity of conceivable social choice principles. The key point is that they can all be represented by functions Ψ:(X→S)→P(X). I've found this a productive way to think about principles of decision-making, and agency more generally.[3] Finding compelling social choice principle is the central problem in social ethics, and different normative frameworks will propose different principles.
3. Two strategies for specifying principles
3.1. Utilitarian principles
Utilitarianism and aggregativism are two strategies for specifying a social choice principle Ψ:(X→S)→P(X). The utilitarian strategy specifies a social choice principle using two components:
Given the social utility function u:S→R and the operator argmaxX:(X→R)→P(X), the utilitarian principle is defined by Ψ(f):=argmaxX(u∘f). Note that if f:X→S is the social context, then the composition u∘f:X→R calculates the social utility resulting from each option, thereby providing a real-valued function r:X→R. The utilitarian principle f↦argmaxX(u∘f) says that the social planner should choose an option that maximizes this function.
As a simplistic example, consider a social utility function u:S→R that measures the gross world product of a social outcome. The resulting utilitarian principle f↦argmaxX(u∘f) would oblige maximizing gross world product. In practice, utilitarians typically endorse more nuanced utility functions that account for factors like individual well-being, fairness, and existential risk.
3.2. Aggregative principles
Aggregativism offers an alternative strategy to specifying social choice principle. Like utilitarianism, it defines the principle Ψ:(X→S)→P(X) using two components:
The function Π:(X→P)→P(X) should model a self-interested human in the following sense: for each personal context g:X→P the subset Π(f)⊆X should contain the options that the hypothetical human might choose in that context. A personal context g:X→P is an assignment of a personal outcome to each of the options, analogously to a social context. For example, if g:X→P maps some options to finding a dollar and the remaining options to drowning in a swamp, then presumably Π(g) contains only the former options.
Given the social zeta function ζ:S→P and a model of self-interested human Π:(X→P)→P(X), the aggregationist principle is defined by Ψ(f):=Π(ζ∘f). Note that if f:X→S is the social context, then the composition ζ∘f:X→P calculates the hypothetical personal outcome resulting from each option, thereby providing a personal context g:X→P. The aggregative principle f↦Π(ζ∘f) says that the social planner should choose an option a self-interested human might choose in this personal context.
For example, consider a social zeta function ζ:S→P that maps each social outcome s to the personal outcome of living every individual's life in sequence, starting with the earliest-born humans. The resulting aggregative principle f↦Π(ζ∘f) obliges affecting society such that living the concatenated lives is personally desirable.
3.3. Structural similarity between the two strategies
This comparison reveals the structural similarity between utilitarianism and aggregativism. Both strategies specify the principle Ψ using a two components:
Both Π, the model of a self-interested human, and the argmaxX operator are choice principles: Π is a personal choice principle, it 'chooses' one the options based on their associated personal outcomes, and argmaxX is a real choice principle, it 'chooses' one of the options based on their associated real value. (Of course, argmaxX doesn’t literally choose anything, it’s simply a mathematical operator, but so too is Π.)
In general, for any space R, let's say an R-context is any function with type-signature X→R, and an R-choice principle is any function with type-signature (X→R)→P(X). That is, an R-choice principle Φ, when provided with an R-context r:X→R, identifies some subset Φ(f)⊆X of the options which are 'acceptable'.
How might one use an R-choice principle Φ to specify a social choice principle Ψ? Well, what's needed is some function σ:S→R from social outcomes to elements of R. This function σ will extends any social context f:X→S to an R-context σ∘f:X→R, which can then be provided to the R-choice principle to identify the acceptable options. Formally, Ψ:f↦Φ(σ∘f). This is how utilitarianism and aggregativism succeed in defining social choice principles. The key difference is that utilitarianism uses real numbers while aggregativism uses personal outcomes.
4. Equivalence between aggregativism and utilitarianism
4.1. Three conditions for equivalence
Despite their differences, there are natural conditions under which the utilitarian and aggregative principles are equivalent, in the sense that a social planner is permitted to choice an option, under the utilitarian principle, if and only if they are permitted to choice the same option under the aggregative principle.
Formally, let Ψu denote the utilitarian principle Ψu:f↦argmaxX(u∘f) and let Ψa denote the aggregative principle Ψa:f↦Π(ζ∘f); under what conditions does Ψu(f)=Ψa(f) for all social contexts f:X→S?
In the previous article, we showed that LELO, HL, and ROI each employ social zeta functions which aggregates the personal outcomes across all individuals in the population. Formally, ζ(s):=α(γ(−,s)M(π)) where I is a fixed set of individuals; γ:I×S→P is a fixed function mapping a social outcome s∈S and an individual i∈I to the personal outcome γ(i,s)∈P that i faces when s obtains; M is the monad capturing a notion of 'collection'; π∈M(I) be a fixed collection of individuals impartially representing the population; and α:M(P)→P is an M-algebra specifying how to aggregate collections of personal outcomes into a single personal outcome.
Supposing ζ has the general form above, and the three conditions below are satisfied, then the utilitarian principle Ψu and the aggregative principle Ψa are mathematically equivalent:
Formally, the first condition states that the function Π:(X→P)→P(X) has the form Π(f)=argmaxX(v∘f) for some personal utility function v:P→R which assigns a real-valued utility v(p)∈R to each personal outcome p∈P. Even by itself, this condition is quite strong. It implies that if, for some personal context f:X→P, two options x1 and x2 result in the same personal outcome, i.e. f(x1)=f(x2), then the human might choose x1 if and only if they might choose x2. Hence, this condition precludes nonconsequential considerations.
Let's call this condition "Humans Maximise Personal Utility" (HMPU).
Let α:M(P)→P denote an M-algebra on personal outcomes, describing how to aggregate a collection of personal outcomes into a single personal outcome. Let β:M(R)→R denote an M-algebra on real numbers, describing how to aggregate a collection of real numbers into a single real number. The second condition states that v∘α=β∘vM. Informally, this condition means that the personal utility of an aggregate of personal outcomes is the aggregate of the personal utilities of each personal outcome being aggregated. In mathematical jargon, the personal utility function v:P→R must be a homomorphism between the M-algebras (P,α) and (R,β), which means it preserves the algebraic structure on P and R.
Let's call this condition "Rationality of Personal Utility" (RUP).
Formally, the third condition states that u(s)=β((v∘γ)(−,s)M(π)), where v:P→R is the personal utility function introduced in HMPU, γ:I×S→P is the function assigning personal outcomes to each individual in each social outcome, π∈M(I) is the distinguished collection of individuals representing the population, and β:M(R)→R is the M-algebra describing how to aggregate a collection of real numbers into a single real number. Informally, this condition states that the social utility of a social outcome is the aggregate of the personal utilities of the personal outcomes faced by all individuals in the population.
Let's call this condition "Social Utility Aggregates Personal Utilities" (SUAPU).
The aggregative principle (when our model of a self-interested human is a rational personal utility maximiser) is equivalent to the utilitarian principle (when social utility is the impartial aggregation of personal utility over each individual). The full proof is elementary and uninsightful.[4]
Now, these three conditions are only approximately true, and they fail in systematic ways. However, the theorem will help elucidate exactly the extent to which the aggregative principle approximates the corresponding utilitarian principle. Namely, the aggregative principle will approximate the utilitarian principle to the degree that these conditions hold.
Because RPU and SUAPU depend on the specific monad M under discussion, I will spell out the details for three paradigm examples: the list monad List (representing finite sequences), the distribution monad Δ (representing probability distributions), and the nonempty finite powerset monad P+f (representing nonempty finite sets).
5. Equivalence between LELO and longtermist total utilitarianism
The previous section proved an equivalence, under certain conditions, between aggregative principles and utilitarian principles. This section will apply that theorem to the monad List, which is used to formalise Live Every Life Once (LELO). We will see that LELO is equivalent to longtermist total utilitarianism.
5.2. Monoidal rationality of personal utility?
The real numbers admit a concatenation operator in the obvious way, i.e., there exists a function sum:List(R)→R defined by sum([r1,…,rk]):=0+r1+⋯+rk. This is simply the well-known summation operator, which sends a list of real values to their sum.
Let's unpack RPU, which formally states that v∘conc=sum∘vList. In other words, for any list of personal outcomes [p1,…,pn], we have equality between v∘conc([p1,…,pn]) and 0+v(p1)+⋯+v(pn). Informally, the personal utility of a concatenated outcome equals the sum of the personal utilities of the outcomes being concatenated. This 'monoidal' rationality condition constrains how humans must value the concatenation of different personal outcomes.
In the previous article, we saw that the concatenation operator conc:List(P)→P can be equivalently presented by a binary operator ▹ and a constant ϵ∈P, with the intended interpretation p▹p′:=conc([p1,p2]) and ϵ:=conc([]). We can restate the RPU condition in terms of ▹ and ϵ with two equations: v(ϵ)=0 and v(p▹p′)=v(p)+v(p′) for all p,p′∈P.
How realistic is the RPU condition? That is, supposing humans do maximise a personal utility function, how monoidally rational is it? I think this condition is approximately true, but unrealistic in several ways. I'll assume here that p▹p′ is interpreted as facing p and then facing p′ in sequence, rather than some exotic notion of concatenation.
Firstly, RPU rules out permutation-dependent values. It precludes a personal utility function v:P→R such that v(p1▹p2)≠v(p2▹p1). Informally, RPU assumes human values must be invariant to the ordering of experiences: they cannot value saving the best till last, nor saving the worse till last. In particular, RPU assumes that humans values are time-symmetric, which seems unrealistic, as illustrated by the following examples. Compare the process of learning, i.e. ending with better beliefs than one started with, with the process of unlearning, i.e. ending with worse beliefs than one started with. Humans seem to value learning above unlearning, but such time-asymmetric values are precluded by RPU. Similarly, humans seem to value a history of improvement over a history of degradation, even if both histories are different permutations of the same list of moments, but such values are precluded by RPU.
Secondly, RPU rules out time-discounted values. Under exponential time-discounting, a common assumption in economics, the personal utility function v:P→R obeys the equation v(p1▹p2)=v(p1)+(1+δ)−duration(p1)⋅v(p2). Here duration:P→R≥0 gives the duration of each outcome and δ>0 is the discount rate. This discounting formula weights the first outcome p1 more than the second outcome p2, with the difference growing exponentially with the duration of p1. For instance, let p1 and p′1 be equally valuable experiences lasting different durations, like a minute of ecstasy and a week of contentment respectively. Time-discounting implies that v(p1▹p2) depends more on v(p2) than v(p′1▹p2) does. However, RPU precludes this possibility, as it requires that δ=0, i.e. that humans are equally concerned with all life stages, not discounting future rewards relative to present ones
Thirdly, RPU rules out path-dependent values. Informally, whether I value a future p more than a future q must be independent of my past experiences. But this is an unrealistic assumption about human values, as illustrated in the following examples. If p denotes reading Moby Dick and q denotes reading Oliver Twist, then humans seem to value p▹p less than p▹q but value q▹p more than q▹q. This is because humans value reading a book higher if they haven't already read it, due to an inherent value for novelty in reading material. Alternatively, if p and q denote being married to two different people, then humans seem to value p▹p more than p▹q but value q▹p less than q▹q. This is because humans value being married to someone for a decade higher if they've already been married to them, due to an inherent value for consistency in relationships.[5] But RPU would precludes such path-dependent values.
5.2. Social utility sums personal utility?
Now let's unpack SUAPU, which formally states that u(s)=sum((v∘γ(−,