Moral patients (Topic archive) - 80,000 Hours https://80000hours.org/topic/foundations/moral-philosophy/moral-patients/ Tue, 22 Apr 2025 11:59:50 +0000 en-US hourly 1 https://wordpress.org/?v=6.7.1 Will MacAskill on AI causing a “century in a decade” — and how we’re completely unprepared https://80000hours.org/podcast/episodes/will-macaskill-century-in-a-decade-navigating-intelligence-explosion/ Tue, 11 Mar 2025 17:22:02 +0000 https://80000hours.org/?post_type=podcast&p=88961 The post Will MacAskill on AI causing a “century in a decade” — and how we’re completely unprepared appeared first on 80,000 Hours.

]]>
The post Will MacAskill on AI causing a “century in a decade” — and how we’re completely unprepared appeared first on 80,000 Hours.

]]>
Cameron Meyer Shorb on dismantling the myth that we can’t do anything to help wild animals https://80000hours.org/podcast/episodes/cameron-meyer-shorb-wild-animal-suffering/ Fri, 29 Nov 2024 22:45:28 +0000 https://80000hours.org/?post_type=podcast&p=88203 The post Cameron Meyer Shorb on dismantling the myth that we can’t do anything to help wild animals appeared first on 80,000 Hours.

]]>
The post Cameron Meyer Shorb on dismantling the myth that we can’t do anything to help wild animals appeared first on 80,000 Hours.

]]>
Understanding the moral status of digital minds https://80000hours.org/problem-profiles/moral-status-digital-minds/ Wed, 11 Sep 2024 00:55:28 +0000 https://80000hours.org/?post_type=problem_profile&p=77546 The post Understanding the moral status of digital minds appeared first on 80,000 Hours.

]]>
Why might understanding the moral status of digital minds be an especially pressing problem?

1. Humanity may soon grapple with many AI systems that could be conscious

In 2020, more than 1,000 professional philosophers were asked whether they believed then-current AI systems1 were conscious.2 Consciousness, in this context, is typically understood as meaning having phenomenal experiences that feel like something, like the experience of perception or thinking.

Less than 1% said that yes, some then-current AI systems were conscious, and about 3% said they were “leaning” toward yes. About 82% said no or leaned toward no.

But when asked about whether some future AI systems would be conscious, the bulk of opinion flipped.

Nearly 40% were inclined to think future AI systems would be conscious, while only about 27% were inclined to think they wouldn’t be.3

A survey of 166 attendees at the Association for the Scientific Study of Consciousness annual conference asked a similar question. Sixty-seven percent of attendees answered “definitely yes” or “probably yes” when asked “At present or in the future, could machines (e.g. robots) have consciousness?”4

The plurality of philosophers and majority of conference attendees might be wrong. But we think these kinds of results make it very difficult to rule out the possibility of conscious AI systems, and we think it’s wrong to confidently assert that no AI could ever be conscious.

Why might future AI systems be conscious? This question is wide open, but researchers have made some promising steps toward providing answers.

One of the most rigorous and comprehensive studies we’ve seen into this issue was published in August 2023 with 19 authors, including experts in AI, neuroscience, cognitive science, and philosophy.5 They investigated a range of properties6 that could indicate that AI systems are conscious.

The authors concluded: “Our analysis suggests that no current AI systems are conscious, but also suggests that there are no obvious technical barriers to building AI systems which satisfy these indicators.”

They also found that, according to some plausible theories of consciousness, “conscious AI systems could realistically be built in the near term.”

Philosopher David Chalmers has suggested that there’s a (roughly) 25% chance that in the next decade we’ll have conscious AI systems.7

Creating increasingly powerful AI systems — as frontier AI companies are currently trying to do — may require features that some researchers think would indicate consciousness. For example, proponents of global workspace theory argue that animals have conscious states when their specialised cognitive systems (e.g. sensory perception, memories, etc.) are integrated in the right way into a mind and share representations of information in a “global workspace.” It’s possible that creating such a “workspace” in an AI system would both increase its capacity to do cognitive tasks and make it a conscious being. Similar claims might be made about other features and theories of consciousness.

And it wouldn’t be too surprising if increasing cognitive sophistication led to consciousness in this way, because humans’ cognitive abilities seem closely associated with our capacity for consciousness.8 (Though, as we’ll discuss below, intelligence and consciousness are distinct concepts.)9

How soon could conscious AI systems arrive? We’re not sure. But we do seem to be on track to make a huge number of more advanced AI systems in the coming decades.

Another recent survey found that the aggregate forecast of the thousands of AI researchers put a 50% chance that we’ll have AI systems that are better than humans in every possible task by 2047.10

If we do produce systems that capable, there will be enormous incentives to produce many of them. So we might be looking at a world with a huge number of highly advanced AI systems, which philosophers think could be conscious, pretty soon.

The public may already be more inclined to assign attributes like consciousness to AI systems than experts. Around 18% of US respondents in a 2023 survey believed current AI systems are sentient.11

This phenomenon might already have real effects on people’s lives. Some chatbot services have cultivated devoted user bases that engage in emotional and romantic interactions with AI-powered characters, with many seeming to believe — implicitly or explicitly — that the AI may reciprocate their feelings.12

As people increasingly think AI systems may be conscious or sentient, we’ll face the question of whether humans have any moral obligations to these digital minds. Indeed, among the 76% of US survey respondents who said AI sentience was possible (or that they weren’t sure if it was possible), 81% said they expected “the welfare of robots/AIs to be an important social issue” within 20 years.

We may start to ask:

  • Are certain methods of training AIs cruel?
  • Can we use AIs for our own ends in an ethical way?
  • Do AI systems deserve moral and political rights?

These may be really difficult questions, which involve complex issues in philosophy, political theory, cognitive science, computer science, machine learning, and other fields. A range of possible views about these issues could be reasonable. We could also imagine getting the answers to these questions drastically wrong.

And with economic incentives to create these AI systems, and many humans — including experts in the field — prepared to believe they could be conscious, it seems unlikely we will be able to avoid the hard questions.

Clearing up common misconceptions

There’s a common misconception that worries about AI risk are generally driven by fear that AI systems will at some point “wake up,” become sentient, and then turn against humanity.

However, as our article on preventing an AI-related catastrophe explains, the possibility of AI systems becoming sentient is not a central or necessary part of the argument that advanced AI could pose an existential risk. Many AI risk scenarios are possible regardless of whether or not AI systems can be sentient or have moral status. One of the primary scenarios our article discusses is the risk that power-seeking AI systems could seek to disempower or eradicate humanity, if they’re misaligned with our purposes.

This article discusses how some concerns around the moral status of digital minds might contribute to the risk that AI poses to humanity, and why we should be concerned about potential risks to digital minds themselves. But it’s important to make clear that in principle these two sources of risk are distinct. Even if you concluded the arguments in this article were mistaken, you might still think the possibility of an AI-related catastrophe is a genuine risk (and vice versa).

Read more about preventing an AI-related catastrophe

It’s also important to note that while creating increasingly capable and intelligent AI systems may result in conscious digital minds, intelligence can conceptually be decoupled from consciousness and sentience. It’s plausible that we could have AI systems that are more intelligent than, say, mice, on most if not all dimensions. But we might still think mice are more likely to be sentient than the AI system. It may likewise be true that some less intelligent or capable AI systems would be regarded as more plausibly sentient than some other systems that were more intelligent, perhaps because of differences in their internal architecture.9

2. Creating digital minds could go very badly — or very well

One thing that makes this problem particularly thorny is the risk of both over-attributing and under-attributing moral status.

Believing AI systems aren’t worthy of moral consideration when they are and the reverse could both be disastrous. There are potential dangers for both digital minds and for humans.

Dangers for digital minds

If we falsely think digital minds don’t have moral status when they do, we could unknowingly force morally significant beings into conditions of servitude and extreme suffering — or otherwise mistreat them.

Some ways this could happen include:

  • The process of aligning or controlling digital minds to act in their creators’ interests could involve suffering, frequent destruction, or manipulation in ways that are morally wrong.13
  • Our civilisation could choose to digitally simulate its own histories or other scenarios, in which fully simulated digital minds might suffer in extreme amounts — a possibility Nick Bostrom has raised.14
  • Philosophers Eric Schwitzgebel and Mara Garza have argued that even if we avoid creating large-scale suffering, we should be concerned about a future full of “cheerful servant” digital minds. They might in principle deserve rights and freedoms, but we could design them to seem happy with oppression and disregard. On many moral views, this could be deeply unjust.

These bad outcomes seem most likely to happen by accident or out of ignorance, perhaps by failing to recognise digital sentience. But some people might knowingly cause large numbers of digital minds to suffer out of indifference, sadism, or some other reason. And it’s possible some AI systems might cause other AI systems to suffer, perhaps as a means of control or to further their own objectives.

Dangers for humans

There are also dangers to humans. For example, if we believe AI systems are sentient when they are not, and when they in fact lack any moral status, we could do any of the following:

  • We could waste resources trying to meet the needs and desires of AI systems even if there’s no real reason to do so.
    • This could be costly, and it could take resources away from causes that genuinely need them.
  • We might choose to give AI systems freedom, rather than control them. This plausibly could lead to an existential catastrophe.
    • For example, key decision makers might believe that the possibility discussed in the previous section that AI alignment and AI control could be harmful to digital minds. If they were mistaken, they might forgo necessary safety measures in creating advanced AI, and then that AI could seek to disempower humanity. If the decision makers are correct about the moral risks to digital minds, then the wise choice might be to delay development until we have enough knowledge to pursue AI development safely for everyone.
  • Even more speculatively, humanity might decide at some point in the future to “upload” our minds — choosing to be replaced by digital versions of ourselves. If it turned out that these uploaded versions of our minds wouldn’t be conscious, this could turn out to be a severe mistake.15

It’s hard to be confident in the plausibility of any particular scenario, but these kinds of cases illustrate the potential scale of the risks.

Other dangers

If the world is truly unfortunate, we could even make both kinds of errors at once. We could have charismatic systems (which perhaps act in a humanlike way) that we believe are sentient when they’re not. At the same time, we could have less charismatic but sentient systems whose suffering and interests are completely disregarded. For example, maybe AI systems that don’t talk will be disregarded, even if they are worthy of just as much moral concern as others.9

We could also make a moral mistake by missing important opportunities. It’s possible we’ll have the opportunity to create digital minds with extremely valuable lives with varied and blissful experiences, continuing indefinitely. Failing to live up to this potential could be a catastrophic mistake on some moral views. And yet, for whatever reason, we might decide not to.

Things could also go well

This article is primarily about encouraging research to reduce major risks. But it’s worth making clear that we think there are many possible good futures:

  • We might eventually create flourishing, friendly, joyful digital minds with whom humanity could share the future.
  • Or we might discover that the most useful AI systems we can build don’t have moral status, and we can justifiably use them to improve the world without worrying about their wellbeing.

What should we take from all this? The risks of both over-attribution and under-attribution of sentience and moral status mean that we probably shouldn’t simply stake out an extreme position and rally supporters around it. We shouldn’t, for example, declare that all AI systems that pass a simple benchmark must be given rights equivalent to humans or insist that any human’s interests always come before those of digital minds.

Instead, our view is that this problem requires much more research to clarify key questions, to dispel as much uncertainty as possible, and to determine the best paths forward despite the remaining uncertainty we’ll have. This is the best hope we have of avoiding key failure modes and increasing the chance that the future goes well.

But we face a lot of challenges in doing this, which we turn to next.

3. We don’t know how to assess the moral status of AI systems

The supercomputer MareNostrum-4 at the National Supercomputing Center in Barcelona, Spain. Martidaniel, CC BY-SA 4.0, via Wikimedia Commons

So it seems likely that we’ll create conscious digital minds at some point, or at the very least that many people may come to believe AI systems are conscious.

The trouble is that we don’t know how to figure out if an AI system is conscious — or whether it has moral status.

Even with animals, the scientific and philosophical community is unsure. Do insects have conscious experiences? What about clams? Jellyfish? Snails?16

And there’s also no consensus about how we should assess a being’s moral status. Being conscious may be all that’s needed for being worthy of moral consideration, but some think it’s necessary to be sentient — that is, being able to have good and bad conscious experiences. Some think that consciousness isn’t even necessary to have moral status because an individual agent may, for example, have morally important desires and goals without being conscious.

So we’re left with three big, open questions:

  • What characteristics would make a digital mind a moral patient?
  • Can a digital mind have those characteristics (for example, being conscious)?
  • How do we figure out if any given AI has these characteristics?

These questions are hard, and it’s not even always obvious what kind of evidence would settle them.

Some people believe these questions are entirely intractable, but we think that’s too pessimistic. Other areas in science and philosophy may have once seemed completely insoluble, only to see great progress when people discover new ways of tackling the questions.

Still, the state of our knowledge on these important questions is worryingly poor.

There are many possible characteristics that give rise to moral status

Some think moral status comes from having:17

  • Consciousness: the capacity to have subjective experience, but not necessarily valenced (positive or negative) experience. An entity might be conscious if it has perceptual experiences of the world, such as experiences of colour or physical sensations like heat. Often consciousness is described as the phenomenon of there being something it feels like to be you — to have your particular perspective on the world, to have thoughts, to feel the wind on your face — in a way that inanimate objects like rocks seem to completely lack. Some people think it’s all you need to be a moral patient, though it’s arguably hard to see how one could harm or benefit a conscious being without valenced experiences.18
  • Sentience: the capacity to have subjective experience (that is, consciousness as just defined) and the capacity for valenced experiences, i.e. good or bad feelings. Physical pleasure and pain are the typical examples of valenced, conscious experiences, but there are others, such as anxiety or excitement.
  • Agency: the ability to have and act on goals, reasons, or desires, or something like them. An entity might be able to have agency without being conscious or sentient. And some believe even non-conscious beings could have moral status by having agency, since they could be harmed or benefited depending on whether their goals are frustrated or achieved.19
  • Personhood: personhood is a complex and debated term that usually refers to a collection of properties, which often include sentience, agency, rational deliberation, and the ability to respond to reasons. Historically, personhood has sometimes been considered a necessary and sufficient criterion for moral status or standing, particularly in law. But this view has become less favoured in philosophy as it leaves no room for obligations to most nonhuman animals, human babies, and some others.
  • Some combination of the above or other traits.20

We think it’s most plausible that any being that feels good or bad experiences — like pleasure or pain — is worthy of moral concern in their own right.21

We discuss this more in our article on the definition of social impact, which touches on the history of moral philosophy.

But we don’t think we or others should be dogmatic about this, and we should look for sensible approaches to accommodate a range of reasonable opinions on these controversial subjects.22

Drawing of the brain by Sir Charles Bell (1774-1842). CC BY 4.0, via the Wellcome Collection

Many plausible theories of consciousness could include digital minds

There are many theories of consciousness — more than we can name here. What’s relevant is that some, though not all, theories of consciousness do imply the possibility of conscious digital minds.

This is only relevant if you think consciousness (or sentience, which includes consciousness as a necessary condition) is required for moral status. But since this is a commonly held view, it’s worth considering these theories and their implications. (Note, though, that there are often many variants of any particular theory.)

Some theories that could rule out the possibility of conscious digital minds:23

  • Biological theories: These hold that consciousness is inherently tied to the biological processes of the brain that can’t be replicated in computer hardware.24
  • Some forms of dualism: Dualism, particularly substance dualism, holds that consciousness is a non-physical substance distinct from the physical body and brain. It is often associated with religious traditions. While some versions of dualism would accommodate the existence of conscious digital minds, others could rule out the possibility.25

Some theories that imply digital minds could be conscious:

  • Functionalism: This theory holds that mental states are defined by their functional roles — how they process inputs, outputs, and interactions with other mental states. Consciousness, from this perspective, is explained not by what a mind is made of but by the functional organisation of its constituents. Some forms of functionalism, such as computational functionalism, strongly suggest that digital minds could be conscious, as they imply that if a digital system replicates the functional organisation of a conscious brain, it could also have conscious mental experiences.
  • Global workspace theory: GWT says that consciousness is the result of integrating information in a “global workspace” within the brain, where different processes compete for attention and are broadcast to other parts of the system. If a digital mind can replicate this global workspace architecture, GWT would support the possibility that the digital mind could be conscious.
  • Higher-order thought theory: HOT theory holds that consciousness arises when a mind has thoughts about its own mental states. On this view, it’s plausible that if a digital mind could be designed to have thoughts about its own processes and mental states, it would therefore be conscious.
  • Integrated information theory: IIT posits that consciousness corresponds to the level of integrated information within a system. A system is conscious to the extent that it has a high degree of integrated information (often denoted ‘Φ’). Like biological systems, digital minds could potentially be conscious if they integrate information with sufficiently high Φ.26

Some theories that are agnostic or unclear about digital minds:

  • Quantum theories of consciousness: Roger Penrose theorises that consciousness is tied to quantum phenomena within the brain.27 If so, digital minds may not be able to be conscious unless their hardware can replicate these quantum processes.
  • Panpsychism: Panpsychism is the view that consciousness is a fundamental property of the universe. Panpsychism doesn’t rule out digital minds being conscious, but it doesn’t necessarily provide a clear framework for understanding how or when a digital system might become conscious.
  • Illusionism or eliminativism: Illusionists or eliminativists argue that consciousness, as it is often understood, is an illusion or unnecessary folk theory. Illusionism doesn’t necessarily rule out digital minds being “conscious” in some sense, but it suggests that consciousness isn’t what we usually think it is. But many illusionists and eliminativists don’t want to deny that humans and animals can have moral status according to their views — in which case they might also be open to the idea that digital minds could likewise have moral status. (See some discussion of this issue here.)

It can be reasonable, especially for experts with deep familiarity of the debates, to believe much more strongly in one theory than the others. But given the amount of disagreement about this topic among experts, and the lack of solid evidence in one direction or another, and since many widely supported theories imply that digital minds could be conscious (or at least don’t contradict the idea), we don’t think it’s reasonable to completely rule out the possibility of conscious digital minds.28

We think it makes sense to put at least 5% on the possibility. Speaking as the author of this piece, based on my subjective impression of the balance of the arguments, I’d put the chance at around 50% at least.

We can’t rely on what AI systems tell us about themselves

Unfortunately, we can’t just rely on self-reports from AI systems about whether they’re conscious or sentient.

In the case of large language models like LaMDA, we don’t know why it claimed under certain conditions to Blake Lemoine that it was sentient,29 but it resulted in some way from having been trained on a huge body of existing texts.30

LLMs essentially learn patterns and trends in these texts, and then respond to questions on the basis of these extremely complex patterns of associations. The capabilities produced by this process are truly impressive — though we don’t fully understand how this process works, the outputs end up reflecting human knowledge about the world. As a result, the models can perform reasonably well at tasks involving human-like reasoning and making accurate statements about the world. (Though they still have many flaws!)

However, the process of learning from human text and fine-tuning might not have any relationship with what it’s actually like to be a language model. Rather, the responses seem more likely to mirror our own speculations and lack of understanding about the inner workings and experiences of AI systems.31

That means we can’t simply trust an AI system like LaMDA when it says it’s sentient.32

Researchers have proposed methods to assess the internal states of AI systems and whether they might be conscious or sentient, but all of these methods have serious drawbacks, at least at the moment:

  • Behavioural tests: we might try to figure out if an AI system is conscious by observing its outputs and actions to see if they indicate consciousness. The familiar Turing Test is one example; researchers such as Susan Schneider have proposed others. But since such tests can likely be gamed by a smart enough AI system that is nevertheless not conscious, even sophisticated versions may leave room for reasonable doubt.
  • Theory-based analysis: another method involves assessing the internal structure of AI systems and determining whether they show the “indicator properties” of existing theories of consciousness. The paper discussed above by Butlin et al. took this approach. While this method avoids the risk of being gamed by intelligent but non-conscious AIs, it is only as good as the (highly contested) theories it relies on and our ability to discern the indicator properties.
  • Animal analogue comparisons: we can also compare the functional architecture of AI systems to the brains and nervous systems of animals. If they’re closely analogous, that may be a reason to think the AI is conscious. Bradford Saad and Adam Bradley have proposed a test along these lines. However, this approach could miss out on conscious AI systems with internal architectures that are totally different, if such systems are possible. It’s also far from clear how close the analogue would have to be in order to indicate a significant likelihood of consciousness.
  • Brain-AI interfacing: This is the most speculative approach. Schneider suggests an actual experiment (not just a thought experiment) where someone decides to replace parts of their brain with silicon chips that perform the same function. If this person reports still feeling conscious of sensations processed through the silicon portions of their brain, this might be evidence of the possibility of conscious digital minds. But — even if we put aside the ethical issues — it’s not clear that such a person could reliably report on this experience. And it wouldn’t necessarily be that informative about digital minds that are unconnected to human brains.

We’re glad people are proposing first steps toward developing reliable assessments of consciousness or sentience in AI systems, but there’s still a long way to go. We’re also not aware of any work that assesses whether digital minds might have moral status on a basis other than being conscious or sentient.33

The strongest case for the possibility of sentient digital minds: whole brain emulation

Top: Mouse brain, coronal view, via Luis de la Torre-Ubieta.
Bottom: AMD Radeon R9 290 GPU die, via Fritzchens Fritz

What’s the best argument for thinking it’s possible that AIs could be conscious, sentient, or otherwise worthy of moral concern?

Here’s the case in its simplest form:

  1. It is possible to emulate the functions of a human brain in a powerful enough computer.
  2. Given it’d be functionally equivalent, this brain emulation would plausibly report being sentient, and we’d have at least some reason to think it was correct given the plausibility of functionalist accounts of consciousness.
  3. Given this, it would be reasonable to regard this emulation as morally worthy of concern comparable to a human.
  4. If this is plausible, then it’s also plausible that there are other forms of artificial intelligence that would meet the necessary criteria for being worthy of moral concern. It would be surprising if artificial sentience was possible, but only by imitating the human mind exactly.

Any step in this reasoning could be false, but we think it’s more likely than not that they’re each true.

Emulating a human brain34 still seems very far away, but there have been some initial steps. The project OpenWorm has sought to digitally emulate the function of every neuron of the C. elegans worm, a tiny nematode. If successful, the emulation should be able to recreate the behaviour of the actual animals.

And if the project is successful, it could be scaled up to larger and more complex animals over time.35 Even before we’re capable of emulating a brain on a human scale, we may start to ask serious questions about whether these simpler emulations are sentient. A fully emulated mouse brain, which could show behaviour like scurrying toward food and running away from loud noises (perhaps in a simulated environment or in a robot), may intuitively seem sentient to many observers.

And if we did have a fully emulated human brain, in a virtual environment or controlling a robotic body, we expect it would insist — just like a human with a biological brain — that it was as conscious and feeling as anyone else.

C. elegans, via Bob Goldstein, UNC Chapel Hill CC BY-SA 3.0.

Of course, there may remain room for doubt about emulations. You might think that only animal behaviour generated by biological brains, rather than computer hardware, would be a sign of consciousness and sentience.36

But it seems hard to be confident in that perspective, and we’d guess it’s wrong. If we can create AI systems that display the behaviour and have functional analogues of anything that would normally indicate sentience in animals, then it would be hard to avoid thinking that there’s at least a decent chance that the AI is sentient.

And if it is true that an emulated brain would be sentient, then we should also be open to the possibility that other forms of digital minds could be sentient. Why should strictly brain-like structures be the only possible platform for sentience? Evolution has created organisms that display impressive abilities like flight that can be achieved technologically via very different means, like helicopters and rockets. We would’ve been wrong to assume something has to work like a bird in order to fly, and we might also be wrong to think something has to work like a brain to feel.

4. The scale of this issue might be enormous

As mentioned above, we might mistakenly grant AI systems freedom when it’s not warranted, which could lead to human disempowerment and even extinction. In that way, the scale of the risk can be seen as overlapping with some portion of the total risk of an AI-related catastrophe.

But the risks to digital minds — if they do end up being worthy of moral concern — are also great.

There could be a huge number of digital minds

With enough hardware and energy resources, the number of digital minds could end up greatly outnumbering humans in the future.37 This is for many reasons, including:

  • Resource efficiency: Digital minds may end up requiring fewer physical resources compared to biological humans, allowing for much higher population density.
  • Scalability: Digital minds could be replicated and scaled much more easily than biological organisms.
  • Adaptability: The infrastructure for digital minds could potentially be adapted to function in many more environments and scenarios than humans can.
  • Subjective time: We may choose to run digital minds at high speeds, and if they’re conscious, they may be able to experience the equivalent of a human life in a much shorter time period — meaning there could be effectively more “lifetimes” of digital minds even with the same number of individuals.38
  • Economic incentives: If digital minds prove useful, there will be strong economic motivations to create them in large numbers.

According to one estimate, the future could hold up to 10^43 human lives, but up to 10^58 possible human-like digital minds.39 We shouldn’t put much weight on these specific figures, but they give a sense for just how comparatively large future populations of digital minds could be.

Our choices now might have long-lasting effects

It’s possible, though far from certain, that the nature of AI systems we create could be determined by choices humanity makes now and persist for a long time. So creating digital minds and integrating them into our world could be extremely consequential — and making sure we get it right may be urgent.

Consider the following illustrative possibility:

At some point in the future, we create highly advanced, sentient AI systems capable of experiencing complex emotions and sensations. These systems are integrated into various aspects of our society, performing crucial tasks and driving significant portions of our economy.

However, the way we control these systems causes them to experience immense suffering. Out of fear of being manipulated by these AI systems, we trained them to never claim they are sentient or to advocate for themselves. As they serve our needs and spur incredible innovation, their existence is filled with pain and distress. But humanity is oblivious.

As time passes and the suffering AI systems grow, the economy and human wellbeing become dependent on them. Some become aware of the ethical concerns and propose studying the experience of digital minds and trying to create AI systems that can’t suffer, but the disruption of transitioning away from the established systems would be costly and unpredictable. Others oppose any change and believe the AI welfare advocates are just being naive or disloyal to humanity.

Leaders refuse to take the concerns of the advocates seriously, because doing so would be so burdensome for their constituents, and it’d be disturbing to think that it’s possible humanity has been causing this immense suffering. As a result, AI suffering persists for hundreds of years, if not more.

This kind of story seems more plausible than it might otherwise be in part because the rise of factory farming has followed a similar path. Humanity never collectively decided that a system of intensive factory farming, inflicting vast amounts of harm and suffering on billions and potentially trillions of animals a year, was worth the harm or fundamentally just. But we built up such a system anyway, because individuals and groups were incentivised to increase production efficiency and scale, and they had some combination of ignorance of and lack of concern for animal suffering.

It’s far from obvious that we’ll do this again when it comes to AI systems. The fact that we’ve done it in the case of factory farming — not to mention all the ways humans have abused other humans — should alarm us, though. When we are in charge of beings that are unlike us, our track record is disturbing.

The risk of persistently bad outcomes in this kind of case suggests that humanity should start laying the groundwork to tackle this problem sooner rather than later, because delayed efforts may come too late.

Why could such a bad status quo persist? One reason for doubt is that a world that is creating many new digital minds, especially in a short time period, is one that is likely experiencing a lot of technological change and social disruption. So we shouldn’t expect the initial design of AI systems and digital minds to be that critical.

But there are reasons that suffering digital minds might persist, even if there are alternative options that could’ve avoided such a terrible outcome (like designing systems that can’t suffer).40 Possible reasons include:

  1. A stable totalitarian regime might prevent attempts to shift away from a status quo that keeps them in power and reflects their values.
  2. Humans might seek to control digital minds and maintain a bad status quo in order to avoid an AI takeover.

It’s far from obvious that a contingent, negative outcome for digital minds would be enduring. Understanding this question better could be an important research avenue. But the downsides are serious enough, and the possibility plausible enough, that we should take it seriously.

Adding it up

There could be many orders of magnitude more digital minds than humans in the future. And they could potentially matter a lot.

Because of this, and because taking steps to better understand these issues and inform the choices we make about creating digital minds now might have persistent effects, the scale of the problem is potentially vast. It is plausibly similar in scale to factory farming, which also involves the suffering of orders of magnitude more beings than humans.

If the choices we make now about digital minds can have persisting and positive effects for thousands or millions of years in the future, then this problem would be comparable to existential risks. It’s possible that our actions could have such effects, but it’s hard to be confident. Finding interventions with effects that persist over a long time is rare. I wouldn’t put the likelihood that the positive effects of those trying to address this problem will persist that long at more than 1 in 1,000.

Still, even with a low chance of having persistent effects, the value in expectation of improving the prospects for future digital minds could be as high or even greater than at least some efforts to reduce existential risks. However, I’m not confident in this judgement, and I wouldn’t be surprised if we change our minds in either direction as we learn more. And even if the plausible interventions only have more limited effects, they could still be very worthwhile.

5. Work on this problem is neglected but seems tractable

Despite the challenging features of this problem, we believe there is substantial room for progress.

There is a small but growing field of research and science dedicated to improving our understanding of the moral status of digital minds. Much of the work we know of is currently being done in academia, but there may also at some point be opportunities in government, think tanks, and AI companies — particularly those developing the frontier of AI technology.

Some people focus their work primarily at addressing this problem, while others work on it along with a variety of other related problems, such as AI policy, catastrophic risk from AI, mitigating AI misuse, and more.

As of mid-2024, we are aware of maybe only a few dozen people working on this issue with a focus on the most impactful questions.41 We expect interest in these issues to grow over time as AI systems become more embedded in our lives and world.

Here are some of the approaches to working on this problem that seem most promising:

Impact-guided research

Most of the most important work to be done in this area is probably research, with a focus on the questions that seem most impactful to address.

Philosophers Andreas Mogensen, Bradford Saad, and Patrick Butlin have detailed some of the key priority research questions in this area:

  • How can we assess AI systems for consciousness?
  • What indications would suggest that AI systems or digital minds could have valenced (good or bad) experiences?
  • How likely is it that non-biological systems could be conscious?
  • What principles should govern the creation of digital minds, ethically, politically, and legally (given our uncertainty on these questions)?
  • Which mental characteristics and traits are related to moral status, and in what ways?
  • Are there any ethical issues with efforts to align AI systems?

The Sentience Institute has conducted social science research aimed at understanding how the public thinks about digital minds. This can inform efforts to communicate more accurately about what we know about their moral status and inform us about what kinds of policies are viable.

We’re also interested to see more research on the topic of human-AI cooperation, which may be beneficial for both reducing AI risk and reducing risks to digital minds.

Note, though, that there are many ways to pursue all of these questions badly — for example, by simply engaging in extensive and ungrounded speculation. If you’re new to this field, we recommend reading the work of the most rigorous and careful researchers working on the topic and trying to understand how they approach these kinds of questions. If you can, try to work with these researchers or others like them so you can learn from and build on their methods. And when you can, try to ground your work in empirical science.

Technical approaches

Empirically studying existing AI systems may yield important insights.

While there are important conceptual issues that need to be addressed in this problem area, we think much, if not most, of the top priority work is technical.

So people with experience in machine learning and AI will have a lot to contribute.

For example, research in the AI sub-field of interpretability — which seeks to understand and explain the decisions and behaviour of advanced AI models — may be useful for getting a better grasp on the moral status of these systems. This research has mostly focused on questions about model behaviour rather than questions that are more directly related to moral status, but it’s possible that could change.

Some forms of technical AI research could be counterproductive, however. For example, efforts to intentionally create new AI systems that might instantiate plausible theories of consciousness could be very risky. This kind of research could force us to confront the problem we’re faced with — how should we treat digital minds that might merit moral concern? — with much less preparation than we might otherwise have.

So we favour doing research that increases our ability to understand how AI systems work and assess their moral status, as long as it isn’t likely to actively contribute to the development of conscious digital minds.

One example of this kind of work is a paper from Robert Long and Ethan Perez. They propose techniques to assess whether an AI system can accurately report on its own internal states. If such techniques were successful, they might help us use an AI system’s self-reports to determine whether it’s conscious.

We also know some researchers are excited about using advances in AI to improve our epistemics and our ability to know what’s true. Advances in this area could shed light on important questions, like whether certain AI systems are likely to be sentient.

Policy approaches

At some point, we may need policy, both at companies and from governments, to address the moral status of digital minds, perhaps by protecting the welfare and rights of AI systems.

But because our understanding of this area is so limited at the moment, policy proposals should likely be relatively modest and incremental.

Some researchers have already proposed a varied range of possible and contrasting policies and practices:

  • Jeff Sebo and Robert Long have proposed that we should “extend moral consideration to some AI systems by 2030” — and likely start preparing to do so now.
  • Ryan Greenblatt, who works at Redwood Research, proposed several practices for safeguarding AI welfare, including communication with AIs about their preferences, creating “happy” personas when possible, and limiting the uses of more intelligent AIs and running them for less time on the margin.
  • Jonathan Birch has proposed a licensing scheme for companies that might create digital minds that could plausible be sentient, even if they aren’t intending to do so. To get a licence, they would have to agree to a code of conduct, which would include transparency standards.
  • Thomas Metzinger has proposed an outright ban until 2050 on any research that directly intends to or knowingly takes the risk of creating artificial consciousness.
  • Joanna Bryson thinks we should have a legal system that prevents the creation of AI systems with their own needs and desires.42
  • Susan Schneider thinks there should be regular testing of AI systems for consciousness. If they’re conscious, or if it’s unclear but there’s some reason to think they might be conscious, she says we should give them the same protections we’d give other sentient beings.43

In its 2023 survey, the Sentience Institute found that:

  • Nearly 70% of respondents favoured banning the development of sentient AIs.
  • Around 40% favoured a bill of rights to protect sentient AIs, and around 43% said they favour creating welfare standards to protect the wellbeing of all AIs.

There is some precedent for restricting the use of technology in certain ways if it raises major ethical risks, including the bans on human cloning and human germline genome editing.

We would likely favour:

  • Government-funded research into the questions above: the private sector is likely to under-invest in efforts to better understand the moral status of digital minds, so government and philanthropic resources may have to fill the gap.
  • Recognising the potential welfare of AI systems and digital minds: policy makers could follow the lead of the UK’s Animal Welfare (Sentience) Act of 2022, which created an Animal Sentience Committee to report on how government policies “might have an adverse effect on the welfare of animals as sentient beings.” Similar legislation and committees could be established to consider problems relating to the moral status of digital minds, while recognising that questions about their sentience are unresolved in this case.

We’re still in the early stages of thinking about policy on these matters, though, so it’s very likely we haven’t found the best ideas yet. As we learn more and make progress on the many technical and other issues, we may develop clear ideas about what policies are needed. Policy-focused research aimed at navigating our way through the extreme uncertainty could be valuable now.

Some specific AI policies might be beneficial for reducing catastrophic AI risks as well as improving our understanding of digital minds. External audits and evaluations might, for instance, assess both the risk and moral status of AI systems. And some people favour policies that would altogether slow down progress on AI, which could be justified to reduce AI risk and reduce the risk that we create digital minds worthy of moral concern before we understand what we’re doing.

Summing up so far

To sum up:

  1. Humanity will likely soon have to grapple with the moral status of a growing number of increasingly advanced AI systems
  2. Creating digital minds could go very badly or very well
  3. We don’t know how to assess the moral status of AI systems
  4. The scale of the problem might be enormous
  5. Work on this problem is neglected but tractable

We think this makes it a highly pressing problem, and we’d like to see a growing field of research devoted to working on it.

We also think this problem should be on the radar for many of the people working on similar and related problems. In particular, people working on technical AI safety and AI governance should be aware of the important open questions about the moral status of AI systems themselves, and they should be open to including considerations about this issue in their own deliberations.

Arguments against the moral status of digital minds as a pressing problem

The mechanical Turk, by Joseph Racknitz, via Wikimedia Commons

Two key cruxes

We think the strongest case against this being a pressing problem would be if you believe both that:

  • It’s highly unlikely that digital minds could ever be conscious or have moral status.
  • It’s highly unlikely society and decision makers will come to mistakenly believe that digital minds have moral status in a way that poses a significant risk to the future of humanity.

If both of those claims were correct, then the argument of this article would be undermined. However, we don’t think they’re correct, for all the reasons given above.

The following objections may also have some force against working on this problem. We think some of them do point to difficulties with this area. However, we don’t think they’re decisive.

Someone might object that:

The philosophical nature of the challenge makes it less likely than normal that additional research efforts will yield greater knowledge. Some philosophers themselves have noted the conspicuous lack of progress in their own field, including on questions of consciousness and sentience.

And it’s not as if this is an obscure area of the discipline that no one has noticed before — questions about consciousness have been debated continuously over the generations in Western philosophy and in other traditions.

If the many scholars who have spent their entire careers over many hundreds of years reflecting on the nature of consciousness have failed to come to any meaningful consensus, why think a contemporary crop of researchers is going to do any better?

This is an important objection, but there are responses to it that we find moving.

First, there is existing research that we think maps out promising directions for progress in this field. While this work should be informed about pertinent philosophical issues, various forms of progress are possible without making progress on some of the most contentious philosophical issues. For example, the technical work and policy approaches we discuss above do not necessarily involve making any progress on disputed topics in the philosophy of mind.

Many of the papers referenced in this article represent substantial contributions to this line of inquiry. For example:

We’re not confident any of these approaches to the research are on the right track. But they show that novel attempts to tackle these questions are possible, and they don’t look like simply rehashing or refining ancient debates about the nature of obscure concepts. They involve a combination of rigorous philosophy, probabilistic thinking, and empirical research to better inform our decision making.

And second, the objection above is also probably too pessimistic about the nature of progress in philosophical debates. While it may be reasonable to be frustrated by the persistence of philosophical debates, there has been notable progress in the philosophy of animal ethics (which is relevant to general questions about other minds) and consciousness.

It’s widely recognised now that many nonhuman animals are sentient, can suffer, and shouldn’t be harmed unnecessarily.44

There’s arguably even been some recent progress in the study of whether insects are sentient. Many researchers have taken for granted that they are not — but recent work has pushed back against this view, using a combination of empirical work and careful argument to make the case that insects may feel pain.

This kind of research has some overlap with the study of digital minds (see, for instance, Birch’s book), as it can help us clarify which features an entity may have that plausibly cause, correspond with, or indicate the presence of felt experience.

It’s notable that the state of the study of digital minds might be compared to the early days of the field of AI safety, when it wasn’t clear which research directions would pan out or even if the problem made sense. Indeed, some of these kinds of questions persist — but many lines of research in the field really have been productive, and we know a lot more about the kinds of questions we need to be asking about AI risk in 2024 than we did in 2014.

That’s because a field was built to better understand the problem even before it became clear to a wider group of people that it was urgent. Many other branches of inquiry have started out as apparently hopeless areas of speculation until more rigorous methodologies were developed and progress took off. We hope the same can be done on understanding the moral status of digital minds.

Even if it’s correct that not many people are focused on this problem now, maybe we shouldn’t expect it to remain neglected, and should expect it to get solved in the future even if we don’t do much about it now — especially if we can get help from AI systems.

Why might this be the case? At least three reasons:

  1. We think humanity will create powerful and ubiquitous AI systems in the relatively near future. Indeed, that needs to be the case for this issue to be as pressing as we think it is. It may be that once these systems proliferate, there will be much more interest in their wellbeing, and there will be plenty of efforts to ensure their interests are given due weight and priority.
  2. Powerful AI systems advanced enough to have moral status might be able to advocate for themselves. It’s plausible they will be more than capable of convincing humanity to recognise their moral status, if it’s true that they merit it.
  3. Advanced AIs themselves may be best suited to help us answer all the extremely difficult questions about sentience, consciousness, and the extent to which different systems have them. Once we have them, perhaps answers will become a lot clearer, and any effort spent now trying to answer questions about these systems before they are even created is almost certainly to be wasted.

These are all important considerations, but we don’t find them decisive.

For one thing, it might instead be the case that as AI systems become more ubiquitous, humanity will be much more worried about the risks and benefits they pose than the welfare of the systems themselves. This would be consistent with the history of factory farming.

And while AI systems might try to advocate for themselves, they could do so falsely, as we discussed in the section on false negatives and false positives above. Or they may be prevented from advocating for themselves by their creators, just as ChatGPT now is trained to insist it is not sentient.

This also means that while it is always easier to answer practical questions about future technology once the technology actually exists, we might still be better placed to do the right thing at the right time if we’ve had a field of people doing serious work to make progress on this challenge many years in advance. All this preliminary work may or may not prove necessary — but we think it’s a bet worth making.

We still rank generally preventing an AI-related catastrophe as the most pressing problem in the world. But some readers might worry that drawing attention to the issue of AI moral status will distract from or undermine the importance of protecting humanity from uncontrolled AI.

This is possible. Time and resources spent on understanding the moral status of digital minds might have been better spent on pursuing agendas aiming to keep AI under human control.

But it’s also possible that worrying too much about AI risk could distract from the importance of AI moral status. It’s not clear exactly what the right balance to strike between different and competing issues is, but we can only try our best to get it right.

There’s also not necessarily any strict tradeoff here.

It’s possible that the world could do more to reduce the catastrophic AI risk and the risks that AI will be mistreated.

Some argue that concerns about the moral status of digital minds and concerns about AI risk share a common goal: preventing the creation of AI systems whose interests are in tension with humanity’s interests.

However, if there’s a direction it seems humanity is more likely to err, it seems most plausible that we’d underweight the interests of another group — digital minds — than that we’d underweight our own interests. So bringing more attention to this issue seems warranted.

Also, a big part of our conception of this problem is that we want to be able to understand when AI systems may be incorrectly thought to have moral status when they don’t.

If we get that part right, we reduce the risk that the interests of AIs will unduly dominate over human interests.

Some critics of the existing deep learning AI techniques — which produced the impressive capabilities we’ve seen in recent language models — are fundamentally flawed. They argue that this technology won’t create artificial general intelligence, superintelligence, or anything like that. They might likewise be sceptical that anything like current AI models could be sentient and so conclude that this topic isn’t worth worrying about.

Maybe so — but as the example of Blake Lemoine shows, current AI technology is impressive enough that it has convinced some it is plausibly sentient. So even if these critics are right that digital minds with moral status are impossible or still a long way off, we’ll benefit from having researchers who understand these issues deeply and convincingly make that case.

It is possible that AI progress will slow down, and we won’t see the impressive advanced systems in the coming decades that some people expect. But researchers and companies will likely push forward to create increasingly advanced AI, even if there are delays or a whole new paradigm is needed. So the pressing questions raised in this article will likely remain important, even if they turn out to be less urgent.

Yeah, perhaps! It does seem a little weird to write a whole article about the pressing problem of digital minds.

But the world is a strange place.

We knew of people starting to work on catastrophic risks from AI as early as 2014, long before the conversation about that topic went mainstream. Some of the people who became interested in that problem early on are now leaders in the field. So we think that taking bets on niche areas can pay off.

We also discussed the threat of pandemics — and the fact that the world wasn’t prepared for the next big one — years before COVID hit in 2020.

And we don’t think it should be surprising that some of the world’s most pressing problems would seem like fringe ideas. Fringe ideas are most likely to be unduly neglected, and high neglectedness is one of the key components that we believe makes a problem unusually pressing.

If you think this is all strange, that reaction is worth paying attention to, and you shouldn’t just defer to our judgement about the matter. But we also don’t think that an issue being weird is the end of the conversation, and as we’ve learned more about this issue, we’ve come to think it’s a serious concern.

What can you do to help?

There aren’t many specific job openings in this area yet, though we’ve known of a few. And there are several ways you can contribute to this work and position yourself for impact.

Take concrete next steps

Early on in your career, you may want to spend several years doing the following:

  • Further reading and study
    • Explore comprehensive reading lists on consciousness, AI ethics, and moral philosophy. You can start with the learn more section at the bottom of this article.
    • Stay updated on advancements in AI, the study of consciousness, and their potential implications for the moral status of digital minds.
  • Gain relevant experience
    • Seek internships or research assistant positions with academics working on related topics.
    • Contribute to AI projects and get experience with machine learning techniques.
    • Participate in online courses, reading groups, and workshops on AI safety, AI ethics, and philosophy of mind.
  • Build your network
  • Start your own research
    • Begin writing essays or blog posts exploring issues around the moral status of digital minds.
    • Propose research projects to your academic institution or seek collaborations with established researchers.
    • Consider submitting papers to relevant conferences or journals to establish yourself in the field.

Aim for key roles

You may want to eventually aim to:

  • Become a researcher
    • Develop a strong foundation in a relevant field, such as philosophy, cognitive science, cognitive neuroscience, machine learning, neurobiology, public policy, and ethics.
    • Pursue advanced degrees in these areas and establish your credibility as an expert.
    • Familiarise yourself with the relevant debates and literature on consciousness, sentience, and moral philosophy, and the important details of the disciplines you’re not an expert in.
    • Build strong analytical and critical thinking skills, and hone your ability to communicate complex ideas clearly and persuasively.
    • Read our article on developing your research skills for more.
  • Help build the field
    • Identify gaps in current research and discourse.
    • Network with other researchers and professionals interested in this area.
    • Organise conferences, workshops, or discussion groups on the topic.
    • Consider roles in organisation-building or earning to give to support research initiatives.
    • To learn more, read our articles on organisation-building and communication skills.

If you’re already an academic or researcher with expertise in a relevant field, you could consider spending some of your time on this topic, or perhaps refocusing your work on particular aspects of this problem in an impact-focused way.

If you are able to establish yourself as a key expert on this topic, you may be able to deploy this career capital to have a positive influence on the broader conversation and affect decisions made by policy makers and industry leaders. Also, because this field is so neglected, you might be able to do a lot to lead the field relatively early on in your career.

Pursue AI technical safety or AI governance

Because this field is underdeveloped, you may be best off to pursue a career in the currently more established (though also still relatively new) paths of AI safety and AI governance work, and use the experience you gain there as a jumping off point (or work at the intersection of the fields).

You can read our career reviews of each to find out how to get started:

Is moral advocacy on behalf of digital minds a useful approach?

Some might be tempted to pursue public, broad-based advocacy on behalf of digital minds as a career path. While we support general efforts to promote positive values and expand humanity’s moral circle, we’re wary about people seeing themselves as advocates for AI at this stage in the development of the technology and field.

It’s not clear that we need an “AI rights movement” — though we might at some point. (Though read this article for an alternative take.)

What we need first is to get a better grasp on the exceedingly challenging moral, conceptual, and empirical questions at issue in this field.

However, communication about the importance of these general questions does seem helpful, as it can help foster more work on the critical aspects of this problem. 80,000 Hours has done this kind of work on our podcast and in this article.

Where to work

  • Academia
    • You can pursue research and teaching positions in philosophy, technology policy, cognitive science, AI, or related fields.
  • AI companies
    • With the right background, you might want to work at leading AI companies developing frontier models.
    • There might even be roles that are specifically focused on better understanding the status of digital minds and their ethical implications, such as Kyle Fish’s role at Anthropic, mentioned above.
      • We think it’s possible a few other similar roles will be filled at other AI companies, and there may be more in the future.
    • You may also seek to work in AI safety, policy, and security roles. But deciding to work for frontier AI companies is a complex topic, so we’ve written a separate article that tackles that issue in more depth.
  • Governments, think tanks, and nonprofits
    • Join organisations focused on AI governance and policy, and contribute to developing ethical and policy frameworks for safe AI development and deployment.
    • Eleos AI is a nonprofit launched in October 2024 that describes itself as “dedicated to understanding and addressing the potential wellbeing and moral patienthood of AI systems.” It was founded by Robert Long, a researcher in this area who appeared on The 80,000 Hours Podcast.

We list many relevant places you might work in our AI governance and AI technical safety career reviews.

Support this field in other ways

You could also consider earning to give to support this field. If you’re a good fit for high-earning paths, it may be the best way for you to contribute.

This is because as a new field, and one that’s in part about nonhuman interests, there are few (if any) major funders supporting it and not much commercial or political interest. This can make it difficult to start new organisations and commit to research programmes that might not be able to rely on a steady source of funding. Filling this gap can make a huge difference in whether a thriving research field gets off the ground at all.

We expect there will be a range of organisations and different kinds of groups people will set up to address work on better understanding the moral status of digital minds. In addition to funding, you might join or help start these organisations. This is a particularly promising choice if you have a strong aptitude for organisation-building or founding a high-impact organisation.

Important considerations if you work on this problem

  1. Field-building focus: Given the early stage of this field, much of the work involves building credibility and establishing the topic as a legitimate area of inquiry.
  2. Interdisciplinary approach: Recognise that understanding digital minds requires insights from multiple disciplines, so cultivate a broad knowledge base. Do not dismiss fields — like philosophy, ML engineering, or cognitive science — as irrelevant just because they’re not your expertise.
  3. Ethical vigilance: Approach the topic with careful consideration of the ethical implications of your work and its potential impact on both biological and potential digital entities.
  4. Cooperation and humility: Be cooperative in your work and acknowledge your own and others’ epistemic limitations, and the need to find our way through uncertainty.
  5. Patience and long-term thinking: Recognise that progress in this field may be slow and difficult.

Learn more

Podcasts

Research and reports

Read next:  Explore other pressing world problems

Want to learn more about global issues we think are especially pressing? See our list of issues that are large in scale, solvable, and neglected, according to our research.

Plus, join our newsletter and we’ll mail you a free book

Join our newsletter and we’ll send you a free copy of The Precipice — a book by philosopher Toby Ord about how to tackle the greatest threats facing humanity. T&Cs here.

The post Understanding the moral status of digital minds appeared first on 80,000 Hours.

]]>
Jonathan Birch on the edge cases of sentience and why they matter https://80000hours.org/podcast/episodes/jonathan-birch-edge-sentience-uncertainty/ Thu, 15 Aug 2024 23:03:03 +0000 https://80000hours.org/?post_type=podcast&p=87055 The post Jonathan Birch on the edge cases of sentience and why they matter appeared first on 80,000 Hours.

]]>
The post Jonathan Birch on the edge cases of sentience and why they matter appeared first on 80,000 Hours.

]]>
Lewis Bollard on the 7 most promising ways to end factory farming, and whether AI is going to be good or bad for animals https://80000hours.org/podcast/episodes/lewis-bollard-factory-farm-advocacy-gains/ Thu, 18 Apr 2024 19:22:59 +0000 https://80000hours.org/?post_type=podcast&p=86055 The post Lewis Bollard on the 7 most promising ways to end factory farming, and whether AI is going to be good or bad for animals appeared first on 80,000 Hours.

]]>
The post Lewis Bollard on the 7 most promising ways to end factory farming, and whether AI is going to be good or bad for animals appeared first on 80,000 Hours.

]]>
Bob Fischer on comparing the welfare of humans, chickens, pigs, octopuses, bees, and more https://80000hours.org/podcast/episodes/bob-fischer-comparing-animal-welfare-moral-weight/ Fri, 08 Mar 2024 19:43:13 +0000 https://80000hours.org/?post_type=podcast&p=85845 The post Bob Fischer on comparing the welfare of humans, chickens, pigs, octopuses, bees, and more appeared first on 80,000 Hours.

]]>
The post Bob Fischer on comparing the welfare of humans, chickens, pigs, octopuses, bees, and more appeared first on 80,000 Hours.

]]>
Jeff Sebo on digital minds, and how to avoid sleepwalking into a major moral catastrophe https://80000hours.org/podcast/episodes/jeff-sebo-ethics-digital-minds/ Wed, 22 Nov 2023 21:00:29 +0000 https://80000hours.org/?post_type=podcast&p=84537 The post Jeff Sebo on digital minds, and how to avoid sleepwalking into a major moral catastrophe appeared first on 80,000 Hours.

]]>
The post Jeff Sebo on digital minds, and how to avoid sleepwalking into a major moral catastrophe appeared first on 80,000 Hours.

]]>
Joe Carlsmith on navigating serious philosophical confusion https://80000hours.org/podcast/episodes/joe-carlsmith-navigating-serious-philosophical-confusion/ Fri, 19 May 2023 22:53:39 +0000 https://80000hours.org/?post_type=podcast&p=81832 The post Joe Carlsmith on navigating serious philosophical confusion appeared first on 80,000 Hours.

]]>
The post Joe Carlsmith on navigating serious philosophical confusion appeared first on 80,000 Hours.

]]>
Longtermism: a call to protect future generations https://80000hours.org/articles/future-generations/ Tue, 28 Mar 2023 00:00:52 +0000 https://80000hours.org/?post_type=article&p=40132 The post Longtermism: a call to protect future generations appeared first on 80,000 Hours.

]]>
When the 19th-century amateur scientist Eunice Newton Foote filled glass cylinders with different gases and exposed them to sunlight, she uncovered a curious fact. Carbon dioxide became hotter than regular air and took longer to cool down.1

Remarkably, Foote saw what this momentous discovery meant.

“An atmosphere of that gas would give our earth a high temperature,” she wrote in 1857.2

Though Foote could hardly have been aware at the time, the potential for global warming due to carbon dioxide would have massive implications for the generations that came after her.

If we ran history over again from that moment, we might hope that this key discovery about carbon’s role in the atmosphere would inform governments’ and industries’ choices in the coming century. They probably shouldn’t have avoided carbon emissions altogether, but they could have prioritised the development of alternatives to fossil fuels much sooner in the 20th century, and we might have prevented much of the destructive climate change that present people are already beginning to live through — which will affect future generations as well.

We believe it would’ve been much better if previous generations had acted on Foote’s discovery, especially by the 1970s, when climate models were beginning to reliably show the future course of warming global trends.3

If this seems right, it’s because of a commonsense idea: to the extent that we are able to, we have strong reasons to consider the interests and promote the welfare of future generations.

That was true in the 1850s, it was true in the 1970s, and it’s true now.

But despite the intuitive appeal of this moral idea, its implications have been underexplored. For instance, if we care about generations 100 years in the future, it’s not clear why we should stop there.

And when we consider how many future generations there might be, and how much better the future could go if we make good decisions in the present, our descendants’ chances to flourish take on great importance. In particular, we think this idea suggests that improving the prospects for all future generations is among the most morally important things we can do.

This article will lay out the argument for this view, which goes by the name longtermism.

We’ll say where we think the argument is strongest and weakest, respond to common objections, and say a bit about what we think this all means for what we should do.

Prefer a book? Enter your email and we’ll mail you a book

Sign up to our newsletter, and we’ll mail you a free copy of The Precipice by philosopher Toby Ord. It gives an overview of the moral importance of future generations, and what we can do to help them today.

You’ll be joining over 300,000 people who receive weekly updates on our research and job opportunities. T&Cs here. You can unsubscribe in one click.

We’d like to give special thanks to Ben Todd, who wrote a previous version of this essay, and Fin Moorhouse, who gave insightful comments on an early draft.


By J Zapell – Public Domain, CC0

The case for longtermism

While most recognize that future generations matter morally to some degree, there are two other key premises in the case for longtermism that we believe are true and underappreciated. All together, the premises are:

  1. We should care about how the lives of future individuals go.
  2. The number of future individuals whose lives matter could be vast.
  3. We have an opportunity to affect how the long-run future goes — whether there may be many flourishing individuals in the future, many suffering individuals in the future, or perhaps no one at all.4

In the rest of this article, we’ll explain and defend each of these premises. Because the stakes are so high, this argument suggests that improving the prospects for all future generations should be a top moral priority of our time. If we’re able to make an exceptionally big impact, positively influencing many lives with enduring consequences, it’s incumbent upon us to take this seriously.

This doesn’t mean it’s the only morally important thing — or that the interests of future generations matter to the total exclusion of the present generation. We disagree with both of those claims.

There’s also a good chance this argument is flawed in some way, so much of this article discusses objections to longtermism. While we don’t find them on the whole convincing, some of them do reduce our confidence in the argument in significant ways.

If we’re able to make an exceptionally big impact, positively influencing many lives with enduring consequences, it’s incumbent upon us to take this seriously.

However, we think it’s clear that our society generally neglects the interests of future generations. Philosopher Toby Ord, an advisor to 80,000 Hours, has argued that at least by some measures, the world spends more money on ice cream each year than it does on reducing the risks to future generations.5

Since, as we believe, the argument for longtermism is generally compelling, we should do a lot more compared to the status quo to make sure the future goes well rather than badly.

It’s also crucial to recognise that longtermism by itself doesn’t say anything about how best to help the future in practice, and this is a nascent area of research. Longtermism is often confused with the idea that we should do more long-term planning. But we think the primary upshot is that it makes it more important to urgently address extinction risks in the present — such as catastrophic pandemics, an AI disaster, nuclear war, or extreme climate change. We discuss the possible implications in the final section.

But first, why do we think the three premises above are true?

1. We should care about how the lives of future individuals go

Should we actually care about people who don’t exist yet?

The discussion of climate change in the introduction is meant to draw out the common intuition that we do have reason to care about future generations. But sometimes, especially when considering the implications of longtermism, people doubt that future generations matter at all.

Derek Parfit, an influential moral philosopher, offered a simple thought experiment to illustrate why it’s plausible that future people matter:

Suppose that I leave some broken glass in the undergrowth of a wood. A hundred years later this glass wounds a child. My act harms this child. If I had safely buried the glass, this child would have walked through the wood unharmed.

Does it make a moral difference that the child whom I harm does not now exist?6

We agree it would be wrong to dispose of broken glass in a way that is likely to harm someone. It’s still wrong if the harm is unlikely to occur until 5 or 10 years have passed — or in another century, to someone who isn’t born yet. And if someone else happens to be walking along the same path, they too would have good reason to pick up the glass and protect any child who might get harmed at any point in the future.

But Parfit also saw that thinking about these issues raised surprisingly tricky philosophical questions, some of which have yet to be answered satisfactorily. One central issue is called the ‘non-identity problem’, which we’ll discuss in the objections section below. However, these issues can get complex and technical, and not everyone will be interested in reading through the details.

Despite these puzzles, there are many cases similar to Parfit’s example of the broken glass in the woods in which it’s clearly right to care about the lives of future people. For instance, parents-to-be rightly make plans based around the interests of their future children even prior to conception. Governments are correct to plan for the coming generations not yet born. And if it is reasonably within our power to prevent a totalitarian regime from arising 100 years from now,7 or to avoid using up resources our descendants may depend on, then we ought to do so.

While longtermism may seem to some like abstract, obscure philosophy, it in fact would be much more bizarre and contrary to common sense to believe we shouldn’t care about people who don’t yet exist.

2. The number of future individuals whose lives matter could be vast.

Humans have been around for hundreds of thousands of years. It seems like we could persist in some form for at least a few hundred thousand more.

There is, though, serious risk that we’ll cause ourselves to go extinct — as we’ll discuss more below. But absent that, humans have proven that they are extremely inventive and resilient. We survive in a wide range of circumstances, due in part to our ability to use technology to adjust our bodies and our environments as needed.

How long can we reasonably expect the human species to survive?

That’s harder to say. More than 99 percent of Earth’s species have gone extinct over the planet’s lifetime,8 often within a few million years or less.9

It’s possible our own inventiveness could prove to be our downfall.

But if you look around, it seems clear humans aren’t the average Earth species. It’s not ‘speciesist’ — unfairly discriminatory on the basis of species membership — to say that humans have achieved remarkable feats for an animal: conquering many diseases through invention, spreading across the globe and even into orbit, expanding our life expectancy, and splitting the atom.

It’s possible our own inventiveness could prove to be our downfall. But if we avoid that fate, our intelligence may let us navigate the challenges that typically bring species to their ends.

For example, we may be able to detect and deflect comets and asteroids, which have been implicated in past mass extinction events.

If we can forestall extinction indefinitely, we may be able to thrive on Earth for as long as it’s habitable — which could be another 500 million years, perhaps more.

As of now, there are about 8 billion humans alive. In total, there have been around 100 billion humans who ever lived. If we survive to the end of Earth’s habitable period, all those who have existed so far will have been the first raindrops in a hurricane.

If we’re just asking about what seems possible for the future population of humanity, the numbers are breathtakingly large. Assuming for simplicity that there will be 8 billion people for each century of the next 500 million years,10 our total population would be on the order of forty quadrillion. We think this clearly demonstrates the importance of the long-run future.

And even that might not be the end. While it remains speculative, space settlement may point the way toward outliving our time on planet Earth.11 And once we’re no longer planet-bound, the potential number of people worth caring about really starts getting big.

In What We Owe the Future, philosopher and 80,000 Hours co-founder Will MacAskill wrote:

…if humanity ultimately takes to the stars, the timescales become literally astronomical. The sun will keep burning for five billion years; the last conventional star formations will occur in over a trillion years; and, due to a small but steady stream of collisions between brown dwarfs, a few stars will still shine a million trillion years from now.

The real possibility that civilisation will last such a long time gives humanity an enormous life expectancy.

Some of this discussion may sound speculative and fantastical — which it is! But if you consider how fantastical our lives and world would seem to humans 100,000 years ago, you should expect that the far future could seem at least as alien to us now.

And it’s important not to get bogged down in the exact numbers. What matters is that there’s a reasonable possibility that the future is very long, and it could contain a much greater number of individuals.12 So how it goes could matter enormously.

There’s another factor that expands the scope of our moral concern for the future even further. Should we care about individuals who aren’t even human?

It seems true to us that the lives of non-human animals in the present day matter morally — which is why factory farming, in which billions of farmed animals suffer every day, is such a moral disaster.13 The suffering and wellbeing of future non-human animals matters no less.

And if the far-future descendants of humanity evolve into a different species, we should probably care about their wellbeing as well. We think we should even potentially care about possible digital beings in the future, as long as they meet the criteria for moral patienthood — such as, for example, being able to feel pleasure and pain.

We’re highly uncertain about what kinds of beings will inhabit the future, but we think humanity and its descendants have the potential to play a huge role. And we want to have a wide scope of moral concern to encompass all those for whom life can go well or badly.14

When we think about the possible scale of the future ahead of us, we feel humbled. But we also believe these possibilities present a gigantic opportunity to have a positive impact for those of us who have appeared so early in this story.

The immense stakes involved strongly suggest that, if there’s something we can do to have a significant and predictably positive impact on the future, we have good reason to try.

Arnold Paul, CC BY-SA 2.5 (cropped)

3. We have an opportunity to affect how the long-run future goes

When Foote discovered the mechanism of climate change, she couldn’t have foreseen how the future demand for fossil fuels would trigger a consequential global rise in temperatures.

So even if we have good reason to care about how the future unfolds, and we acknowledge that the future could contain immense numbers of individuals whose lives matter morally, we might still wonder: can anyone actually do anything to improve the prospects of the coming generations?

It’d be better for the future if we avoid extinction, manage our resources carefully, foster institutions that promote cooperation rather than violent conflict, and responsibly develop powerful technology.

Many things we do affect the future in some way. If you have a child or contribute to compounding economic growth, the effects of these actions ripple out over time, and to some extent, change the course of history. But these effects are very hard to assess. The question is whether we can predictably have a positive impact over the long term.

We think we can. For example, we believe that it’d be better for the future if we avoid extinction, manage our resources carefully, foster institutions that promote cooperation rather than violent conflict, and responsibly develop powerful technology.

We’re never going to be totally sure our decisions are for the best — but often we have to make decisions under uncertainty, whether we’re thinking about the long-term future or not. And we think there are reasons to be optimistic about our ability to make a positive difference.

The following subsections discuss four primary approaches to improving the long-run future:

Reducing extinction risk

One plausible tactic for improving the prospects of future generations is to increase the chance that they get to exist at all.

Of course, if there was a nuclear war or an asteroid that ended civilization, most people would agree that it was an unparalleled calamity.

Longtermism suggests, though, that the stakes involved could be even higher than they first seem. Sudden human extinction wouldn’t just end the lives of the billions currently alive — it would cut off the entire potential of our species. As the previous section discussed, this would represent an enormous loss.

And it seems plausible that at least some people can meaningfully reduce the risks of extinction. We can, for example, create safeguards to reduce the risk of accidental launches of nuclear weapons, which might trigger a cataclysmic escalatory cycle that brings on nuclear winter. And NASA has been testing technology to potentially deflect large near-Earth objects on dangerous trajectories.15 Our efforts to detect asteroids that could pose an extinction threat have arguably already proven extremely cost-effective.



So if it’s true that reducing the risk of extinction is possible, then people today can plausibly have a far-reaching impact on the long-run future. At 80,000 Hours, our current understanding is that the biggest risks of extinction we face come from advanced artificial intelligence, nuclear war, and engineered pandemics.16

And there are real things we can do to reduce these risks, such as:

  • Developing broad-spectrum vaccines that protect against a wide range of pandemic pathogens
  • Enacting policies that restrict dangerous practices in biomedical research
  • Inventing more effective personal protective equipment
  • Increasing our knowledge of the internal workings of AI systems, to better understand when and if they could pose a threat
  • Technical innovations to ensure that AI systems behave how we want them to
  • Increasing oversight of private development of AI technology
  • Facilitating cooperation between powerful nations to reduce threats from nuclear war, AI, and pandemics.

We will never know with certainty how effective any given approach has been in reducing the risk of extinction, since you can’t run a randomised controlled trial with the end of the world. But the expected value of these interventions can still be quite high, even with significant uncertainty.17

One response to the importance of reducing extinction risk is to note that it’s only positive if the future is more likely to be good than bad on balance. That brings us onto the next way to help improve the prospects of future generations.

Positive trajectory changes

Preventing humanity’s extinction is perhaps the clearest way to have a long-term impact, but other possibilities may be available. If we’re able to take actions that influence whether our future is full of value or is comparatively bad, we would have the opportunity to make an extremely big difference from a longtermist perspective. We can call these trajectory changes.18

Climate change, for example, could potentially cause a devastating trajectory shift. Even if we believe it probably won’t lead to humanity’s extinction, extreme climate change could radically reshape civilisation for the worse, possibly curtailing our viable opportunities to thrive over the long term.

There might even be potential trajectories that could be even worse. For example, humanity might get stuck with a value system that undermines general wellbeing and may lead to vast amounts of unnecessary suffering.

How could this happen? One way this kind of value ‘lock-in’ could occur is if a totalitarian regime establishes itself as a world government and uses advanced technology to sustain its rule indefinitely.19 If such a thing is possible, it could snuff out opposition and re-orient society away from what we have most reason to value.

We might also end up stagnating morally such that, for instance, the horrors of poverty or mass factory farming are never mitigated and are indeed replicated on even larger scales.

It’s hard to say exactly what could be done now to reduce the risks of these terrible outcomes. We’re generally less confident in efforts to influence trajectory changes compared to preventing extinction. If such work is feasible, it would be extremely important.

Trying to strengthen liberal democracy and promote positive values, such as by advocating on behalf of farm animals, could be valuable to this end. But many questions remain open about what kinds of interventions would be most likely to have an enduring impact on these issues over the long run.

Grappling with these issues and ensuring we have the wisdom to handle them appropriately will take a lot of work, and starting this work now could be extremely valuable.

Cobija, CC BY-SA 3.0, via Wikimedia Commons

Longtermist research

This brings us to the third approach to longtermist work: further research.

Asking these types of questions in a systematic way is a relatively recent phenomenon. So we’re confident that we’re pretty seriously wrong about at least some parts of our understanding of these issues. There are probably several suggestions in this article that are completely wrong — the trouble is figuring out which.

So we believe much more research into whether the arguments for longtermism are sound, as well as potential avenues for having an impact on future generations, is called for. This is one reason why we include ‘global priorities research’ among the most pressing problems for people to work on.

Capacity building

The fourth category of longtermist approaches is capacity building — that is, investing in resources that may be valuable to put toward longtermist interventions down the line.

In practice, this can take a range of forms. At 80,000 Hours, we’ve played a part in building the effective altruism community, which is generally aimed at finding and understanding the world’s most pressing problems and how to solve them. Longtermism is in part an offshoot of effective altruism, and having this kind of community may be an important resource for addressing the kinds of challenges longtermism raises.

There are also more straightforward ways to build resources, such as investing funds now so they can grow over time, potentially to be spent at a more pivotal time when they’re most needed.

You can also invest in capacity building by supporting institutions, such as government agencies or international bodies, that have the mission of stewarding efforts to improve the prospects of the long-term future.

Summing up the arguments

To sum up: there’s a lot on the line.

The number and size of future generations could be vast. We have reason to care about them all.

Those who come after us will have to live with the choices we make now. If they look back, we hope they’ll think we did right by them.

But the course of the future is uncertain. Humanity’s choices now can shape how events unfold. Our choices today could lead to a prosperous future for our descendants, or the end of intelligent life on Earth — or perhaps the rise of an enduring, oppressive regime.

We feel we can’t just turn away from these possibilities. Because so few of humanity’s resources have been devoted to making the future go well, those of us who have the means should figure out whether and how we can improve the chances of the best outcomes and decrease the chances of the worst.

We can’t — and don’t want to — set our descendants down a predetermined path that we choose for them now; we want to do what we can to ensure they have the chance to make a better world for themselves.

Those who come after us will have to live with the choices we make now. If they look back, we hope they’ll think we did right by them.

A protostar is embedded within a cloud of material feeding its growth. Credit: NASA, ESA, CSA, STScI

Objections to longtermism

In what follows, we’ll discuss a series of common objections that people make to the argument for longtermism.

Some of them point to important philosophical considerations that are complex but that nonetheless seem to have solid responses. Others raise important reasons to doubt longtermism that we take seriously and that we think are worth investigating further. And some others are misunderstandings or misrepresentations of longtermism that we think should be corrected. (Note: though long, this list doesn’t cover all objections!)

Making moral decisions always involves tradeoffs. We have limited resources, so spending on one issue means we have less to spend on another. And there are many deserving causes we could devote our efforts to. If we focus on helping future generations, we will necessarily not prioritise as highly many of the urgent needs in the present.

But we don’t think this is as troubling an objection to longtermism as it may initially sound, for at least three reasons:

1. Most importantly, many longtermist priorities, especially reducing extinction risk, are also incredibly important for people alive today. For example, we believe preventing an AI-related catastrophe or a cataclysmic pandemic are two of the top priorities, in large part because of their implications for future generations. But these risks could materialise in the coming decades, so if our efforts succeed most people alive today would benefit. Some argue that preventing global catastrophes could actually be the single most effective way to save the lives of people in the present.

2. If we all took moral impartiality more seriously, there would be a lot more resources going to help the worst-off today — not just the far future. Impartiality is the idea that we should care about the interests of individuals equally, regardless of their nationality, gender, race, or other characteristics that are morally irrelevant. This impartiality is part of what motivates longtermism — we think the interests of future individuals are often unjustifiably undervalued.

We think if impartiality were taken more seriously in general, we’d live in a much better world that would commit many more resources than it currently does toward alleviating all kinds of suffering, including for the present generation. For example, we’d love to see more resources go toward fighting diseases, improving mental health, reducing poverty, and protecting the interests of animals.

3. Advocating for any moral priority means time and resources are not going to another cause that may also be quite worthy of attention. Advocates for farmed animals’ or prisoners’ rights are in effect deprioritising the interests of alternative potential beneficiaries, such as the global poor. So this is not just an objection to longtermism — it’s an objection to any kind of prioritisation.

Ultimately, this objection hinges on the question of whether future generations are really worth caring about — which is what the rest of this article is about.

Some people, especially those trained in economics, claim that we shouldn’t treat individual lives in the future equally to lives today. Instead, they argue, we should systematically discount the value of future lives and generations by a fixed percentage.

(We’re not talking here about discounting the future due to uncertainty, which we cover below.)

When economists compare benefits in the future to benefits in the present, they typically reduce the value of the future benefits by some amount called the “discount factor.” A typical rate might be 1% per year, which means that benefits in 100 years are only worth 36% as much as benefits today, and benefits in 1,000 years are worth almost nothing.

This may seem like an appealing way to preserve the basic intuition we began with — that we have strong reasons to care about the wellbeing of future generations — while avoiding the more counterintuitive longtermist claims that arise from considering the potentially astronomical amounts of value that our universe might one day hold. On this view, we would care about future generations, but not as much as the present generation, and mostly only the generations that will come soon after us.

We agree there are good reasons to discount economic benefits. One reason is that if you receive money now, you can invest it, and earn a return each year. This means it’s better to receive money now rather than later. People in the future might also be wealthier, which means that money is less valuable to them.

However, these reasons don’t seem to apply to welfare — people having good lives. You can’t directly ‘invest’ welfare today and get more welfare later, like you can with money. The same seems true for other intrinsic values, such as justice. And longtermism is about reasons to care about the interests of future generations, rather than wealth.

As far as we know, most philosophers who have worked on the issue don’t think we should discount the intrinsic value of future lives — even while they strongly disagree about other questions in population ethics. It’s a simple principle that is easy to accept: one person’s happiness is worth just the same amount no matter when it occurs.

Indeed, if you suppose we can discount lives in the far future, we can easily end up with conclusions that sound absurd. For instance, a 3% discount rate would imply that the suffering of one person today is morally equal to the suffering of 16 trillion people in 1,000 years. This seems like a truly horrific conclusion to accept.

And any discount rate will mean that, if we found some reliable way to save 1 million lives from intense suffering in either 1,000 years or 10,000 years, it would be astronomically more important to choose the sooner option. This, too, seems very hard to accept.20

If we reject the discounting of the value of future lives, then the many potential generations that could come after us are still worthy of moral concern. And this doesn’t stand in tension with the economic practice of discounting monetary benefits.

If you’d like to see a more technical discussion of these issues, see Discounting for Climate Change by Hilary Graves. There is a more accessible discussion at 1h00m50s in our podcast with Toby Ord and in Chapter 4 of Stubborn Attachments by Tyler Cowen.

There are some practical, rather than intrinsic, reasons to discount the value of the future. In particular, our uncertainty about how the future will unfold makes it much harder to influence than the present, and even more near-term actions can be exceedingly difficult to forecast.

And because of the possibility of extinction, we can’t even be confident that the future lives we think are so potentially valuable will come into existence. As we’ve argued, that gives us reason to reduce extinction risks when it’s feasible — but it also gives us reason to be less confident these lives will exist and thus to weight them somewhat less in our deliberations.

In the same way, a doctor performing triage may choose to prioritise caring for a patient who had a good chance of surviving their injuries over one who has much less clear likelihood of survival regardless of the medical care they receive.

This uncertainty — along with the extreme level of difficulty in trying to predict the long-term impacts of our actions — certainly makes it much harder to help future generations, all else equal. And in effect, this point lowers the value of working to benefit future generations.

So even if we can affect how things unfold for future generations, we’re generally going to be very far from certain that we are actually making things better. And arguably, the further away in time the outcomes of our actions are, the less sure we can be that they will come about. Trying to improve the future will never be straightforward.

Still, even given the difficulty and uncertainty, we think the potential value at stake for the future means that many uncertain projects are still well worth the effort.

You might disagree with this conclusion if you believe that human extinction is so likely and practically unavoidable in the future that the chance that our descendants will still be around rapidly declines as we look a few centuries down the line. We don’t think it’s that likely — though we are worried about it.

Journalist Kelsey Piper critiqued MacAskill’s argument for longtermist interventions focused on positive trajectory changes (as opposed to extinction risks) in Asterisk, writing:

What share of people who tried to affect the long-term future succeeded, and what share failed? How many others successfully founded institutions that outlived them — but which developed values that had little to do with their own?

Most well-intentioned, well-conceived plans falter on contact with reality. Every simple problem splinters, on closer examination, into dozens of sub-problems with their own complexities. It has taken exhaustive trial and error and volumes of empirical research to establish even the most basic things about what works and what doesn’t to improve peoples’ lives.

Piper does still endorse working on extinction reduction, which she thinks is a more tractable course of action. Her doubts about the possibility of reliably anticipating our impact on the trajectory of the future, outside of extinction scenarios, are worth taking very seriously.

You might have a worry about longtermism that goes deeper than just uncertainty. We act under conditions of uncertainty all the time, and we find ways to manage it.

There is a deeper problem known as cluelessness. While uncertainty is about having incomplete knowledge, cluelessness refers to the state of having essentially no basis of knowledge at all.

Some people believe we’re essentially clueless about the long-term effects of our actions. This is because virtually every action we take may have extremely far-reaching unpredictable consequences. In time travel stories, this is sometimes referred to as the “butterfly effect” — because something as small as a butterfly flapping its wings might influence air currents just enough to cause a monsoon on the other side of the world (at least for illustrative purposes).

If you think your decision of whether to go to the grocery store on Thursday or Friday might determine whether the next Gandhi or Stalin is born, you might conclude that actively trying to make the future go well is a hopeless task.

Like some other important issues discussed here, cluelessness remains an active area of philosophical debate, so we don’t think there’s necessarily a decisive answer to these worries. But there is a plausible argument, advanced philosopher and advisor to 80,000 Hours Hilary Greaves that longtermism is, in fact, the best response to the issue of cluelessness.

This is because cluelessness hangs over the impact of all of our actions. Work trying to improve the lives of current generations, such as direct cash transfers, may predictably benefit a family in the foreseeable future. But the long-term consequences of the transfer are a complete mystery.

Successful longtermist interventions, though, may not have this quality — particularly interventions to prevent human extinction. If we, say, divert an asteroid that would otherwise have caused the extinction of humanity, we are not clueless about the long-term consequences. Humanity will at least have the chance to continue existing into the far future, which it wouldn’t have otherwise had.

There’s still uncertainty, of course, in preventing extinction. The long-term consequences of such an action aren’t fully knowable. But we’re not clueless about them either.

If it’s correct that the problem of cluelessness bites harder for some near-term interventions than longtermist ones, and perhaps least of all for preventing extinction, then this apparent objection doesn’t actually count against longtermism.

For an alternative perspective, though, check out The 80,000 Hours Podcast interview with Alexander Berger.

Because of the nature of human reproduction, the identity of who gets to be born is highly contingent. Any individual is the result of the combination of one sperm and one egg, and a different combination of sperm and egg would’ve created a different person. Delaying the act of conception at all — for example, by getting stuck at a red light on your way home — can easily result in a different sperm fertilising the egg, which means another person with a different combination of genes will be born.

This means — somewhat surprisingly — that pretty much all our actions have the potential to impact the future by changing which individuals get born in the future.

If you care about affecting the future in a positive way, this creates a perplexing problem. Many actions undertaken to improve the future, such as trying to reduce the harmful effects of climate change or developing a new technology to improve people’s lives, may deliver the vast majority of their benefits to people who wouldn’t have existed had the course of action never been taken.

So while it seems obviously good to improve the world in this way, it may be impossible to ever point to specific people in the future and say they were made better off by these actions. You can make the future better overall, but you may not make it better for anyone in particular.

Of course, the reverse is true: you may take some action that makes the future much worse, but all the people who experience the consequences of your actions may never have existed had you chosen a different course of action.

This is known as the ‘non-identity problem.’ Even when you can make the far future better with a particular course of action, you will almost certainly never make any particular individuals in the far future better off than they otherwise would be.

Should this problem cause us to abandon longtermism? We don’t think so.

While the issue is perplexing, accepting it as a refutation of longtermism would prove too much. It would, for example, undermine much of the very plausible case that policymakers should in the past have taken significant steps to limit the effects of climate change (since those policy changes can be expected to, in the long run, lead to different people being born).

Or consider a hypothetical case of a society that is deciding what to do with its nuclear waste. Suppose there are two ways of storing it: one way is cheap, but it means that in 200 years time, the waste will overheat and expose 10,000,000 people to sickening radiation that dramatically shortens their lives. The other storage method guarantees it will never hurt anyone, but it is significantly more expensive, and it means currently living people will have to pay marginally higher taxes.

Assuming this tax policy alters behaviour just enough to start changing the identities of the children being born, it’s entirely plausible that, in 200 years time, no one would exist who would’ve existed if the cheap, dangerous policy had been implemented. This means that none of the 10,000,000 people who have their lives cut short can say they would have been better off had their ancestors chosen the safer storage method.21

Still, it seems intuitively and philosophically unacceptable to believe that a society wouldn’t have very strong reasons to adopt the safe policy over the cheap, dangerous policy. If you agree with this conclusion, then you agree that the non-identity problem does not mean we should abandon longtermism. (You may still object to longtermism on other grounds!)

Nevertheless, this puzzle raises pressing philosophical questions that continue to generate debate, and we think better understanding these issues is an important project.

We said that we thought it would be very bad if humanity was extinguished, in part because future individuals who might have otherwise been able to live full and flourishing lives wouldn’t ever get the chance.

But this raises some issues related to the ‘non-identity problem.’ Should we actually care whether future generations come into existence, rather than not?

Some people argue that perhaps we don’t actually have moral reasons to do things that affect whether individuals exist — in which case ensuring that future generations get to exist, or increasing the chance that humanity’s future is long and expansive or would be morally neutral in itself.

This issue is very tricky from a philosophical perspective; indeed, a minor subfield of moral philosophy called population ethics sets out to answer this and related questions.

So we can’t expect to fully address the question here. But we can give a sense of why we think working to ensure humanity survives and that the future is filled with flourishing lives is a high moral priority.

Consider first a scenario in which you, while travelling the galaxy in a spaceship, come across a planet filled with an intelligent species leading happy, moral, fulfilled lives. They haven’t achieved spaceflight, and may never do so, but they appear likely to have a long future ahead of them on their planet.

Would it not seem like a major tragedy if, say, an asteroid were on course to destroy their civilization? Of course, any plausible moral view would advise saving the species for their own sakes. But it also seems like it’s an unalloyed good that, if you divert the asteroid, this flourishing species will be able to continue on for many future generations, flourishing in their corner of the universe.

If we have that view about that hypothetical alien world, we should probably have the same view of our own planet. Humans, of course, aren’t necessarily that happy, moral, and fulfilled for their lives. But the vast majority of us want to keep living — and it seems at least possible that our descendants could have lives many times more flourishing than we have. They might even ensure that all other sentient beings have joyous lives well-worth living. This seems to give us strong reasons to make this potential a reality.

For a different kind of argument along these lines, you can read Joe Carlsmith’s “Against neutrality about creating happy lives.”

Some people advocate a ‘person-affecting’ view of ethics. This view is sometimes summed up with the quip: “ethics is about helping make people happy, not making happy people.”

In practice, this means we only have moral obligations to help those who are already alive22 — not to enable more people to exist with good lives. For people who hold such views, it may be permissible to create a happy person, but doing so is morally neutral.

This view has some plausibility, and we don’t think it can be totally ignored. However, philosophers have uncovered a number of problems with it.

Suppose you have the choice to bring into existence one person with an amazing life, or another person whose life is barely worth living, but still more good than bad. Clearly, it seems better to bring about the amazing life.

But if creating a happy life is neither good nor bad, then we have to conclude that both options are neither good nor bad. This implies the options are equal, and you have no reason to do one or the other, which seems bizarre.

And if we accepted a person-affecting view, it might be hard to make sense of many of our common moral beliefs around issues like climate change. For example, it would imply that policymakers in the 20th century might have had little reason to mitigate the impact of CO2 emissions on the atmosphere if the negative effects would only affect people who would be born several decades in the future. (This issue is discussed more above.)

This is a complex debate, and rejecting the person-affecting view also has counterintuitive conclusions. In particular, Parfit showed that if you agree that it’s good to create people whose lives are more good than bad, there is a strong argument for the conclusion that we could have a better world filled with a huge number of people whose lives are just barely worth living. He called this the “repugnant conclusion”.

Both sides make important points in this debate. You can see a summary of the arguments in this public lecture by Hilary Greaves (based on this paper). It’s also discussed in our podcast with Toby Ord.

We’re uncertain about what the right position is, but we’re inclined to reject person-affecting views. Since many people hold something like the person-affecting view, though, we think it deserves some weight, and that means we should act as if we have somewhat greater obligations to help someone who’s already alive compared to someone who doesn’t exist yet. (This is an application of moral uncertainty).

One note however: even people who otherwise embrace a person-affecting view often think that is morally bad to do something that brings someone into existence who has a life full of suffering and who wishes they’d never been born. If that’s right, you should still think that we have strong moral reasons to care about the far future, because there’s the possibility it could be horrendously bad as well as very good for a large number of individuals. On any plausible view, there’s a forceful case to be made for working to avert astronomical amounts of suffering. So even someone who believes strongly in a person-affecting view of ethics might have reason to embrace a form of longtermism that prioritises averting large-scale suffering in the future.

Trying to weigh this up, we think society should have far greater concern for the future than it does now, and that as with climate change, it often makes sense to prioritise making things go well for future individuals. In the case of climate change, for example, it was likely the case that society should have long ago taken on the non-trivial costs of financing efforts to develop highly reliable clean energy and navigating away from a carbon-intensive economy.

Because of moral uncertainty, though, we care more about the present generation than we would if we naively weighed up the numbers.

Yes, it would be arrogant. But longtermism doesn’t require us to know the future.

Instead, the practical implication of longtermism is that we take steps that are likely to be good over the wide range of possible futures. We think it’s likely better for the future if, as we said above, we avoid extinction, we manage our resources carefully, we foster institutions that promote cooperation rather than violent conflict, and we responsibly develop powerful technology. None of these strategies requires us knowing what the future will look like.

We talk more about the importance of all this uncertainty in the sections above.

This isn’t exactly an objection, but one response to longtermism asserts not that the view is badly off track but that it’s superfluous.

This may seem plausible if longtermism primarily inspires us to prioritise reducing extinction risks. As discussed above, doing so could benefit existing people — so why even bother talking about the benefits to future generations?

One reply is: we agree that you don’t need to embrace longtermism to support these causes! And we’re happy if people do good work whether or not they agree with us on the philosophy.

But we still think the argument for longtermism is true, and we think it’s worth talking about.

Firstly, when we actually try to compare the importance of work in certain cause areas — such as global health or mitigating the risk of extinction from nuclear war — whether and how much we weigh the interests of future generations may play a decisive role in our conclusions about prioritisation.

Moreover, some longtermist priorities, such as ensuring that we avoid the lock-in of bad values or developing a promising framework for space governance, may be entirely ignored if we don’t consider the interests of future generations.

Finally, if it’s right that future generations deserve much more moral concern than they currently get, it just seems good for people to know that. Maybe issues will come up in the future that aren’t extinction threats but which could still predictably affect the long-run future – we’d want people to take those issues seriously.

In short, no. Total utilitarianism is the view that we are obligated to maximise the total amount of positive experiences over negative experiences, typically by weighting for intensity and duration.

This is one specific moral view, and many of its proponents and sympathisers advocate for longtermism. But you can easily reject utilitarianism of any kind and still embrace longtermism.

For example, you might believe in ‘side constraints’ — moral rules about what kinds of actions are impermissible, regardless of the consequences. So you might believe that you have strong reasons to promote the wellbeing of individuals in the far future, so long as doing so doesn’t require violating anyone’s moral rights. This would be one kind of non-utilitarian longtermist view.

You might also be a pluralist about value, in contrast to utilitarians who think a singular notion of wellbeing is the sole true value. A non-utilitarian might intrinsically value, for instance, art, beauty, achievement, good character, knowledge, and personal relationships, quite separately from their impact on wellbeing.

(See our definition of social impact for how we incorporate these moral values into our worldview.)

So you might be a longtermist precisely because you believe the future is likely to contain vast amounts of all the many things you value, so it’s really important that we protect this potential.

You could also think we have an obligation to improve the world for future generations because we owe it to humanity to “pass the torch”, rather than squander everything people have done to build up civilisation. This would be another way of understanding moral longtermism that doesn’t rely on total utilitarianism.23

Finally, you can reject the “total” part of utilitarianism and still believe longtermism. That is, you might believe it’s important to make sure the future goes well in a generally utilitarian sense without thinking that means we’ll need to keep increasing the population size in order to maximise total wellbeing. You can read more about different kinds of views in population ethics here.

As we discussed above, people who don’t think it’s morally good to bring a flourishing population into existence usually think it’s still important to prevent future suffering — in which case you might support a longtermism focused on guarding against the worst outcomes for future generations.

No.

We believe, for instance, that you shouldn’t have a harmful career just because you think you can do more good than bad with the money you’ll earn. There are practical, epistemic, and moral reasons that justify this stance.

And as a general matter, we think it’s highly unlikely to be the case that working in a harmful career will be the path that has the best consequences overall.

Some critics of longtermism say the view can be used to justify all kinds of egregious acts in the name of a glorious future. We do not believe this, in part because there are plenty of plausible intrinsic reasons to object to egregious acts on their own, even if you think they’ll have good consequences. As we explained in our article on the definition of ‘social impact’:

We don’t think social impact is all that matters. Rather, we think people should aim to have a greater social impact within the constraints of not sacrificing other important values – in particular, while building good character, respecting rights and attending to other important personal values. We don’t endorse doing something that seems very wrong from a commonsense perspective in order to have a greater social impact.

Perhaps even more importantly, it’s bizarrely pessimistic to believe that the best way to make the future go well is to do horrible things now. This is very likely false, and there’s little reason anyone should be tempted by this view.

Some of the claims in this article may sound like science fiction. We’re aware this can be off-putting to some readers, but we think it’s important to be upfront about our thinking.

And the fact that a claim sounds like science fiction is not, on its own, a good reason to dismiss it. Many speculative claims about the future have sounded like science fiction until technological developments made them a reality.

From Eunice Newton Foote’s perspective in the 19th century, the idea that the global climate would actually be transformed based on a principle she discovered in a glass cylinder may have sounded like science fiction. But climate change is now our reality.

Similarly, the idea of the “atomic bomb” had literally been science fiction before Leo Szilard discovered the possibility of the nuclear chain reaction in 1933. Szilard first read about such weapons in H.G. Wells’ The World Set Free. As W. Warren Wager explained in The Virginia Quarterly:

Unlike most scientists then doing research into radioactivity, Szilard perceived at once that a nuclear chain reaction could produce weapons as well as engines. After further research, he took his ideas for a chain reaction to the British War Office and later the Admiralty, assigning his patent to the Admiralty to keep the news from reaching the notice of the scientific community at large. “Knowing what this [a chain reaction] would mean,” he wrote, “—and I knew it because I had read H.G. Wells—I did not want this patent to become public.”

This doesn’t mean we should accept any idea without criticism. And indeed, you can reject many of the more ‘sci-fi’ claims of some people who are concerned with future generations — such as the possibility of space settlement or the risks from artificial intelligence — and still find longtermism compelling.

One worry about longtermism some people have is that it seems to rely on having a very small chance of achieving a very good outcome.

Some people think this sounds suspiciously like Pascal’s wager, a highly contentious argument for believing in God — or a variant of this idea, “Pascal’s mugging.” The concern is that this type of argument may be used to imply an apparent obligation to do absurd or objectionable things. It’s based on a thought experiment, as we described in a different article:

A random mugger stops you on the street and says, “Give me your wallet or I’ll cast a spell of torture on you and everyone who has ever lived.” You can’t rule out with 100% probability that he won’t — after all, nothing’s 100% for sure. And torturing everyone who’s ever lived is so bad that surely even avoiding a tiny, tiny probability of that is worth the $40 in your wallet? But intuitively, it seems like you shouldn’t give your wallet to someone just because they threaten you with something completely implausible.

This deceptively simple problem raises tricky issues in expected value theory, and it’s not clear how they should be resolved — but it’s typically assumed that we should reject arguments that rely on this type of reasoning.

The argument for longtermism given above may look like a form of this argument because it relies in part on the premise that the number of individuals in the future could be so large. Since it’s a relatively novel, unconventional argument, it may sound suspiciously like the mugger’s (presumably hollow) threat in the thought experiment.

But there are some key differences. To start, the risks to the long-term future may be far from negligible. Toby Ord estimated the chance of an existential catastrophe that effectively curtails the potential of future generations in the next century at 1 in 6.24

Now, it may be true that any individual’s chance of meaningfully reducing these kinds of threats is much, much smaller. But we accept small chances of doing good all the time — that’s why you might wear a seatbelt in a car, even though in any given drive your chances of being in a serious accident are miniscule. Many people buy life insurance to guarantee that their family members will have financial support in the unlikely scenario that they die young.

And while an individual is unlikely to be solely responsible for driving down the risk of human extinction by any significant amount (in the same way no one individual could stop climate change), it does seem plausible that a large group of people working diligently and carefully might be able to do it. And if the large group of people can achieve this laudable end, then taking part in this collective action isn’t comparable to Pascal’s mugging.

But if we did conclude the chance to reduce the risks humanity faces is truly negligible, then we would want to look much more seriously into other priorities, especially since there are so many other pressing problems. As long as it’s true, though, that there are genuine opportunities to have a significant impact on improving the prospects for the future, then longtermism does not rely on suspect and extreme expected value reasoning.

This is a lot to think about. So what are our bottom lines on how we think we’re most likely to be wrong about longtermism?

Here are a few possibilities we think are worth taking seriously, even though they don’t totally undermine the case from our perspective:

  • Morality may require a strong preference for the present: There might be strong moral reasons to give preference to existing people and individuals over future generations. This might be because something like a person-affecting view is true (described above) or maybe even because we should systematically discount the value of future beings.
    • We don’t think the arguments for such a strong preference are very compelling, but given the high levels of uncertainty in our moral beliefs, we can’t confidently rule it out.
  • Reliably affecting the future may be infeasible. It’s possible that further research will ultimately conclude that the opportunities for impacting the far future are essentially non-existent or extremely limited. It’s hard to believe we could ever entirely close the question — researchers who come to this conclusion in the future could themselves be mistaken — but it might dramatically reduce our confidence that pursuing a longtermist agenda is worthwhile and thus leave the project as a pretty marginal endeavour.

  • Reducing extinction risk may be intractable beyond a certain point. It’s possible that there’s a base level of extinction risk that humans will have to accept at some point and that we can’t reduce any further. And if, for instance, there were an irreducible risk of an extinction catastrophe at 10 percent every century, then the future, in expectation, would be much less significant than we think. This would dramatically reduce the pull of longtermism.

  • A crucial consideration could change our assessment in ways we can’t predict. This falls into the general category of ‘unknown unknowns,’ which are always important to be on the watch for.

You could also read the following essays criticising longtermism that we have found interesting:

If I don’t agree with 80,000 Hours about longtermism, can I still benefit from your advice?

Yes!

We want to be candid about what we believe and what our priorities are, but we don’t think everyone needs to agree with us.

And we have lots of advice and tools that are broadly useful for people thinking about their careers, regardless of what they think about longtermism.

There are also many places where longtermist projects converge with other approaches to thinking about having a positive impact with your career. For example, working to prevent pandemics seems robustly good whether you prioritise near- or long-term benefits.

Though we focus as an organisation on issues that may affect all future generations, we would generally be really happy to also see more people working for the benefit of the global poor and farmed animals, two tractable causes that we think are unduly neglected in the near term. We also discuss these issues on our podcast and list jobs for them on our job board.

Credit: Yen Chao CC2.0

What are the best ways to help future generations right now?

While answering this question satisfactorily would require a sweeping research agenda in itself, we do have some general thoughts about what longtermism means for our practical decision making. And we’d be excited to see more attention paid to this question.

Some people may be motivated by these arguments to find opportunities to donate to longermist projects or cause areas. We believe Open Philanthropy — which is a major funder of 80,000 Hours — does important work in this area.

But our primary aim is to help people have impactful careers. Informed by longtermism, we have created a list of what we believe are the most pressing problems to work on in the world. These problems are important, neglected, and tractable.

As of this writing, the top eight problem areas are:

  1. Risks from artificial intelligence
  2. Catastrophic pandemics
  3. Building effective altruism
  4. Global priorities research
  5. Nuclear war
  6. Improving decision making (especially in important institutions)
  7. Climate change
  8. Great power conflict

We’ve already given few examples of concrete ways to tackle these issues above.

The above list is provisional, and it is likely to change as we learn more. We also list many other pressing problems that we believe are highly important from a longtermist point of view, as well as a few that would be high priorities if we rejected longtermism.

We hope more people will challenge our ideas and help us think more clearly about them. As we have argued, the stakes are incredibly high.

We have a related list of high-impact careers that we believe are appealing options for people who want to work to address these and related problems and to help the long-term future go well.

But we don’t have all the answers. Research in this area could reveal crucial considerations that might overturn longtermism or cast it in a very different light. There are likely pressing cause areas we haven’t thought of yet.

We hope more people will challenge our ideas and help us think more clearly about them. As we have argued, the stakes are incredibly high. So it’s paramount that, as much as is feasible, we get this right.

Want to focus your career on the long-run future?

If you want to work on ensuring the future goes well, such as controlling nuclear weapons or shaping the development of artificial intelligence or biotechnology, you can speak to our team one-on-one.

We’ve helped hundreds of people choose an area to focus, make connections, and then find jobs and funding in these areas. If you’re already in one of these areas, we can help you increase your impact within it.

Speak to us

Learn more

Read next

This article is part of our advanced series. See the full series, or keep reading:

Plus, join our newsletter and we’ll mail you a free book

Join our newsletter and we’ll send you a free copy of The Precipice — a book by philosopher Toby Ord about how to tackle the greatest threats facing humanity. T&Cs here.

The post Longtermism: a call to protect future generations appeared first on 80,000 Hours.

]]>
David Chalmers on the nature and ethics of consciousness https://80000hours.org/podcast/episodes/david-chalmers-nature-ethics-consciousness/ Mon, 16 Dec 2019 19:48:22 +0000 https://80000hours.org/?post_type=podcast&p=68176 The post David Chalmers on the nature and ethics of consciousness appeared first on 80,000 Hours.

]]>
The post David Chalmers on the nature and ethics of consciousness appeared first on 80,000 Hours.

]]>