[D66] Existential Risk and the Future of Humanity | NYRB
R.O.
juggoto at gmail.com
Thu Feb 11 21:26:32 CET 2021
* 25 Feb 2021
* The New York Review of Books
* Jim Holt
The Power of Catastrophic Thinking
Toba Khedoori: Untitled (Clouds—Drawing), 2004–2005
*The Precipice: Existential Risk and the *
**
*Future of Humanity by Toby Ord. *
**
*Hachette, 468 pp., $30.00; *
**
*$18.99 (paper; to be published in March) *
T. S. Eliot, in his 1944 essay “What Is a Classic?,” complained that a
new kind of provincialism was becoming apparent in our culture: “a
provincialism, not of space, but of time.” What Eliot had in mind was
provincialism about the past: a failure to think of dead generations as
fully real. But one can also be guilty of provincialism about the
future: a failure to imagine the generations that will come after us, to
take seriously our responsibilities toward them.
In 1945, not long after Eliot wrote that essay, the first atomic bomb
was exploded. This made the matter of provincialism about the future all
the more acute. Now, seemingly, humanity had acquired the power to
abolish its own future. A decade later Bertrand Russell and Albert
Einstein issued a joint manifesto warning that nuclear weaponry posed
the risk of imminent human extinction, of “universal death.” (In a
letter to Einstein, Russell also predicted that the same threat would
eventually be posed by biological warfare.)
By the early 1980s, more precise ideas were being put forward about how
this could occur. In 1982 Jonathan Schell, in a much-discussed series of
articles in The New Yorker (later published as a book, The Fate of the
Earth), argued that nuclear war might well result in the destruction of
the ozone layer, making it impossible for human life to survive on
earth. In 1983 Carl Sagan and four scientific colleagues introduced the
“nuclear winter” hypothesis, according to which firestorms created by a
nuclear exchange, even a limited one, would darken the upper atmosphere
for years, causing global crop failures, universal famine, and human
extinction—an alarming scenario that helped move Ronald Reagan and
Mikhail Gorbachev to negotiate reductions in their countries’ nuclear
arsenals. Neither Schell nor Sagan was a philosopher. Yet each raised a
philosophical point: with the advent of nuclear weapons and other
dangerous new technologies, we ran the risk not only of killing off all
humans alive today, but also of depriving innumerable generations of the
chance to exist. Humanity’s past has been relatively brief: some 300,000
years as a species, a few thousand years of civilization. Its potential
future, by contrast, could extend for millions or billions of years,
encompassing many trillions of sentient, rational beings yet to be born.
It was this future—the adulthood of humanity— that was now in jeopardy.
“If our species does destroy itself,” Schell wrote, “it will be a death
in the cradle—a case of infant mortality.”
The idea that potential future lives as well as actual ones must be
weighed in our moral calculus was soon taken up by professional
philosophers. In 1984 Derek Parfit published his immensely influential
treatise Reasons and Persons, which, in addition to exploring issues of
rationality and personal identity
with consummate subtlety, also launched a new (and currently
flourishing) field of moral philosophy known as “population ethics.”1 At
its core is this question: How ought we to act when the consequences of
our actions will affect not only the well-being of future people but
their very existence?
It was on the final pages of Reasons and Persons that Parfit posed an
arresting hypothetical. Consider, he said, three scenarios:
(1) World peace.
(2) A nuclear war that kills 99 percent of the world’s population.
(3) A nuclear war that kills 100 percent of the world’s population.
Clearly, he observed, (2) is worse than (1), and (3) is worse than (2).
But which is the greater of the two moral differences? Most people,
Parfit guessed, would say the difference between (1) and (2) is greater
than the difference between (2) and (3). He disagreed. “I believe that
the difference between (2) and (3) is very much greater,” he wrote.
Killing off that last one percent, he observed, would mean destroying
the entire future of humanity—an inconceivably vast reduction in the sum
of possible human happiness.
Toby
Ord, the author of The Precipice, studied at Oxford under Parfit
1This field is sometimes also called “population axiology,” from the
Greek word for “value,” axía.
(who died in 2017) and calls him his “mentor.” Today Ord too is a
philosopher at Oxford and among the most prominent figures who think
deeply and systematically about existential risks to humanity.2 Ord is a
model of the engaged thinker. In addition to his academic work in
applied ethics, he has advised the World Health Organization, the World
Bank, and the British government on issues of global health and poverty.
He helped start the “effective altruism” movement and founded the
organization Giving What We Can, whose members have pledged more than $2
billion to “effective charities.” (Their donations to charities that
distribute malaria nets have already saved more than two thousand
lives.) The society’s members are governed by a pledge to dedicate at
least a tenth of what they earn to the relief of human suffering, which
grew out of a personal commitment that Ord had made. He has now made a
further pledge to limit his personal spending to £18,000 a year and give
away the rest. And he tells us that he has “signed over the entire
advance and royalties from this book to
2Others include Nick Bostrom, who directs the Future of Humanity
Institute at Oxford (and who was profiled in The New Yorker in 2015);
Martin Rees, Britain’s astronomer royal and the author of Our Final Hour
(2003); and John Leslie, a Canadian philosopher whose book The End of
the World (1996) furnished the first analytical survey of the full range
of humanextinction possibilities. charities helping protect the longterm
future of humanity.”
Ord is, in short, an admirable man. And The Precipice is in many ways an
admirable book. In some 250 brisk pages, followed by another 200 or so
pages of notes and technical appendices, he gives a comprehensive and
highly readable account of the evidence bearing on various human
extinction scenarios. He tells harrowing stories of how humanity has
courted catastrophe in the past—nuclear close calls, deadly pathogens
escaping labs, and so forth. He wields probabilities in a cogent and
often counterintuitive manner. He surveys current philosophical thinking
about the future of humanity and addresses issues of “cosmic
significance” with a light touch. And he lays out an ambitious
three-step “grand strategy” for ensuring humanity’s flourishing into the
deep future—a future that, he thinks, may see our descendants colonizing
entire galaxies and exploring “possible experiences and modes of thought
beyond our present understanding.”
These are among the virtues of The Precipice. Against them, however,
must be set two weaknesses, one philosophical, the other analytical. The
philosophical one has to do with the case Ord makes for why we should
care about the long-term future of humanity—a case that strikes me as
incomplete. Ord confesses that as a younger man he “sometimes took
comfort in the idea that perhaps the outright destruction of humanity
would not be bad at all,” since merely possible people cannot suffer if
they never come into existence. His reasons for changing his mind—for
deciding that safeguarding humanity’s future “could well be our most
important duty”—turn out to be a mixture of classical utilitarian and
“ideal goods”–based considerations that will be familiar to
philosophers. But he fails to take full account of why the future
disappearance of humanity should matter to us, the living, in the here
and now; why we should be motivated to make sacrifices today for
potential future people who, if we don’t make those sacrifices, won’t
even exist. From this philosophical weakness, which involves a why
question, stems an analytical weakness, which involves a how much
question: How much should we be willing to sacrifice today in order to
ensure humanity’s longterm future? Ord is ethically opposed to the
economic practice of “discounting,” which is a way of quantitatively
shrinking the importance of the far future. I’m with him there. But this
leaves him with a difficulty that he does not quite acknowledge. If we
are obliged to weigh the full (undiscounted) value of humanity’s
potential future in making our decisions today, we are threatened with
becoming moral slaves to that future. We will find it our duty to make
enormous sacrifices for merely potential people who might exist millions
of years from now, while scanting the welfare of actual people over the
next few centuries. And the mathematics of this, as we shall see, turn
out to be perverse: the more we sacrifice, the more we become obliged to
sacrifice.
This is not merely a theoretical problem. It leads to a distorted
picture of how we should distribute our present moral concerns,
suggesting that we should be relatively less worried about real and
ongoing developments that will gravely harm humanity without wiping it
out completely (like climate change), and relatively more worried about
notional threats that, however unlikely, could conceivably result in
human extinction (like rogue AI). Ord does not say this explicitly, but
it is implied by his way of thinking. And it should give us pause.
What is the likelihood that humanity will survive even the present
century? In 1980 Sagan estimated that the chance of human extinction
over the next hundred years was 60 percent— meaning that humanity had
less than even odds of making it beyond 2080. A careful risk analysis,
however, suggests that his estimate was grossly too pessimistic. Ord’s
accounting puts the existential risk faced by humanity in the current
century at about one in six: much better, but still the same odds as
Russian roulette. He arrives at this estimate by surveying the “risk
landscape,” whose hills and peaks represent the probabilities of all the
various threats to humanity’s future. This landscape turns out to have
some surprising features.
What apocalyptic scenario looms largest in your mind? Do you imagine the
world ending as the result of an asteroid impact or a stellar explosion?
In a nuclear holocaust or a global plague caused by biowarfare? The
former possibilities fall under the category of “natural” risks, the
latter under “anthropogenic” (human-caused) risks. Natural risks have
always been with us: ask the dinosaurs. Anthropogenic risks, by
contrast, are of relatively recent vintage, dating from the beginning of
the atomic era in 1945. That, Ord says, was when “our rapidly
accelerating technological power finally reached the threshold where we
might be able to destroy ourselves,” as Einstein and Russell warned at
the time.
Which category, natural or anthropogenic, poses the greater threat to
humanity’s future? Here it is not even close. By Ord’s reckoning, the
total anthropogenic risk over the next century is a thousand times
greater than the total natural risk. In other words, humanity is far
more likely to commit suicide than to be killed off by nature. It has
thus entered a new age of unsustainably heightened risk, what Ord calls
“the Precipice.”
We know that the extinction risk posed by natural causes is relatively
low because we have plenty of actuarial data. Humans have been around
for about three thousand centuries. If there were a sizable per-century
risk of our perishing because of a nearby star exploding, or an asteroid
slamming into the earth, or a supervolcanic eruption blackening the sky
and freezing the planet, we would have departed the scene a long time
ago. So, with a little straightforward math, we can conclude that the
total risk of our extinction by natural causes over the next century is
no more than one in 10,000. (In fact, nearly all of that risk is posed
by the supervolcanic scenario, which is less predictable than an
asteroid impact or stellar explosion.) If natural risks were all that we
had to worry about, Homo sapiens could expect to survive on earth for
another million years—which, not coincidentally, is the longevity of a
typical mammalian species.
Over, then, to the anthropogenic category. Here, as Ord observes, we
have hardly any data for calculating risks. So far, we’ve survived the
industrial era for a mere 260 years and the nuclear era for 75. That
doesn’t tell us much, from a statistical point of view, about whether
we’ll get through even the next century. So we have to rely on
scientific reasoning. And such reasoning suggests that the greatest
human-made dangers to our survival are not what you might think.
Start with the seemingly most obvious one: nuclear war. How could that
result in the absolute extinction of humanity? It is often claimed that
there are enough nuclear weapons in the world today to kill off all
humans many times over. But this, as Ord observes, is “loose talk.” It
arises from naively extrapolating from the destruction visited on
Hiroshima. That bomb killed 140,000 people. Today’s nuclear arsenal is
equivalent to 200,000 Hiroshima bombs. Multiply these two numbers, and
you get a death toll from an all-out nuclear war of 30 billion
people—about four times the world’s current population. Hence the “many
times over” claim. Ord points out that this calculation makes a couple
of big mistakes. First, the world’s population, unlike Hiroshima’s, is
not densely concentrated but spread out over a wide land area. There are
not nearly enough nuclear weapons to hit every city, town, and village
on earth. Second, today’s bigger nuclear bombs are less efficient at
killing than the Hiroshima bomb was.3 A reasonable estimate for the
death toll arising from the local effects of a full-scale nuclear
war—explosions and firestorms in large cities—is 250 million:
unspeakable, but a long way from the absolute extinction that is Ord’s
primary worry.
That leaves the global effects of nuclear war to consider. Fallout?
Spreading deadly radiation across the entire surface of the earth would
require a nuclear arsenal ten times the size of the current one.
Destruction of the ozone layer? This was the danger cited by Schell in
The Fate of the Earth, but the underlying theory has not held up.
Nuclear winter? Here lies the greatest threat, and it is one that Ord
examines in fascinating (if depressing) detail, before coming to the
conclusion that “nuclear winter appears unlikely to lead to our
extinction.” As for the chance that it would lead merely to the
unrecoverable collapse of civilization—another form of “existential
catastrophe”—he observes that New Zealand at least, owing to its coastal
location, would likely survive nuclear winter “with most of their
technology (and institutions) intact.” A cheerful thought.
All told, Ord puts the existential risk posed by nuclear war over the
next century at one in one thousand, a relatively small peak in the risk
landscape. 3
The blast damage scales up as the two-thirds root of the bomb’s
kilotonnage—a fun fact for those who, like Herman Kahn, enjoy thinking
about the unthinkable. So whence the rest of the onein-six risk figure
he arrives at? Climate change? Could global warming cause unrecoverable
collapse or even human extinction? Here too, Ord’s prognosis, though
dire, is not so dire as you might expect. On our present course, climate
change will wreak global havoc for generations and drive many nonhuman
species to extinction. But it is unlikely to wipe out humanity entirely.
Even in the extreme case where global temperatures rise by as much as 20
degrees centigrade, there will still be enough habitable land mass,
fresh water, and agricultural output to sustain at least a miserable
remnant of us.
There is, however, at least one scenario in which climate change might
indeed spell the end of human life and civilization. Called the “runaway
greenhouse effect,” this could arise— in theory—from an amplifying
feedback loop in which heat generates water vapor (a potent greenhouse
gas) and water vapor in turn traps heat. Such a feedback loop might
raise the earth’s temperature by hundreds of degrees, boiling off all
the oceans. (“Something like this probably happened on Venus,” Ord tells
us.) The runaway greenhouse effect would be fatal to most life on earth,
including humans. But is it likely? Evidence from past geological eras,
when the carbon content of the atmosphere was much higher than it is
today, suggests not. In Ord’s summation, “It is probably physically
impossible for our actions to produce the catastrophe—but we aren’t sure.”
So he puts the chance of existential doom from climate change over the
next century at one in one thousand— not quite negligible, but still a
comparatively small peak in the risk landscape. He assigns similarly
modest odds to our being doomed by other types of environmental damage,
like resource depletion or loss of biodiversity. (For me, one of the
saddest bits in the book is the claim that humans could survive the
extinction of honeybees and other pollinators, whose disappearance
“would only create a 3 to 8 percent reduction in global crop
production.” What a world.)
If neither nuclear war nor environmental collapse accounts for the
Russian roulette–level threat of doom we supposedly face over the next
century, then what does? In Ord’s analysis, the tallest peaks in the
existential risk landscape turn out to be “unaligned artificial
intelligence” and “engineered pandemics.”
Start with the lesser of the two: pandemic risk. Natural pandemics have
occurred throughout the existence of the human species, but they have
not caused our extinction. The worst of them, at least in recorded
history, was the Black Death, which came to Europe in 1347 and killed
between one quarter and one half of its inhabitants. (It also ravaged
the Middle East and Asia.) The Black Death “may have been the greatest
catastrophe humanity has seen,” Ord observes. Yet by the sixteenth
century Europe had recovered. In modern times, such “natural” pandemics
are, because of human activities, in some ways more dangerous: our
unwholesome farming practices make it easy for diseases to jump from
animals to humans, and jet travel spreads pathogens across the globe.
Still, the fossil record suggests that there is only a tiny per-century
chance that a natural pandemic could result in universal death: about
one in 10,000, Ord estimates.
Factor in human mischief, though, and the odds shorten drastically.
Thanks to biotechnology, we now have the power to create deadly new
pathogens and to resurrect old ones in more lethal and contagious forms.
As Ord observes, this power will only grow in the future. What makes
biotech especially dangerous is its rapid “democratization.” Today,
“online DNA synthesis services allow anyone to upload a DNA sequence of
their choice then have it constructed and shipped to their address.” A
pandemic that would wipe out all human life might be deliberately
engineered by “bad actors” with malign intent (like the Aum Shinrikyo
cult in Japan, dedicated to the destruction of humanity). Or it might
result from well-intentioned research gone awry (as in 1995 when
Australian scientists released a virus that unexpectedly killed 30
million rabbits in just a few weeks). Between bioterror and bio-error,
Ord puts the existential risk from an “engineered pandemic” at one in
thirty: a major summit in the risk landscape. That leaves what Ord deems
the greatest of all existential threats over the next century:
artificial intelligence. And he is hardly eccentric in this judgment.
Fears about the destructive potential of AI have been raised by figures
like Elon Musk, Bill Gates, Marvin Minsky, and Stephen Hawking.4
How might AI grow potent enough to bring about our doom? It would happen
in three stages. First, AI becomes able to learn on its own, without
expert programming. This stage has already arrived, as was demonstrated
in 2017 when the AI company DeepMind created a neural network that
learned to play Kasparov-level chess on its own in just a few hours.
Next, AI goes broad as well as deep, rivaling human intelligence not
just in specialized skills like chess but in the full range of cognitive
domains. Making the transition from specialized AI to AGI—artificial
general intelligence—is the focus of much cutting-edge research today.
Finally, AI comes not just to rival but to exceed human intelligence—a
development that, according to a 2016 survey of three hundred top AI
researchers, has a fifty-fifty chance of occurring within four decades,
and a 10 percent chance of occurring in the next five years.
But why should we fear that these ultra-intelligent machines, assuming
they do emerge, will go rogue on us? Won’t they be programmed to serve
our interests? That, as it turns out, is precisely the problem. As Ord
puts it, “Our values are too complex and subtle to specify by hand.” No
matter how careful we are in drawing up the machine’s “reward
function”—the rule-like algorithm that steers its behavior—its actions
are bound to diverge from what we really want. Getting AI in sync with
human values is called the “alignment problem,” and it may be an
insuperable one. Nor have AI researchers figured out how 4There are also
prominent skeptics— like Mark Zuckerberg, who has called Musk
“hysterical” for making so much of the alleged dangers of AI.
to make a system that, when it notices that it’s misaligned in this way,
updates its values to coincide with ours instead of ruthlessly
optimizing its existing reward function (and cleverly circumventing any
attempt to shut it down). What would you command the superintelligent AI
system to do? “Maximize human happiness,” perhaps? The catastrophic
result could be something like what Goethe imagined in “The Sorcerer’s
Apprentice.” And AI wouldn’t need an army of robots to seize absolute
power. It could do so by manipulating humans to do its destructive
bidding, the way Hitler, Stalin, and Genghis Khan did.
“The case for existential risk from AI is clearly speculative,” Ord
concedes. “Indeed, it is the most speculative case for a major risk in
this book.” But the danger that AI, in its coming superintelligent and
misaligned form, could wrest control from humanity is taken so seriously
by leading researchers that Ord puts the chance of its happening at one
in ten: by far the highest peak in his risk landscape.5 Add in some
smaller peaks for less well understood risks (nanotechnology,
high-energy physics experiments, attempts to signal possibly hostile
extraterrestrials) and utterly unforeseen technologies just over the
horizon—what might be called the “unknown unknowns”—and the putative
risk landscape is complete.
So what is to be done? The practical proposals Ord lays out for
mitigating existential risk—greater vigilance, more research into safer
technologies, strengthening international institutions—are well thought
out and eminently reasonable. Nor would they be terribly expensive to
implement. We currently spend less than a thousandth of a percent of
world gross world product on staving off technological
selfdestruction—not even a hundredth of what we spend on ice cream. Just
raising our expenditure to the ice cream threshold, as Ord suggests,
would go far in safeguarding humanity’s longterm potential.
But let’s consider a more theoretical issue: How much should we be
willing to pay in principle to ensure humanity’s future? Ord does not
explicitly address this question. Yet his way of thinking about the
value of humanity’s future puts us on a slippery slope to a preposterous
answer.
Start, as he does, with a simplifying assumption: that the value of a
century of human civilization can be captured by some number V. To make
things easy, we’ll pretend that V is constant from century to century.
(V might be taken to quantify a hundred years’ worth of net human
happiness, or of cultural achievement, or some such.) Given this
assumption, the longer humanity’s future continues, the greater its
total value will be. If humanity went on forever, the value of its
future would be infinite. But this is unlikely: eventually
5As far as I can tell, he arrives at this one-in-ten number by assuming,
in broad agreement with the AI community, that the chance of AI
surpassing human intelligence in the next century is 50 percent, and
then multiplying this number by the probability that the resulting
misalignment will prove catastrophic, which he seems to put at one in
five. the universe will come to some sort of end, and our descendants
probably won’t be able to survive that. And in each century of
humanity’s existence there is some chance that our species will fail to
make it to the next. In the present century, as we have seen, Ord puts
that chance at one in six. Let’s suppose—again, to simplify—that this
risk level remains the same in the future: a one in six risk of doom per
century. Then humanity’s expected survival time would be another six
centuries, and the value of its future would be V multiplied by six.
That is, the expected value of humanity’s future is six times the value
of the present century.
Now suppose we could take actions today that would enduringly cut this
existential risk in half, from one in six down to one in twelve. How
would that affect the expected value of humanity’s future? The answer is
that the value would double, going from 6V (the old expected value) to
12V (the new expected value). That’s a net gain of six centuries worth
of value! So we should be willing to pay a lot, if necessary, to reduce
risk in this way.
And the math gets worse. Suppose that we could somehow eliminate all
anthropogenic risk. We might achieve this, say, by going Luddite and
stamping out each and every potentially dangerous technology, seeking
fulfillment instead in an Arden-like existence of foraging for nuts and
berries, writing lyric poems, composing fugues, and proving theorems in
pure mathematics. Then the only existential risks remaining would be the
relatively tiny natural ones, which come to one in ten thousand per
century. So the expected value of humanity’s future would go from 6V to
10,000V—a truly spectacular gain. How could we not be obliged to make
whatever sacrifice this might entail, given the expected payoff in the
increased value of humanity’s future? Clearly there is something amiss
with this reasoning. Ord would say— indeed, does say—that humanity needs
risky technologies like AI if it is to flourish, so “relinquishing
further technological progress is not a solution.” But the problem is
more general than that. The more we do to mitigate risk, the longer
humanity’s expected future becomes. And by Ord’s logic, the longer that
future becomes, the more its potential value outweighs the value of the
present. As we push the existential risk closer and closer to zero,
expected gains in value from the very far future become ever more
enormous, obliging us to make still greater expenditures to ensure their
ultimate arrival. This combination of increasing marginal costs (to
reduce risk) and increasing marginal returns (in future value) has no
stable equilibrium point short of bankruptcy. At the limit, we should
direct 100 percent of our time and energy toward protecting humanity’s
long-term future against even the remotest existential threats—then wrap
ourselves in bubble wrap, just to be extra safe. When a moral theory
threatens to make unlimited demands on us in this way, that is often
taken by philosophers as a sign there is something wrong with it. (This
is sometimes called the “argument from excessive sacrifice.”) What could
be wrong with Ord’s theory? Why does it threaten to make the demands of
humanity’s future on us unmanageable? Perhaps the answer is to be sought
in asking just why we value that future—especially the parts of it that
might unfold long after we’re gone. What might go into that hypothetical
number V that we were just bandying about?
Philosophers have traditionally taken two views of this matter. On one
side, there are the classical utilitarians, who hold that all value
ultimately comes down to happiness. For them, we should value humanity’s
future because of its potential contribution to the sum of human
happiness. All those happy generations to come, spreading throughout the
galaxy! Then there are the more Platonic philosophers, who believe in
objective values that transcend mere happiness. For them, we should
value humanity’s future because of the “ideal goods”—knowledge, beauty,
justice— with which future generations might adorn the cosmos. (The term
“ideal goods” comes from the nineteenthcentury moral philosopher Henry
Sidgwick, who had both utilitarian and Platonizing tendencies.)
Ord cites both kinds of reasons for valuing humanity’s future. He
acknowledges that there are difficulties with the utilitarian account,
particularly when considerations of the quantity of future people are
balanced against the quality of their lives. But he seems more
comfortable when he doffs his utilitarian hat and puts on a Platonic one
instead. What really moves him is humanity’s promise for achievement—for
exploring the entire cosmos and suffusing it with value. If we and our
potential descendants are the only rational beings in the universe—a
distinct possibility, so far as we know—then, he writes, “responsibility
for the history of the universe is entirely on us.” Once we have reduced
our existential risks enough to back off from the acute danger we’re
currently in—the Precipice— he encourages us to undertake what he calls
“the Long Reflection” on what is the best kind of future for humanity: a
reflection that, he hopes, will “deliver a verdict that stands the test
of eternity.” Ord’s is a very moralizing case for why we should care
about humanity’s future. It cites values—both utilitarian happiness and
Platonic ideal goods— that might be realized many eons from now, long
after we and our immediate descendants are dead. And since values do not
diminish because of remoteness in time, we are obligated to take those
remote values seriously in our current decision-making. We must not
“discount” them just because they lie far over the temporal horizon.
That is why the future of humanity weighs so heavily on us today, and
why we should make the safeguarding of that future our greatest duty,
elevating it in importance above all nonexistential threats—such as
world poverty or climate change. Though Ord does not explicitly say
that, it is the conclusion to which his reasoning seems to commit him.
As a corrective, let’s try to take a nonmoralizing view of the matter.
Let’s consider reasons for caring about humanity’s future that do not
depend on value-based considerations, whether of happiness or ideal
goods. How would our lives today change if we knew that humanity was
doomed to imminent extinction—say, a century from now? That is precisely
the question that the philosopher Samuel Scheffler posed in his 2012
Tanner Lectures at Berkeley, later published in his book Death and the
Afterlife.6 Suppose we discovered that the world was guaranteed to be
wiped out in a hundred years’ time by a nearby supernova. Or suppose
that the whole human race was suddenly rendered infertile, so that no
new babies could be born.7 How would the certain prospect of humanity’s
absolute extinction, not long after your own personal extinction, make
you feel?
It would be “profoundly depressing”—so, at least, Scheffler plausibly
maintains. And the reason is that the meaning and value of our own lives
depend on their being situated in an ongoing flow of generations.
Humanity’s extinction soon after we ourselves are gone would render our
lives today in great measure pointless. Whether you are searching for a
cure for cancer, or pursuing a scholarly or artistic project, or engaged
in establishing more just institutions, a threat to the future of
humanity is also a threat to the significance of what you do. True,
there are some aspects of our lives—friendship, sensual pleasures,
games—that would retain their value even in an imminent doomsday
scenario. But our long-term, goal-oriented projects would be robbed of
their point. “Most of the time, we don’t think much about the question
of humanity’s survival one way or the other,” Scheffler observes:
6Reviewed in these pages by Thomas Nagel, January 9, 2014.
7Something of the sort threatens to happen in the P.D. James novel
Children of Men (Faber and Faber, 1992).
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.tuxtown.net/pipermail/d66/attachments/20210211/c4563e1c/attachment-0001.html>
More information about the D66
mailing list