Subtitles section Play video
Chris Anderson: Nick Bostrom.
So, you have already given us so many crazy ideas out there.
I think a couple of decades ago,
you made the case that we might all be living in a simulation,
or perhaps probably were.
More recently,
you've painted the most vivid examples of how artificial general intelligence
could go horribly wrong.
And now this year,
you're about to publish
a paper that presents something called the vulnerable world hypothesis.
And our job this evening is to give the illustrated guide to that.
So let's do that.
What is that hypothesis?
Nick Bostrom: It's trying to think about
a sort of structural feature of the current human condition.
You like the urn metaphor,
so I'm going to use that to explain it.
So picture a big urn filled with balls
representing ideas, methods, possible technologies.
You can think of the history of human creativity
as the process of reaching into this urn and pulling out one ball after another,
and the net effect so far has been hugely beneficial, right?
We've extracted a great many white balls,
some various shades of gray, mixed blessings.
We haven't so far pulled out the black ball --
a technology that invariably destroys the civilization that discovers it.
So the paper tries to think about what could such a black ball be.
CA: So you define that ball
as one that would inevitably bring about civilizational destruction.
NB: Unless we exit what I call the semi-anarchic default condition.
But sort of, by default.
CA: So, you make the case compelling
by showing some sort of counterexamples
where you believe that so far we've actually got lucky,
that we might have pulled out that death ball
without even knowing it.
So there's this quote, what's this quote?
NB: Well, I guess it's just meant to illustrate
the difficulty of foreseeing
what basic discoveries will lead to.
We just don't have that capability.
Because we have become quite good at pulling out balls,
but we don't really have the ability to put the ball back into the urn, right.
We can invent, but we can't un-invent.
So our strategy, such as it is,
is to hope that there is no black ball in the urn.
CA: So once it's out, it's out, and you can't put it back in,
and you think we've been lucky.
So talk through a couple of these examples.
You talk about different types of vulnerability.
NB: So the easiest type to understand
is a technology that just makes it very easy
to cause massive amounts of destruction.
Synthetic biology might be a fecund source of that kind of black ball,
but many other possible things we could --
think of geoengineering, really great, right?
We could combat global warming,
but you don't want it to get too easy either,
you don't want any random person and his grandmother
to have the ability to radically alter the earth's climate.
Or maybe lethal autonomous drones,
massed-produced, mosquito-sized killer bot swarms.
Nanotechnology, artificial general intelligence.
CA: You argue in the paper
that it's a matter of luck that when we discovered
that nuclear power could create a bomb,
it might have been the case
that you could have created a bomb
with much easier resources, accessible to anyone.
NB: Yeah, so think back to the 1930s
where for the first time we make some breakthroughs in nuclear physics,
some genius figures out that it's possible to create a nuclear chain reaction
and then realizes that this could lead to the bomb.
And we do some more work,
it turns out that what you require to make a nuclear bomb
is highly enriched uranium or plutonium,
which are very difficult materials to get.
You need ultracentrifuges,
you need reactors, like, massive amounts of energy.
But suppose it had turned out instead
there had been an easy way to unlock the energy of the atom.
That maybe by baking sand in the microwave oven
or something like that
you could have created a nuclear detonation.
So we know that that's physically impossible.
But before you did the relevant physics
how could you have known how it would turn out?
CA: Although, couldn't you argue
that for life to evolve on Earth
that implied sort of stable environment,
that if it was possible to create massive nuclear reactions relatively easy,
the Earth would never have been stable,
that we wouldn't be here at all.
NB: Yeah, unless there were something that is easy to do on purpose
but that wouldn't happen by random chance.
So, like things we can easily do,
we can stack 10 blocks on top of one another,
but in nature, you're not going to find, like, a stack of 10 blocks.
CA: OK, so this is probably the one
that many of us worry about most,
and yes, synthetic biology is perhaps the quickest route
that we can foresee in our near future to get us here.
NB: Yeah, and so think about what that would have meant
if, say, anybody by working in their kitchen for an afternoon
could destroy a city.
It's hard to see how modern civilization as we know it
could have survived that.
Because in any population of a million people,
there will always be some who would, for whatever reason,
choose to use that destructive power.
So if that apocalyptic residual
would choose to destroy a city, or worse,
then cities would get destroyed.
CA: So here's another type of vulnerability.
Talk about this.
NB: Yeah, so in addition to these kind of obvious types of black balls
that would just make it possible to blow up a lot of things,
other types would act by creating bad incentives
for humans to do things that are harmful.
So, the Type-2a, we might call it that,
is to think about some technology that incentivizes great powers
to use their massive amounts of force to create destruction.
So, nuclear weapons were actually very close to this, right?
What we did, we spent over 10 trillion dollars
to build 70,000 nuclear warheads
and put them on hair-trigger alert.
And there were several times during the Cold War
we almost blew each other up.
It's not because a lot of people felt this would be a great idea,
let's all spend 10 trillion dollars to blow ourselves up,
but the incentives were such that we were finding ourselves --
this could have been worse.
Imagine if there had been a safe first strike.
Then it might have been very tricky,
in a crisis situation,
to refrain from launching all their nuclear missiles.
If nothing else, because you would fear that the other side might do it.
CA: Right, mutual assured destruction
kept the Cold War relatively stable,
without that, we might not be here now.
NB: It could have been more unstable than it was.
And there could be other properties of technology.
It could have been harder to have arms treaties,
if instead of nuclear weapons
there had been some smaller thing or something less distinctive.
CA: And as well as bad incentives for powerful actors,
you also worry about bad incentives for all of us, in Type-2b here.
NB: Yeah, so, here we might take the case of global warming.
There are a lot of little conveniences
that cause each one of us to do things
that individually have no significant effect, right?
But if billions of people do it,
cumulatively, it has a damaging effect.
Now, global warming could have been a lot worse than it is.
So we have the climate sensitivity parameter, right.
It's a parameter that says how much warmer does it get
if you emit a certain amount of greenhouse gases.
But, suppose that it had been the case
that with the amount of greenhouse gases we emitted,
instead of the temperature rising by, say,
between three and 4.5 degrees by 2100,
suppose it had been 15 degrees or 20 degrees.
Like, then we might have been in a very bad situation.
Or suppose that renewable energy had just been a lot harder to do.
Or that there had been more fossil fuels in the ground.
CA: Couldn't you argue that if in that case of --
if what we are doing today
had resulted in 10 degrees difference in the time period that we could see,
actually humanity would have got off its ass and done something about it.
We're stupid, but we're not maybe that stupid.
Or maybe we are.
NB: I wouldn't bet