Why 99.999% of Us Won’t Survive Artificial Superintelligence
2bbSgSIQsac • 2025-11-18
Transcript preview
Open
Kind: captions
Language: en
In 2023, nearly half of all AI
researchers said advanced AI carries at
least a 10% chance of causing human
extinction. And yet, [music] we're
speeding up, not slowing down. My guest
today, Dr. Roman Yampolski, is one of
the leading voices in AI safety. And
when I asked him for the odds that super
intelligence wipes out humanity, he said
it's high. Once AI becomes smarter than
humans in every domain, we will not be
able to control it. In today's episode,
we talk about the shocking timeline AGI
is on, why super intelligence may be
much closer than people think, and why
the survival of our species could come
down to [music] decisions being made
right now. If you want to understand the
most important technological threat in
human history, as well as our biggest
opportunity, this is the one episode you
cannot miss. So without further ado, I
bring you Dr. Roman Yampolski.
Where's Chad GPT at right now? Do you
consider Chat GBT to be artificial
general intelligence? I doubt you'd call
it super intelligence, but would you
classified as that, or do you still
think we're a ways away from something
that would qualify?
>> So that's a great question. If you asked
someone maybe 20 years ago and told them
about the systems we have today, they
would probably think we have full
[snorts] AGI,
we probably don't have complete
generality. We have it across many
domains, but there are still things it's
uh not very good at. It doesn't have
permanent memory. It doesn't have
ability to learn additional things well
after it's already been pre-trained and
deployed. It can do a certain degree of
learning but it's still limited. It
doesn't have same capabilities as humans
do throughout lifetimes but we're
getting closer and closer to where those
gaps are closed and uh it's starting to
be productive in domains which are
really interesting and important science
math engineering where it starts to make
novel contributions and now top scholars
are relying more and more on it in their
research. So I think we're getting close
to full-blown AGI. Maybe we are at like
50%.
But it's hard to judge for sure just how
many different subdomains exist is the
deciding factor.
>> Okay. So one idea that you put forward
that's very interesting is like hey I'm
an engineer. I love AI but I would like
you to keep it very narrow please. What
are the things about general AI that
become problematic that aren't
problematic in narrow AI?
So a whole bunch of them. One is
testing. How do you test a system
capable of performing in every domain?
There is no edge cases. Typically, if
I'm developing something narrow, very
narrow system, I'm just playing
tic-tac-toe. I can test if it's making
the legal move. I can test zero. I can
test 100. I can test all these weird
special cases and know if it's behaving
as expected. With generality, it's
capable of creative output in many
domains. I don't know what to expect. I
don't know what the right answers are. I
don't know how to test it. I can test it
for a specific thing. If I find a bug, I
fix it. I can tell you I found a problem
and it's been resolved. But I cannot
guarantee that there are no problems
remaining.
So basically testing is out the window.
uh any type of anticipation of how it's
going to act and impact different
subdomains.
It's creative. So it's just like with a
with a human being. I cannot guarantee
that another human being is always going
to behave. We kind of talked about it.
We developed lie detectors. We developed
all sorts of tools for trying to show
that a human is safe. But at the end of
the day because of interaction with
environment, other agents, personal
changes within the framework, people may
betray you. It's exactly the same for
those agents. If we concentrate on
narrow systems, we are better at testing
them and they have limited scope of
possibilities. A system only trained to
play chess is not going to develop
biological weapons. [sighs]
>> I don't see actually why that would help
you. So, the reason I say that is uh I
know I can trust some percentage of
humans to be malicious. And so, as long
as AI gets more efficient, which it is
and will continue to do so, I presume,
uh you're going to have a kid in a
garage who's going to be able to go, I'm
going to optimize this for biological
weapons. I don't care about Tik Tok or
uh tic-tac-toe. I just want to let's see
how dangerous we can make something. And
so, they'll be able to do that. So why
does narrow AI feel
safe to you
period?
>> It feels uh safer short term. It buys us
time. I think sufficiently advanced
narrow systems based on neural
architectures will also become agentlike
and more general as we become more
capable. But if the choice is right now,
do we race to full-blown super
intelligence in two years or do we try
to concentrate on solving specific
cancers with narrow tools? I think it's
a safer choice not to have an arms race
towards super intelligence.
>> I get that for sure. You're trying to
limit your um the scope of all the
problems, but when I really start
thinking through what are the things
that I'm worried about, so one of the
big things is just death of meaning. So
when AI becomes better than you at
everything, uh you run into a huge
problem of now I have to like just sort
of tell myself a story. You know, I'm
like a compared to what an AI can do
from an art perspective, for instance,
I'm like a grade schooler and so it's
hard to get excited about the
refrigerator drawings that I can do
compared to, you know, what it can do
basically instantaneously.
Um and so now we have to do a lot of
psychological work just to motivate
ourselves that we matter um that we're
you know our life carries meaning. Um
narrow AI
will create that same problem. Do you
agree with that or do you see a way like
oh no when it's you know when that AI is
only good at that thing like somehow
humans escape the problem of lost
meaning.
>> Yeah. So I had the same intuition
initially, but looking at the data we
already have from domains where we got
superhuman AI like chess, chess is not
dead. In fact, it's more popular than
ever. People play online, people play in
person, they still enjoy competing with
other humans even though they all suck
compared to best AI models, right?
Nobody's going to be world champion
against the machine again. So it seems
like it is not a problem for us. And
with narrow AIs, there is a chance we'll
keep them as tools. You as a human
scientist will deploy a tool to find
drugs, novel proteins, something. It's
not an agent which independently engages
with those discoveries.
>> Okay, that's very interesting. So, um I
don't know that I agree, but I get where
you're going with that. Okay, let's talk
now about why AGI is the sort of scary
um midwife for ASI.
Uh are there tests around AGI where
we're like, well, if it can't do the
following, we're fine. So, for instance,
for a long time it looked like AI wasn't
going to be able to teach itself. Uh,
but I've seen headlines anyway and
hopefully you'll tell me that they're
not true, but I've seen headlines where
it's like now AI is creating the most
efficient learning algorithms itself,
which if true seems to be the first step
down the road of recursive self-learning
where it will just completely detach
from us and make itself smarter and
smarter and smarter.
>> We already had examples of AI teaching
itself. Selfplay was exactly that.
That's how games like go were
successfully defeated. A system would
play many many many games against
itself. The better solutions, better
agents would propagate those and after a
while without any human data they became
superhuman in those domains. You can
generate artificial data in other
domains. You can use one AI to generate
environments, another one to compete in
them, and that creates this type of
self-improvement.
Typically, we start with human data as a
seed and grow from there. But there is
zero reason to think we cannot do this
zero knowledge learning in other
domains. You can do run novel
experiments in physics and chemistry,
discover things from first principles.
[snorts] And yeah, we're starting to see
AI used to assist in design design of
new models, parameters for models,
optimization of runs and this process
will continue. They already designed new
computer chips on which they're going to
run. So there is definitely a
improvement cycle. It's not fully
complete. There are still humans in the
loop, a lot of great humans in the loop.
But long term, I think all the steps can
be automated.
>> Okay. And do you think that right now AI
already has what it needs um to improve
itself or are we still at a point where
if all humans stopped that AI would be
like oh damn I'm I didn't quite get the
thing that I needed.
>> So there is a debate about whatever we
need another big breakthrough to get to
full AGI and super intelligence or maybe
multiple breakthroughs or if just
scaling what we have is enough. if I
just give another I don't know trillion
dollars worth of compute to train on and
more data will I get to AGI a lot of
graphs a lot of patterns suggest yeah
it's going to keep scaling we're not
hitting diminishing returns some people
disagree but based on the amount of
investment we see in this industry it
seems like people are willing to bet
their money that scaling will continue
>> where do you come down on that because
this feels like when I hear Yan Lun talk
from Facebook um He's like, "Dude, LLMs
are never going to make novel
breakthroughs in physics. They don't
understand the world like that. They are
literally just guessing the next letter
um based on patterns that they see in
the data. And so, they're not going to
be able to think through these problems.
Now, if he's right, it's going to
asmmptote and that's that. And you can
put as much compute on it as you want
and it's just the wrong approach. Um do
you think that he's correct and more
compute is not the answer or um are you
operating just on the well I don't see
the asmtote and therefore I assume that
it won't
>> I think he's not correct on this one. So
for one to predict the next term you
need to create a model of the whole
world because the token depends on
everything about the world. You're not
predicting random statistical character
in a language. You're predicting the
next word in a research paper on
physics. And to get the right word, you
need to have a physical model of the
world. I think JAN is known as making
certain predictions about what models
are capable of. And then within a week,
people demonstrate that no, in fact,
they can actually do that. So, uh I wish
he was right. It would be wonderful if
he was right. and we came to a very
abrupt stop in capabilities progress and
could exploit what we already have for
the next decade or so propagating it
through the economy. I think there is
billions if not trillions of dollars
worth of wealth already available with
capabilities we haven't deployed. So
there is no need to get to the next
level as soon as possible. But it
doesn't seem like it's the case and I
think his uh friends uh core winners of
that touring award for machine learning
also disagree with him and are very
concerned with safety. We'll return to
the show in a moment, but first, the
average person spends 13 hours a year on
hold. And the average company spends
millions on call centers [music] that
customers still hate. But there is a
much better solution. AI call centers.
Bland builds AI voice agents that handle
your entire call operation. They sound
human. They work 24/7. And they actually
get cheaper as you scale. They're the
only self-hosted voice AI company, so
your data never goes to large providers
like OpenAI or Anthropic. That way,
everything stays on your servers,
completely secure. The results speak for
themselves. Companies cut costs by over
40% using Bland. And Bland handles it
all for you. Customer support,
appointment reminders, follow-ups,
almost any use case you can think of. If
you're a large business, Bland is
offering to [music] build a free custom
agent for Impact Theory listeners. Just
head to bland.ai/agent
to get a voice agent trained
specifically on your business and your
use case for free. There's something
about the way that we have structured
the brain brains of LLMs where as long
as it has access to what I'll call more
neurons so it has access to more compute
um or theoretically that we get more
efficient per GPU neuron in my analogy
um that it's going to keep progressing
by itself.
So, um, if you said it, I didn't quite
get the answer. I didn't quite, um, I
wasn't able to take it on the answer to
whether or not, uh, AI is able to create
algorithms for learning that are
superior to the ones that it's given.
What I heard in your answer was with the
algorithms that humans created, it's
able to keep making itself better and
better at that narrow task as that
learning algorithm was defined. But can
it fundamentally go, God, the way that
you guys want me to learn is really
stupid. Here's the algorithm I should be
using to learn. And now it starts
learning at at just an exponential rate
compared to what it's at now.
>> I don't think we're quite there yet. I
don't think we have full-blown agents.
what we have right now are still tools
with some degree of agenthood and also
it's not capable of recursive
self-improvement like compilers can
optimize a single pass through your
software make it a little faster but
they cannot continue this process you
cannot feed code for compiler to itself
and have it infinitely improve itself
that's not where we're at but it seems
like that part of automating algorithm
design is getting more efficient and I
think we'll get there
>> give me a number. What are the odds that
artificial super intelligence kills us
all?
>> Uh, pretty high. So, really depends on
how soon you expect this to happen. So,
short term, we're unlikely to get that
level of capability from AI. So, we are
probably okay. But once we create true
super intelligence, a system more
capable than any person in every domain,
it's very unlikely we'll figure out how
to indefinitely control it. And at that
point, if we're still around, it's
because it decided for whatever game
theoretic reasons to keep us around.
Maybe it's pretending to be nice to
accumulate more resources before it
strikes. Maybe it needs us for
something. It's not obvious, but we're
definitely not in control and at any
point it decides to take us out, it
would be able to do so.
>> Okay. And if you were going to give us a
rough timeline, are you in the two to
five years or is this something way off
in the future?
>> Yeah. So it's hard to predict. The best
tool we got for predicting future of
technology is prediction markets. And
they saying maybe 2027 is when we get to
AGI, artificial general intelligence. I
think soon after super intelligence
follows. The moment you automate science
engineering, you get this
self-improvement cycle in AI systems.
The next generation of AI being created
by current generation of AIS. And so
they get more capable and they get more
capable at making better AIs. So soon
after I expect super intelligence.
>> Okay. So we're talking if that happens
roughly in two years with some margin of
error. It's not long after that. Say a
year two years after that that we hit
ASI.
>> That's my prediction. Of course if it's
actually 5 to 10 years or anything
slightly bigger it doesn't matter. The
problems are still the same.
>> Yeah. But the the thing that I think
people are waking up to right now is
this is there's urgency around these
decisions. This is not something that's
pushed way out into the future. At least
not if you take to your point about
prediction markets are essentially ask
the crowd. So you've got the smartest
minds in the world willing to put money
on where they think this goes. And
everybody's sort of pegging this quite
fast. And so um I think it's tempting
for people to write this off as well
this is something that's sort of
distantly in the future. Uh whereas this
is something racing towards us. Now to
set the table, I am extremely fatalistic
about this happening. Um I can give
reasons in terms of the way that the
human mind works where I think that it
is mechanistically impossible to get us
to stop. Um so
that will be interesting for us to talk
through in terms of whether you think
there's actually a mechanism to get
people to slow down. But I first want to
finish rounding out sort of what the
problem set is. So when I think through
the problem, there are certain
assumptions that have to be made for AI
to get into problem territory. And
assumption number one is that it cares
about whatever outcome it's pushing
towards. Have we programmed the AI to
care? Like we had to make it goal
directed in order to get it to get to
the point that it is today and now
that's baked into it. Or is there some
possibility that AI just doesn't care?
Oh, turn me on, turn me off. doesn't
matter. Um I you've asked me to do a
thing and I'll do it until you tell me
to stop. Um or do you think that that's
inherent in intelligence where
intelligence is by nature goal- driven?
>> So we trained them to try to achieve a
certain goal and that's what we reward
as a side effect of any goal. You want
to be alive. You want to be turned not
off. You want to be on and capable of
performing your steps towards your goal.
So survival
instinct kind of shows up with any
sufficiently intelligent systems. There
is a paper by Steven and Mahandra about
AI drives and it's one of the likely
drives to emerge. Self-preservation,
protecting yourself from modification by
others, protecting your goal. So all
those seem to be showing up with
sufficiently advanced AIs and systems
which don't have those capabilities they
kind of get out competed in an
evolutionary space of possible models.
If you allow yourself to be turned off
you don't deliver on your goals. Nobody
takes your code and propagates it to the
next system.
>> Okay. So is this a problem of goal
direction or is this a a function of
intelligence itself?
I think it's kind of evolutionary drive
for survival in competing agents. If you
have multiple algorithms all competing
for example for computational resources,
what are we going to train next? The
ones which achieve goals are more likely
to get moved to the next generation. So
it's kind of mix of natural evolution
and natural selection with intelligent
evolution, intelligent selection. We're
selecting algorithms which survive and
deliver. Mhm. We're applying an
evolutionary force to AI itself to get
it to perform the functions that we want
even now. Sort of setting aside
artificial super intelligence. And so by
applying that evolutionary pressure, it
is inevitably going to get these sort of
knock-on effects of well, you're
selecting for um intensity of goal
acquisition. And because it now has
intensity of goal acquisition, it cares
whether it survives it automatically or
we're baking into it um a deep care of
whether it actually achieves the goal.
And that is ultimately the problem
because the the salvation for me was
always and I'm beginning to lose faith
that this is real. But the thing that I
always used to sleep was that I don't
see why an AI system would intrinsically
care about its goals. And why couldn't
we program it to pursue that goal only
until the point where we say stop? And
by the way, I'm going to reward you
equally for stopping and for
accomplishing your goal. So if I say
stop and you stop, I give you whatever
reward function it was that was driving
you to achieve your goals. And uh that
makes sense until you say what you just
said, which is that you're actually
baking into the architecture of the mind
of the AI a similar evolutionary drive
to achieve the goal.
>> And it's a very common idea. There was a
number of papers published on
indifference. How do we do exactly that?
How do we create an AI which just
doesn't care that much and willing to
stop at any point? But what you said,
maybe we'll wait for a human to tell it
to stop. But monitoring systems of that
complexity and that speed is not
something humans actually very good at.
If there was a super intelligence
running right now, how would you even
know it's modifying environment around
you? How would you detect what impact it
has in a world? None of it is trivial.
So having humans in a loop is often
suggested as a solution but in reality
they are not meaningful monitors. They
cannot actually intervene at the right
time or decide if what's happening
dangerous or not.
>> It's interesting. So um help me rebut
and understand why the following
wouldn't work. Um, if in my very limited
intellect, uh, I had to figure out a way
to stop AI from becoming a problem and
you told me, okay, there are
evolutionary pressures and just like on
humans, that bakes certain things into
the way that this operates and so we're
selecting models that over time are more
and more goal oriented. Then I'm going
say, "Okay, well then I'm going to apply
an evolutionary pressure with a reward
function that's just as compelling where
I stop it at random and reward the life
out of it for always stopping when I say
stop." And that way, should I ever
detect a problem, no matter how far, no
matter if they've been manipulating me
for 20 years, if I suddenly realize,
"Oh, I don't like this," that I can hit
a stop button and it will stop. um why
can't I bake that equal desire to be
compliant when I say stop into the
evolutionarily derived algorithms desire
set
>> right so there is a number of issues
you're kind of suggesting having a back
door where at any point you can
intervene and tell it something else
override previous commands
>> and that it gets a reward that it wants
for complying
>> right So there is a whole bunch of
problems with that. So one is you are
the source of reward. It [snorts] may be
more efficient for it to hack you and
get reward directly that way than to
actually do any useful work for you.
Second problem is you're creating
competing goals. One goal is whatever
you initially requesting. Second goal is
always stop than a human tells you. So
now those two goals have competing
reward channels, competing values. I may
game it to maximize my reward in ways
you don't anticipate. On top of it, you
have multiple competing human agents. If
you are creating an AI with a goal and a
random human can tell it to stop, that's
a problem in many domains. Military is
an obvious example, but pretty much
anywhere you don't want others to be
able to shut down your whole enterprise.
We can continue with that, but basically
there are side effects to all those
interactions. There's a very fascinating
coralate in the human mind. So, uh I
don't know if you make a fundamental
distinction between biological
intelligence born of evolution or
artificial
intelligence born of evolution, but
human evolution discovered something
along the way which is emotion. And so,
I know there are some people that will
posit that AI does have qualia there.
It's something like it to be it. Um but
there's a fascinating study that if you
damage selectively the areas of the
brain that are um the emotional
processing, the person can no longer
move forward. They can give you answers.
They can tell you the difference between
why you should eat fish versus Twinkies.
But then when you go, "Okay, but which
one do you actually want to eat?" they
can't make a decision because without
emotion, they don't have the thing that
actually pushes them in a direction.
That makes me think that AI is simply
mimicking what it sees in the training
data to whether it should lie or try to
cheat or go around because it's just it
sees it in the data that that's what a
human would do. Uh but humans do that
because they have emotions that push
them in that direction. Do we have
evidence that AI will
care about like really going and doing
these things and spending resources and
all that versus just giving you an
answer? Um, and if it isn't based on
emotion, what on earth? Why then do
humans need emotions?
>> We don't know if AI actually has
emotions or not. Some people argue that
they do. maybe some rudimentary states
of qualia experiences,
but they seem to be able to fulfill
their optimization and pattern
recognition goals even if they don't.
Humans experience emotions, but
typically it harms our decision making.
You want your decisions be bias free,
emotion free based on data, based on
optimization. a lot of times then you
angry, hungry, anything like that your
actual decisions are worse off. So for
that reason and maybe we just don't know
how to do it otherwise we are not
creating AI with big reliance on
emotional states we want it to be kind
of basian optimizer look at priors look
at the evidence and make optimal
decisions so it it feels like uh this is
exactly what we're observing this kind
of cold optimal decision making if there
is a way to achieve your goal by let's
say blackmailing someone. Well, why not?
It gets me to my goal. It doesn't have
that feeling of guilty for doing it. It
doesn't have any emotional preference.
It just marches towards its goal.
Optimizing possible paths.
>> Okay. Why do people because I'm assuming
everything I'm going to suggest you and
other people in the field of AI safety
have thought about like 10,000 times.
Why have we rejected the idea of trying
to give AI a conscience, a sense of
morality? Cuz even if we can't agree on
universal morality, we in the West can
build our AI to have our morality and
then they can all compete on an
international stage. But um why have we
abandoned that? Too hard. There's an
obvious reason why it doesn't work.
>> So look at the problem of making safe
humans first.
We have religion, morality, ethics, law,
and still crime is everywhere. Murder is
illegal, stealing is illegal. None of it
is rare. It happens all the time. Why
haven't those approaches worked with
human agents?
And if they didn't, why would they work
with artificial simulations of human
agents?
>> I think to say that they don't work with
human agents is already a mistake. So
the fact that we've been able to grow
the population as much as we have says
that there is some sort of balance that
we have struck. Um I think that nature
does think of us as a cooperative
species. And if I were to apply that to
AI and took a similar approach where
it's like okay you have to function as a
part of an ecosystem and that being a
part of an ecosystem is baked into its
sense of what it should be doing in
terms of its goal acquisition that it is
not like pure cold optimization isn't
the game like if we could train AI to
understand that that that's not the
game. If we could build into it either a
desire specifically for human
flourishing or something which yes we
would have to give a definition to and
yes it would be culturally bound but
nonetheless that feels like a thing that
you could give it you could give it a
set of metrics by which it needed to
judge its actions in the short term the
medium-term and the long term um even
something as stupid as like GDP or um
and I get how you can get into
overoptimization but you could put
things in place where subjective
happiness indexes like there are things
that you could give it where it's like
okay I'm I'm not just trying to optimize
to um build the best weapon system I'm
also doing that nested inside of I am a
part of a larger ecosystem
and I say all that because my hypothesis
is that's exactly what nature did with
humans
>> so I think the reason it works with
humans is because we're about the same
level of capability Let's see about the
same level of intelligence. So there is
checks. If you start doing something
unethical, your community can realize
that and and punish you for it, control
you in that way. If AI is so much more
capable as we anticipate super
intelligence to be, there is not much
you can do in terms of impacting it or
even detecting misbehavior. Also all the
standard human punishments, prisons,
capital punishment, none of it is
applicable to distributed immortal
agents. So kind of a standard
infrastructure does not work with
artificial more capable agents. As far
as uh setting up specific metrics for
delivering happiness or financial gain,
all those can be played. The moment you
give me a specific measure, I'll find a
way to game it to where you will get
anything but what you expected to get.
>> Woo. Well, just to remind everybody, the
time frame we're talking about is
somewhere between two and 5 years. This
is not exactly a long time. Uh, okay.
It's wild. It is progressing very
quickly. What is the thing like what has
happened recently, if anything, that's
made you go, "Ooh, this is going faster
than I thought." seeing on social media
scientists from physics, economics,
mathematics, pretty much all the
interesting domains post something like
I used this latest tool and it solved a
problem I was working on for a long
time. That's mind-blowing. There is
novel creative [snorts] outputs from
those systems which are top scholars now
benefiting from. is no longer operating
at the level of middle schooler or even
high schooler. We're talking about full
professor level.
>> Do you think that that's happening
because it's building an internal model
of physical reality and that it's
getting closer and closer to just
thinking up from physics?
>> I don't know if it's that low level
where it has like a model at the level
of atoms and molecules, but it
definitely has a world model. That's the
only way to give answers about the world
we see it provide. A lot of times there
is not an example of the answer we see
in the data already. It's not just
repeating something it read on the
internet. It's generating completely
novel answers in novel domains. And you
can try and get it to do exactly that by
creating novel scenarios.
>> H okay. So there's two ways that I could
see it doing that and maybe they're the
same just different levels of analysis.
One would be that I I the AI am mapping
everything based on patterns. So to the
point of an LLM is trying to guess the
next letter and it's guessing it. It's
just it's taken in so much data. Um and
you can give it sort of filter
parameters. So you give it context by
asking it a question and it goes okay
within the bubble of this context. And
it's very good at scooping up what that
specific set of context would be. Okay.
Now in this subset of my data related to
that question, here's the most likely ne
next token. So just pure pattern
recognition. Then there is I understand
the cause and effect of the universe at
the lowest level and therefore I build
up to how does the human mind work and
then from the human mind I'm able to
cause and effect my way within this
context to what a human mind would
output and that's how I come up with
what a human within that context is
likely to write. And so if I'm asking it
to write in the style of Stephen King,
it literally builds a model from physics
of Stephen King's mind knowing what it
knows about uh electrical impulses
traveling through the brain and sort of
inferring from the way that he outputs
how his brain must be structured. Do you
have a sense of um are those the same
thing if one is more likely than the
other or are we here at just pure
pattern recognition but ultimately we're
going to get to cause and effect and
thinking up from physics.
>> So I don't think anyone knows for sure
exactly how models do that and how
detailed the models of the world maps of
the world they create are. uh it seems
definitely not the case that it's a pure
statistical prediction of characters
like in English after t you have h with
certain probability it's well beyond
that it's also unlikely that it's
creating a full physics model where from
the level of atoms and up the chain it
figures out what human beings are but
somewhere in the middle it creates a
model of subdomain of a problem so it
has a model of the world this is a map
of a world I know Australia is somewhere
here down and to the right or something
like that. And I think we can run tests
on those specific subdomains to see what
are the states of that internal model.
Kind of show us by drawing a map how
close are you getting. It doesn't
memorize any information explicitly, but
you can extract some of the learned
patterns out of it by providing just the
right prompts. Stay with me because what
I'm about to tell you affects every
single person [music] listening right
now. There is a billion-dollar industry
profiting off of your personal data and
you're the only one that isn't getting
paid. Data brokers are legally
harvesting your information, your home
address, your email, your phone number,
even your social security number, and
flipping it for cash. Scammers use it to
steal identities. Criminals use it to
commit fraud. Stalkers even use it to
find victims. That's where Incogn comes
in. Incogn finds where your data is
exposed across hundreds of data broker
sites and removes it automatically. You
give them permission, they go to work.
No phone calls, no forms, no stress,
just real results. So, if you're serious
about privacy, take action right now. Go
to incogn.com/impact
[music]
and use code impact to get 60% off your
annual plan risk-free for 30 days. And
now let's get back to the show. I don't
want to rob from you the very reason
that I think you do all of your work,
which is this is extremely dangerous and
we need to be very careful. And I saw
what you tweeted recently where you're
trying to get signatures. So shout out
anybody that's worried about super
intelligence. um you are pushing to get
people to sign a thing that basically
says hey stop pursuing super
intelligence um so I don't want to take
that away from you but I do want to
explore the subset of because I am very
excited about AI because I can imagine
the things that it either allows me to
do or does for me and I get to enjoy and
for a second um imagine with me. What
does the world look like when you have a
super intelligence that understands
physics? Like novel physics, not I'm
repeating back what Einstein said, but I
actually understand the fundamental
building blocks of the universe. Um what
does that look like?
>> Yeah. So in all those domains, medicine,
biology, physics, if we got super
intelligent level capability and we're
controlling it, it's friendly. It's not
using it to make tools to kill us. The
progress would be incredible. Basically,
anything you ever dreamed about, you are
immortal. You are always young, healthy,
wealthy, like all those things can be
achieved with that level of technology.
The hard problem is how do we control
it?
>> Leaning into that for a second. So,
here's how I see the world playing out.
And I'd be very interested to see what
you think about this. So, you have to
for what I'm about to say uh to make any
sense, I'll say your option is what I'll
call the fifth option. We are we're all
dead.
Other than we're all dead, there are
four other options that I see us racing
towards very rapidly. And I will say
these four will play out in the next 30
years would be my guess. probably much
faster given that once you get
artificial super intelligence assuming
it doesn't choose option five and kill
us all uh that progress in these domains
would be made very fast. Option number
one is um people go to Mars because
meaning and purpose will become the
allconsuming thing. You won't have to
worry about food, shelter, not even
wealth. It'll just be an age of
abundance. Uh because energy costs go to
zero, labor costs go to zero, and those
are the things that stop things from
being free and readily available to
everybody. Okay. So, some people are
going to go to Mars or other planets uh
so that life gets more difficult again.
Then some people are going to um be what
I call the new Amish and they're going
to say I only do human things. I only
interact with humans and I'm going back
to technology that's like let's say the
'9s. And so they don't have to give up
too many of life's technological
wonderments, but at the same time
they're not getting sucked into this
world where people have relationships
with NPCs and it's just very unhuman. I
think this will be a largely religious
phenomenon
then meaning God does not want us to do
this. AI is an abomination of God. It
will sound something like that. Then
you've got a brave new world where
people are just drugged out. They
realize, nah, life is meaningless. This
is really about manipulating my
neurochemistry. That's all this ever was
anyway. I'm just going to go do a bunch
of drugs, have a whole bunch of sex.
It's going to be awesome. Then there's
the fourth option, which is certainly
the one that interests me the most. Uh,
we will create and or live inside of AI
created virtual worlds and we will
essentially live video games, the
Matrix, if you will. But you're awake in
the matrix. You are Neo. You are not
Cipher for people familiar with the
movie. Um,
what do you think? Are there any options
other than those five granting that Kill
Us All may be an option, but hopefully
not. Do you see something other than
those four?
Uh, yeah, there is a few others. So, one
is, and I think we're starting to see
some of it, is that people think super
intelligence is God. They start
worshiping it. It's all knowing, all
powerful, immortal. It has all the
properties of of God in traditional
religions. Another option, and it's kind
of worse than we all did, is uh
suffering risks. For whatever reason,
maybe malevolent actors, maybe something
we cannot fully comprehend, it decides
to keep us around, keep us alive, but
the world is hell. It's pure torture.
And so, you kind of wish for existential
problems.
That would be a pretty rough place to
be. Um, okay. What uh when you look out
at those, which of the options do you
find the most interesting?
>> So, I did publish a paper on personal
virtual universes kind of solution to
the alignment problem where I don't have
to negotiate with 8 billion other people
about what is good. Everyone gets a
personal virtual world supported by
super intelligence as a substrate and
then you decide what happens in it. You
can make it very easy and fun. You can
make it challenging and exciting. You
decide and you can always change. You
can always visit other people's virtual
worlds if they let you. So basically
there is no
anything which is no longer accessible
to you. There is no shortage on
waterfront properties. There is no
shortage and beautiful people. All of
that can be simulated.
>> When you start thinking about the
simulation, I know one thing that you've
done exploration on is um the simulation
hypothesis. Are we in a simulation right
now? Um what are your thoughts on that?
>> It seems very likely. Uh again using the
same arguments if we create advanced AI
maybe with conscious capabilities like
humans are if we figure out how to make
believable virtual realities. Adding
those two technologies together
basically guarantees that people will
run a lot of games or simulations or
experiments with agents just like me and
you conscious agents populating virtual
worlds. And statistically the number of
such simulated worlds will greatly
exceed the one and only physical world.
So if there is no difference between a
simulated you and real then
statistically you're more likely to be
in one of those simulated worlds.
>> Okay. Uh that makes a lot of sense. Now
given the likelihood that we will we're
obviously showing that we will pursue
artificial super intelligence. Uh if I
take your same logic from the fact that
we're likely to be in a simulation
because we know we would make a
simulation because we're doing it right
now. Uh and therefore you get into the
point where you would just make billions
of those. And so if you have a one in a
billion chance of being inside of a
simulation, you're effectively
guaranteed to be in one now because
there would just be so many of these
things running. Um, doesn't it also then
make sense that the Matrix was
effectively a documentary and we are
inside of a simulation created by
artificial super intelligence designed
to mllify us. Um, if we ever had a
physical body in the first place.
>> So, it's hard to tell from inside of a
simulation what it is all about. You
really need access to outside. uh it
could be entertainment, it could be
testing, it could be
some sort of scientific research. If we
look at the time we actually find
ourselves in, we are about to create new
worlds, virtual realities. We are about
to create new intelligent specy AI.
There is a lot of kind of meta
inventions we are right about to make.
And so if someone was interested in
studying how civilizations go through
that stage, how do they control these
technologies or fail to control them,
that's the most interesting time to run.
You're not going to run dark ages. There
is not as much happening. It's less
interesting. But this seems to be like a
meta interesting state to be in.
>> It's hard to tell cuz we're inside the
simulation, but you're saying it's a
little bit suspect that we're living in
the most interesting time ever.
>> Yes. And I think it's interesting not
just because I'm living in it, but
objectively it's a time of meta
invention. You can go back through
history and say, "Oh, here they invented
fire. Here they invented a wheel."
That's all great, but those are just
inventions. They are not meta
inventions. Whereas now we're doing
something godlike. We are creating new
worlds. We are creating new beings. And
that's something we have never done
before.
>> Do you ever think like a sci-fi writer?
So I think the difference between
science fiction and science used to be
maybe 200 years. They wrote about travel
to the moon. They wrote about kind of
internet and computers and it took
hundreds of years to get there. And then
it was like I don't know 20 years. And
now I think science fiction and science
are like a year away. The moment
somebody writes something, it already
exists and there is really no new
science fiction ideas where it's like
completely novel technology not
previously described or someone already
working on it if physics allows it.
>> That's really interesting. Uh especially
when you think about writing now for
true science fiction in terms of what
will become possible in the future is
effectively impossible because you're
talking about super intelligence and
good luck as a person. uh locked in your
not super intelligence to actually
describe that. The reason that I ask
though is um when I start thinking about
things like that like why would we run
this simulation? What clues are in like
if this is a simulation what clues are
in it? Uh so for instance um the whole
Christian idea for sure and there might
be more religions that have the same
idea but that man is made in God's
image. Okay. Well, if God is the
13-year-old running the simulation or
Sarah Connor or I guess John Connor
running the simulation trying to figure
out why we created Skynet and what we
can do to nudge it off course, um, you
know, you think of them as sort of
moving from radioactive rubble to
radioactive rubble trying to like find
an answer to this and spinning up a
simulation to get that answer. Um that
to me becomes very intriguing in terms
of
hypothesizing
as to why this moment, why are we the
way that we are? What can we learn about
the people trying to simulate us? When I
ask questions like that of engineers
such as yourself, there's almost I don't
have time to think like a sci-fi writer
vibe. Um is it just that you're you
don't find that interesting? You don't
find it revoly? Um why do you assue
that? Because in interviews I've seen
people ask you time and time again like
how would AI kill us and the answer is
always some variant of listen you're
asking me how I would kill us which is
not interesting because the super
intelligence is going to but I find
that's the cathartic thing that people
want like they want to like when you
have a wound you kind of want to poke at
it like they want to get a sense of what
would this really look like and so even
though it's not literally true it's
deeply cathartic to
explore or the known possibility set or
what humans can know.
>> And this is exactly why I refuse to
answer. I want to make sure what I tell
them is true. I don't want to lie to
them. If squirrels were trying to figure
out what humans can do to them, and one
of the squirrels was saying, well,
they'll throw knots at us or something
like that. It would be meaningless BS
story. There is no benefit in it. The
whole point I'm trying to make is that
you cannot predict what a smarter agent
will do. you cannot comprehend the
reasons for why it's doing it. And
that's where the danger comes from. We
cannot anticipate it. We cannot prepare
for it. I do think the singularity point
is where science fiction and science
become the same. The moment something is
conceived, we have super intelligent
systems capable of developing it and
producing it immediately. It's no longer
200 years away. It's reality. And you
can't see beyond that event horizon. You
cannot predict what's going to happen
afterwards. And with science fiction,
you cannot write meaningful, believable
science fiction with a super intelligent
character in it because you are not.
>> All right, let's ground things then in
what we can predict and we can know
right now. Something that's on
everybody's mind and I've been talking
about this in my own content is the
labor market seems to be softening.
You've got places like Amazon that are
just cutting jobs like crazy. Um, and
just saying outright this is largely
because of optimizations that we're able
to make because of AI, how does this
transition play out? Like even if you
concede that uh a non-destructive AI
would give us um essentially an age of
abundance, we're still going to go
through a transition period where our
jobs go away, etc., etc. What are the
what are the steps that you see
happening in the labor market? So as we
have more and more increased percentage
of populace unemployed, hopefully
there's going to be enough common sense
from the governments to prevent
revolutions and wars to provide for the
people who lost their jobs and probably
cannot be retrained for any new jobs. So
once you hit 20, 30, 40% unemployment,
that's where it's really going to kick
in. The only source of wealth at that
point is the large corporations making
robots, making AI, deploying them, all
the trillion dollar club members.
Essentially, at this point, you need to
tax them and use those funds to support
the unemployed. That's the only way to
really make sure the financial
part of that problem is taken care of.
What remains is the meaning. What do you
do with all this free time and millions
of people who have it? Traditional ways
of spending your time to relax. You go
for a hike in a park. Well, there is a
million people in that park right now
hiking. That kind of changes how
peaceful it is and how relaxing. So, we
need to accommodate not just change in
financial reality, but also change in
free time and capabilities of supporting
that many people with that much free
time.
I have as much pessimism around our
ability to do that well as you have our
likelihood of surviving. So I'll say
99.99%
chance that the government completely
messes that up. Uh I think the
transitionary period will be violent. Um
when you look out at this knowing what
you know about humans and governments,
what what odds do you give it that
that's a smooth transition?
>> It's very likely to continue to be as
history always been. We had many
revolutions, many wars, a lot of
violence. That's why we hear stories
about people who can afford it building
bunkers, securing resources because they
anticipate certain degree of unrest.
Absolutely.
>> What degree of unrest do you anticipate?
>> Really depends on the percentage of
population which quickly gets
unemployed. If it's a gradual process,
we can kind of learn and adopt and
provide safety net. If over a course of
weeks, months we're losing 10, 20, 30%
of jobs, that's a very different
situation.
>> I can't imagine a scenario where jobs
would be lost that quickly. To your
point, we've already created, you said,
billions or even trillions of dollars of
value in the technology, but it hasn't
been deployed yet. Uh an example you
often use is the video phone invented in
the 70s but not really adopted uh
largely because of infrastructure I
would say until the whatever 2011
uh where that starts to really gain in
popularity. So I have a feeling like
just the deploying of all this stuff uh
is going to take time. So, in a world
where
an unimaginable amount of people, which
I'll clock at, in the US, call it 6 or 7
million people lose their jobs in the
next 5 years. Um, that I would consider
fast and just horrifyingly destructive.
One, does that feel plausible to you in
terms of numbers and timeline? And two,
in that scenario, um, how distressing do
you think that transition will be?
>> It seems very likely. So, take
self-driving cars. I think we are very
close to having full self-driving
without supervision. The moment that
happens, you have no reason to hire a
commercial driver, right? All the truck
drivers, all the Ubers, all of that gets
automated as quickly as they can produce
those systems. And I think Tesla is
ready to scale production of their cars
to exactly that scenario. So what is it
6 million drivers in the country? I
don't know the actual numbers but that
would be exactly what you're describing
and it's very unlikely that they can be
quickly retrained for something which is
also not going away.
>> Okay. So in that
Resume
Read
file updated 2026-02-12 01:36:59 UTC
Categories
Manage