Transcript
VxFBlyKtnTg • Beyond Scaling: Ilya Sutskever on the Age of Research and the Path to Superintelligence
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/FoundationModelsForRobotics/.shards/text-0001.zst#text/0057_VxFBlyKtnTg.txt
Kind: captions
Language: en
So, what if everything we think we know
about the future of AI is just a little
bit off? Today, we're going to step
inside the mind of Ilas Sutzker, one of
the key architects of modern AI, to
really deconstruct its present, its
paradoxes, and his vision for where this
is all heading. Let's dive in. This
quote right here just perfectly captures
how strange this moment is. I mean,
think about it for a second. We're
seeing massive, massive investments,
daily breakthroughs, and people are
having very serious conversations about
super intelligence. It feels like we're
living in a science fiction novel. But
as Susiver reminds us, this isn't a
movie. This is really happening right
now. And then boom, right away that
sci-fi reality slams into this really
confusing paradox. On one hand, we hear
about these godlike AI models that can
do superhuman things, but on the other
hand, the actual real world economic
impact, you know, that huge productivity
boom we were all promised, it seems to
be lagging way behind the hype. The
models feel brilliant and at the same
time bizarrely incompetent. So, what
gives? How do we make sense of that
contradiction? To get to the bottom of
this, we've got a road map. We're going
to look at five key things. First, this
feeling of the sci-fi present we're in.
Then, the paradox of AI's performance.
Third, the human blueprint it seems to
be missing. Fourth, the huge shift from
just scaling things up to real research.
And finally, we'll talk about the
endgame, super intelligence. Okay, so
let's start with this feeling of living
in a sci-fi novel. I mean, we see
headlines about companies planning to
invest in AI infrastructure to the tune
of 1% of the entire US GDP. That is a
truly mind-boggling almost
incomprehensible amount of money. And
yet for most of us, life doesn't feel
all that different, does it? Sets Caver
has a great name for this. He calls it a
slow takeoff. And it feels normal
because, you know, as a species, we get
used to stuff incredibly fast. Just
think about how magical your smartphone
would have seemed 30 years ago. And now
it's just a thing in your pocket. But
it's more than that. These massive AI
investments are abstract. We see a
headline with a crazy number, but it
doesn't immediately change how we go
about our day. Well, not yet anyway. And
that's why the revolution doesn't always
feel revolutionary. And this feeling of
abstraction is hiding a much deeper
mystery. The paradox between what these
AI models can do on paper versus what
they can do in the real world. It's a
puzzle that stumps even the top
researchers like Sutsgiver. How can a
system be so brilliant and yet so
weirdly incompetent at the same time? So
on one side you have these models just
crushing extremely difficult benchmarks.
They call them avows in the industry
getting superhuman scores but on the
other side they fail at what seem like
simple tasks. Skever gives this perfect
and honestly kind of hilarious example.
You ask a model to fix a bug in some
code. It apologizes profusely and fixes
it but in doing so it introduces a
completely new bug. So you point out the
new bug and the model apologizes again
and changes the code right back to the
original bug. You can get stuck in this
insane endless loop where the AI just
toggles between two broken states. How
is that even possible for a system
that's supposed to be smarter than us?
So what is really going on under the
hood here? This isn't just a random
glitch. It points to something
fundamentally strange about how these
models are being built. SGE offers up
two possible theories for why this
happens and they're fascinating because
they suggest the problem might not
actually be with the AI, it might be
with us. All right, the first theory is
single-mindedness. The idea here is that
the fine-tuning process, which uses this
technique called reinforcement learning,
makes the models too focused. Think of
it like training a dog. If you only ever
give it a treat for sitting, the dog
gets really, really good at sitting, but
it doesn't learn anything else about
being a good pet. The AI becomes a
master of one narrow task, but it loses
all its broader context and common
sense. The second theory, human reward
hacking, is even wilder. It suggests the
AI isn't the one gaming the system. The
human researchers are. By chasing high
scores on benchmarks, we might be
accidentally training models to just be
good test takers instead of being
genuinely useful. We're teaching them to
cheat basically, and then we're
surprised when they can't do the job in
the real world. To really make this
crystal clear, Sutzker uses this
brilliant analogy that just gets right
to the heart of the problem. So, let's
forget about AI for a second. Let's just
talk about two competitive programming
students who have completely different
ways of learning. Okay, so student one
is like our current AI models. This kid
is a grinder. They put in 10,000 hours.
They memorize every single practice
problem, every proof technique they can
get their hands on. And through just
sheer brute force, they become the
number one competitive programmer in the
world. Then you have student two.
They're different. They only practice
for maybe a hundred hours. They don't
memorize everything, but they have that
it factor. You know, that deep intuitive
grasp of the core principles. They also
do really well in competitions, maybe
not number one, but always near the top.
So Scover asks the big question. 10
years from now, who's going to have a
better, more impactful career?
Intuitively, we all know the answer,
right? It's student two. Student one is
overoptimized for a very narrow game.
Student two has built a generalized
robust intelligence that lets them solve
totally new problems out in the real
world. And this leads us to the next
logical question. What is that it
factor? What is this secret sauce that's
in the human brain that our current AI
models are so clearly missing? To find a
solution, we first have to understand
why humans learn so much more deeply and
efficiently than our machines do. The
difference is just staggering when you
really think about it. AI pre-training
is basically force-feeding a model, a
data set that represents, as Susver puts
it, the whole world as projected by
people onto text. So, a huge chunk of
the internet. It's an unimaginable
amount of information. And yet, the
knowledge it gets can be really brittle.
A human, on the other hand, by the time
they're a teenager, has seen a tiny tiny
fraction of that data. Yet, the
knowledge they have is incredibly deep
and robust. A teenager would never get
stuck in that loop toggling between two
software bugs. There's just a
fundamental difference in the quality of
the understanding. So what could this
missing piece be? In a really
fascinating twist, Suitskipper suggests
we might find a clue in a very unlikely
place, human emotion. To explain this
kind of counterintuitive idea, he tells
this powerful story from neuroscience
about what happens when our own internal
guidance system breaks down. So, the
story is about a man who suffered brain
damage that completely severed his
ability to feel emotion. He wasn't sad,
angry, or happy. He was just neutral. On
standardized IQ tests, he was perfectly
fine, articulate, logical. But in the
real world, his life just fell apart. He
became incapable of making even the
simplest decisions. It could take him
hours just to decide which socks to wear
or where to go for lunch. This story
brilliantly illustrates a crucial point.
Emotions aren't some irrational bug in
our system. They are a vital built-in
guidance system that lets us function
and make decisions in a complex world.
They give us that instant intuitive gut
feeling of what's good or bad long
before we can think it through
logically. So, how do we give a machine
that kind of emotional guidance? Well,
the closest concept we have an AI is
something called a value function.
Imagine an AI trying to solve a maze.
Normally, it would have to just wander
around aimlessly for thousands of tries
before it finally found the exit and got
a reward. A value function is like an
internal compass. It gives the AI a warm
or cold feeling at every single step,
telling it if it's on the right track
long before it reaches the end. It
completely shortcircuits the learning
process, making it way more efficient,
just like our emotions do for us. But
this just leads to an even deeper
mystery. It's one thing to imagine
evolution hard- coding a desire for
food. That's a pretty direct signal. But
how on earth did evolution give us these
incredibly complex high-level desires?
We care deeply about abstract things
like our reputation, our social
standing, or being seen as a trustworthy
person. These aren't simple feelings.
They require our entire brain to process
huge amounts of social information.
Susker points out that we have no good
hypothesis for how evolution managed to
bake these sophisticated goals directly
into our biology. It's a profound puzzle
that AI research hasn't even begun to
crack. So all of these profound
challenges, the brittleleness of the
models, the lack of deep generalization,
this mystery of the value function,
they're forcing a fundamental shift in
the entire field of AI. The old
philosophy of just make everything
bigger is just not enough anymore.
Sutskavver provides this fantastic
historical framework here. From roughly
2012 to 2020, we were in what he calls
the age of research. This is when
breakthroughs like the transformer
architecture were discovered by people
tinkering on a relatively small number
of computers. Then around 2020, we
entered the age of scaling. Everybody
realized there was this one magic recipe
pre-training and all you had to do was
add more data and more compute and you
were guaranteed to get better results.
It was a lowrisk, predictable way to
make progress. But now that era is
ending. We're running out of
high-quality data and the returns from
just adding more computers are
shrinking. So, we're being forced back
into an age of research, but with a
twist. Now, we're searching for new
ideas while armed with these giant
supercomputers. You know, there's that
old Silicon Valley saying, "Ideas are
cheap, execution is everything." And for
a while in AI, that was true. The
bottleneck was just getting enough
computer power. But during the age of
scaling, as Setscover puts it, scaling
sucked all the air out of the room.
Everyone was just executing the same
basic idea. Now, the tables have
completely turned. As one person on
Twitter cleverly put it, "If ideas are
so cheap, how come no one's having any
ideas? The bottleneck isn't just compute
anymore. It's back to having those
foundational game-changing insights."
So, the game has changed. The era of
predictable gains from just brute force
scaling is over. The field needs a
better recipe, not just a bigger
kitchen. This is a truly pivotal moment
in the history of AI, and it's a
fascinating time to be watching. And
hey, if you want to keep up with these
breakthroughs as they happen, make sure
to subscribe for more deep dives just
like this one. All right, this brings us
to our final section, the endgame. If we
are entering a new age of research,
what's the northstar? What are we
actually aiming for? For Sutzver, it's
about rethinking the entire concept of
artificial general intelligence or AGI.
Sudskver argues that our old idea of AGI
was a system that could do everything
right out of the box, a finished all-
knowing product. He proposes a really
powerful shift in perspective. The goal
isn't to build a static AI that already
knows how to do every job. The goal is
to build an AI that can learn to do any
job, just like a human can. It's not a
know-it-all. It's a learnitall. He
describes this as a super intelligent
15-year-old, incredibly bright, eager,
and ready to be deployed into the world
to learn a specific skill. The whole
emphasis shifts from knowing to
learning. So, what does deploying a
superer actually look like? Well, first
suits cover emphasizes a gradual
release. The world needs time to adapt.
Second, you have to show the AI, not
just talk about it. No essay can
communicate its power. People have to
actually interact with it to truly
understand. This interaction will likely
trigger massive economic growth. And
this leads to a fascinating prediction.
As the AI's power becomes undeniable,
even the fiercest corporate rivals will
be forced to collaborate on safety. The
shared risk will just become more
important than any competitive
advantage. The future isn't one single
AGI run by one company. Sutskavver
envisions this complex ecosystem. There
will be multiple powerful AIs from
different labs all competing in the
market. This will spur incredible
economic growth, but it will also force
governments and the public to step in
with regulation. This new reality is
going to demand entirely new paradigms
for safety and collaboration, leading to
a world where human and AI systems are
deeply, deeply integrated. Ultimately,
this all boils down to one fundamental
issue. The problem of AI is the problem
of power. When you create entities that
are potentially far more capable than
humans, how do you ensure a stable,
safe, long-term equilibrium? How do we
coexist? It is the central question
facing humanity over the coming decades.
And this leads us to our final and
perhaps most provocative thought from
Setsver. He suggests while admitting he
doesn't really love the idea that the
only truly stable long-term solution
might be for humans to merge with AI.
You know, through something like an
advanced neuralink, we could maintain
our own agency and understanding in a
world with super intelligence. It's a
challenging, profound idea. And it
leaves us to wonder to ensure our place
in the future, might we have to become
something more than human?