Interview: Deepfake Detection and the Future of AI with Hany Farid | Particles of Thought
nG2_GhNdTek • 2025-08-26
Transcript preview
Open
Kind: captions
Language: en
We have to put our hands back on the
steering wheel and we have to start
getting serious about how do we put
guard rails on the system because what
we know is that if you unleash Silicon
Valley they will burn the place to the
ground to get to the finish line first
and we've got to start putting
guardrails in place and the thing is is
that this is a borderless problem right
we this can't be an only US or an only
EU or an only we have got to start
thinking about this globally
>> and I don't think we are doing that very
[Music]
[Applause]
[Music]
Hello everybody. I sat down with Hani
Farid, one of the leading voices in AI
research. He's a professor at UC
Berkeley where he works on digital
forensics and misinformation, especially
things like deep fakes, image analysis,
and how we perceive fake content. He's
also the chief science officer at GetRal
Labs, a company that focuses on the
authentication of digital media, telling
us what's real out there and what's
fake. So, yeah, he's busy, but he made
time for us. This isn't your standard AI
interview, cuz I know y'all have heard
too much about AI and you're tired of
that same old conversation. This time,
we talk about how AI actually works and
how he can detect the fake stuff. and of
course what we all want to know, what he
thinks the future holds. If you enjoy
the show, I'd love your help spreading
the word. So, take a moment to rate,
review, or leave a comment. And don't
forget to subscribe so you don't miss
anything. Your support really helps us
reach new audiences. So, let's go.
Honey, welcome to Particles of Thought.
>> It's great to be here, Hickey.
>> All right, man. So, look, this isn't
going to be like your normal interview
because this first question I got for
you
>> Yeah.
>> is something that the people need to
know.
>> Good.
>> All right. So,
>> I can't tell you I'm a little nervous
already, but go ahead.
>> All right. So, listen. You're you're an
expert on AI
>> and you specialize in identifying deep
fakes.
>> So, there's been three occurrences in
recent history that has everybody in a
tizzy.
>> Yeah.
>> The first one was the movie The Matrix.
>> Yeah. The second one is when physicists
came up with their holographic theory
which seems to indicate that life could
be a simulation.
>> Yeah.
>> All right. So, and the third one now is
AI.
>> Yeah.
>> And the question everyone has is
>> is reality a deep fake?
>> And since you are an expert in
uncovering deep fakes, are we cold, man?
Is reality what we think it is? And if
it is a deep fake, would a deep fake
expert have an idea of how to uncover
that?
>> No. I mean, if this was all a
simulation, my job would be a lot
easier, honestly, because then you sort
of give up, right? There's there's no
more reality anymore, I think. But this
is sort of where we are, Hakee.
>> Yeah.
>> Is we are now questioning everything,
not just what I see on social media, but
our our existence. we are starting to
question and our existence today and
yesterday and in the future
>> and it does feel to me I've been I've
been thinking about these problems for
25 years that we went from if I was
having a podcast with you 25 years ago
your questions would have been of the
form hey let's talk about Photoshop and
how people splice together two images
and maybe we would talk about the Lee
Harvey Oswald photo or the moonlanding
photo and today a perfectly reasonable
question is let's talk about the nature
of reality and that's happened in 25
years. Wow. Imagine the next 25 years,
the kind of conversations we're going to
be having. So, to get back to your
question, I don't know. I honestly don't
know. And I also don't know where this
latest AI
boom is taking us. And I I don't think
anybody knows honestly. But here's what
I can tell you.
>> If you look at the personal computer
revolution, that took roughly 50 years
to unfold, which at the time felt really
fast.
>> Where do you start from? like
>> I'm going to start in 19 let's start in
1950 1945 and let's go to about 2000
when more than half of US um homes had
uh a personal computer
>> then you look at the internet revolution
right from the beginning of the htt http
or html protocol Tim Berners Lee to
about half the world's population being
online that was 25 years the mobile
revolution was less than 10 years and
the AI revolution has been well the AI
revolution of course started in 1950 but
the one we are experiencing now was
about 2 to 3 years we have gone from 0
to 100 miles an hour where we are now
talking about existential threats of AI.
We're talking about 50% of jobs being
eliminated in the next 5 years. We are
talking about general artificial
intelligence. We're talking about the
Terminator. And we would not have had
this conversation 5 years ago. And by
the way, on top of all of that, we are
don't even know what's real anymore
because we consume all of our content
from online sources. Online sources have
been polluted for a while. out there
getting more polluted thanks to AI and
suddenly our whole notion of reality is
up in the air and that is it's
unsettling. I think people feel
unanchored and I don't
>> I don't know how to help everybody with
that.
>> Yeah. All right. Well, listen man, let's
step back
>> because everybody listening may not know
what's even meant by AI, what's meant by
artificial general intelligence, what's
meant by deep fakes. So give us some
bearing and foundation and define AI and
I will tell you I started in
>> big data
>> back in 2008.
>> Yeah.
>> And I was doing what was called machine
learning, right? Classified I mean
supervised and unsupervised learning and
I was working on data sets in astronomy
and astrophysics. And it was for the
Vera Rubin telescope which has just
debuted its its images this year.
>> And I was being told, hey, you're
building the software infrastructure for
analyzing that data now.
>> Yeah.
>> No, I wasn't.
>> No, I wasn't cuz AI didn't exist. So
>> good.
>> I think of machine learning as just
statistics. Define AI for us. Define
deep learning. I mean deep fakes and
define general intelligence.
>> All right. Let's start with AI because I
think you raised a really good point
which is that everything we are talking
about today is in fact not AI. It is
machine learning. It is statistics.
Almost everything we see from the chat
GPTs of the world to deep fakes which
I'll define in a minute to all of the
things that we are seeing impacting our
day-to-day lives is machine learning. So
let me rewind to 1950. 1950 was when the
term give or take when Alan Turing and
John McCarthy conceived of the term or
the concept of AI and the idea was then
quite bold given where computers were
can we imbue intelligence the way we
conceive of intelligence with humans
into machines
>> and the idea was yes we should be able
to do this and I think it's because when
we think about our own brain and how we
operate it seems like it should be
straightforward but of course it's not
talk to any neuroscientist and they'll
tell you that but for a long time for
about 30 years the field of AI struggled
to be relevant
>> because there was no path to creating
human intelligence in a machine
>> and then around the 1980s the field
splintered and the field of machine
learning came about which was what you
were just referring to and here the idea
is we are not going to imbue
intelligence into machines we are going
to learn it from data So we are simply
going to present to a machine a bunch of
data and have it infer the rules and the
logic that we think that we have and
frankly it failed for about 20 years. So
1980 to about 2000 the field really
struggled for relevance and it struggled
for relevance for two reasons. One is
there was not enough data because there
was no internet and there was not enough
computing power because there was no
Nvidia.
>> So what happened at the turn of the
millennium right the internet rose. What
did the internet give rise to? lots of
things, but one of the things it gave
rise to is a boatload of data
>> from us. Right? The ultimate irony, by
the way, is if AI comes for us, we
created it by giving it all of our our
data to learn from.
>> So around the turn of the millennia,
around 2000, huge rise in data being
pushed online and then of course a huge
rise in computing power with things like
Nvidia.
>> So define Nvidia for the audience for
people who don't know.
>> Good. So Nvidia is the company of course
that makes GPUs, graphical processing
units. This is the core computip.
Yeah. That is particularly good at the
types of computation you need to do
machine learning.
>> And suddenly we saw this explosion. It
really started around 2015 2016 in the
last decade has really exploded with
phenomenal amounts of data, phenomenal
amounts of computing power. And you got
to give credit to the statistics and
machine learning community. Some real
insights on how to do machine learning
and how to do the types of things that
you did. you were way ahead of the curve
in the in the early knots and then we
got better at doing that because we had
some insights from statistics and
physics and mathematics and computer
science and engineering. So that is so
everything you hear today in my opinion
that we talk about AI is machine
learning. We've sort of conceded that
we're just going to call it AI because
that's just the term that has gotten
adopted. But underneath it, understand
that everything that is happening is
pattern matching.
>> That you take a bunch of data and you
learn the rules implicitly, not
explicitly from the data.
>> And that's both very powerful and very
dumb. Right? It's powerful because you
can learn complicated patterns, but it's
dumb because you don't know what the
rules are. Like it's not learning that
gravity is - 9.8 m/s squared. It's
simply learning that if you film
something falling, this is how it will
fall. But it doesn't know what the
physics of it are.
>> Right? It simply knows this is where the
ball will be at any given moment. Okay?
>> Now, so that's AI/Machine Learning.
>> Um, artificial general intelligence is a
term that I don't think anybody knows
how to define, but I'm going to give you
my definition of it, or at least we
don't agree on a definition. So,
typically, historically, when you've
looked at machine learning and AI
systems, they've been bespoke. So,
you'll have a system that does medical
imaging. You'll have a system that does
drug discovery. You'll have a system
that drives um self-driving cars.
>> The idea of AGI, artificial general
intelligence, is that it does
everything.
>> What humans do, right?
>> Um
>> or stem cells
>> or that's right. Exactly. That's a good
example of it. Right. So, so is that a
year away, 5 years, 10 years? I don't
know. I don't think anybody knows. But
there's a lot of speculation about that.
I would argue by the way that we have
some notion of AGI because if you go
over to your favorite large language
model like chat GPT yeah or claude you
can ask it a lot of different questions
about physics about medicine about
computer science about televisions
anything and so is that AGI I don't know
I suppose it depends on how you define
it
>> that that takes me to the touring test
right that was the big so so so what was
the touring test is it if you can't tell
the difference between when you're
talking to a machine or
>> Yeah. So you you got to by the way, you
got to have respect for Alan Turing.
>> Oh, absolutely.
>> In the 1950s, this guy was thinking
about things that were really outrageous
to be thinking about at the time.
>> Maybe he was a time travel.
>> That's right. That's really funny. Don't
don't give anybody any ideas by the way.
>> Okay. So what is the Turing test? So
when Alan Turing first conceived of this
notion of imbuing intelligence in
machines, he came up with a mechanism to
determine if you have solved that
problem. And the idea was that you would
have two computer screens uh and behind
one computer screen was a human and
behind another computer screen was a AI
system. You of course couldn't see what
was behind it. And you were allowed to
interact with it with a keyboard. Okay?
And if you and you can ask it questions,
you can have a conversation the way you
and I are having right now. And if you
could not tell the difference, then that
machine has passed the so-called touring
test and it has what probably today we
would call AGI. Okay. So now we've done
AI and we've done AGI, right? Okay. So
let's talk about deep fakes which is
this sort of sliver of all of this.
>> Yeah.
>> So deep fakes is an umbrella term for
using machine learning AI to whole cloth
create images, audio and video of things
um that have never existed or happened.
So for example, I can go to my favorite
deep fake generator and say give me an
image of Hakee in a studio doing a
podcast with Professor Hani Fared
>> and actually it would do a pretty good
job because you have a presence online.
I have somewhat of a presence online. It
knows what we look like and it would
generate an image that's not exactly
this but something like that. Or I can
say please I by the way I still say
please when I ask AI for for things. One
of my students told me that this is a
good idea because when the AI overlords
come they're going to remember you were
polite to them. Ah,
>> I actually really like this advice.
>> Wait a minute. So, I read an article.
>> Yes. It cost tens of millions of
dollars.
>> The energy ultimate. Yes. Just saying
please and thank you. I still do it by
the way. And even in my head right there
when I was asked when I was I I still in
my head say please.
>> Well, listen. I have AI connected to my
AI, right? And so my AI corrects my AI
prompts
>> to proper grammar and it's like
>> please. It puts please in there.
>> I know. And it does cost tens of
millions of dollars for that extra
token. Okay.
>> So, I will ask it for an image of a um
of a unicorn wearing a red clown hat um
walking down the street at Times Square
and it will generate that image. Um I
can ask uh generate an audio uh of
Professor Hani Fared saying the
following, right?
>> Um I can generate a video of me saying
and doing things I never did. And you
can clearly see the power of that
technology from a creative perspective.
If you and I are having a conversation
and in post we said something we didn't
mean to, we can just fill it in with AI
now.
>> Well, here here's the thing that makes
me you just mentioned how we're only two
three years into this. So, however good
it is now, you know, this is the worst
it will ever be,
>> right?
>> So, if you look at the So, I can tell
you by the way how good it is.
>> So, in addition to being trained as a
computer scientist and applied
mathematician, I've been somewhat
trained as a as a cognitive
neuroscientist. And we do perceptual
studies. So what we do is we recruit
participants. We show them images, audio
clips and video. And we tell them half
of the things you're going to look at
are real. Half of the things are AI
generated. We explain to them what AI
generated is. We give them examples of
that.
>> And for images as of last year, people
are roughly at chance at distinguishing
a real photo from an AI generated photo.
>> So what you mean by that is if they were
just if you had a a monkey behind a
keyboard,
>> flipping a coin.
>> Flipping a coin.
>> Yeah. Yeah. The monkeyy's probably
better than you. By the way, I'm I'm
going to go off and guess. Um, so with
audio, so we play a clip of somebody
speaking like you and then we play an AI
generated version. They're slightly
above chance, not like 65%.
>> On image at chance, at audio, slightly
better than chance.
>> And video, they're a little bit better,
but all of those trends are going
towards chance. So here's what we know.
everything in the next 12 months, 18
months, 24 months, I don't know what the
number is,
>> it will be indistinguishable to the
average person online, right? And that
is
>> that is a weird world we're living in
because think about how much in first of
all, the vast majority of Americans now
get the the the majority of their
information from online sources and
unfortunately from social media too.
>> And that and because it is so easy to
create this content, understand all this
is is a text prompt away. I type,
"Please give me an image of this,
generate this audio, generate this
video." There are dozens of services
that will do this extremely inexpensive
or for free. And you can carpet bomb the
internet with fake images of the
conflict in uh Gaza.
>> Fake images.
>> I have seen them too. Fake images of the
flood in Texas, fake images and video of
the fires in name it across the boards,
right? Fake images of people stuffing
ballot boxes. Now we have a threat to
our democracy.
>> Wow. So suddenly our sense of reality
coming back to your first very good
question is up in the air because I can
create whatever reality I want and
understand that there's sort of three
things happening here when we talk about
deep fakes. There's the creation of it.
That's what we've been talking about.
>> There's the distribution which we
democratized 20 years ago. So anybody
can
>> publish to the world and that's very
powerful and very terrifying because
there's no editorial standards on social
media. And then there's the
amplification that we have become so
polarized as a society that when you see
things that conform to your world view,
you are more than happy to click like,
reshare. And now you have creation,
distribution, amplification.
>> Wow.
>> That's the ball game,
>> right? That's the ball game for
spreading massive lies, conspiracies,
and disinformation campaigns that affect
our global health, our planet's health,
our democracy, our economy, everything.
Everything. So let's get into how these
fakes are generated. So start with
images.
>> Good. So let's start with images because
in some ways it's the easiest one, but
all of these have a similar theme. And
one of my favorite techniques for
generating images called a generative
adversarial network or a GAN. And here's
how it works.
>> Wait a minute. Wait a minute.
Adversarial.
>> Adversarial.
>> So that means that you're fighting your
computer.
>> Two two computer systems are fighting
each other. And this is sort of the
genius of this technique. So here's how
it works.
>> You have two systems.
One system's job is to make an image of
a person or a landscape or whatever you
want. Yeah. And so what it does, it
starts by, this is literally true, it
just splats down a bunch of random
pixels. So I say, generate an image of a
of a person and it says, "Okay, here's a
bunch of so so think uh the monkeys at
the keyboard typing randomly. Let's see
if this is Shakespeare,
>> right?
>> And then it takes that image and it
hands it to a second system and it says,
is this a face?" And that system has
access to millions and millions of
images that it scraped from the internet
that are faces.
>> I see.
>> And that system says that thing that you
generated doesn't look like these things
over here.
>> And it gives the feedback to the
generator and it says, "Nope, try again.
>> Modify some pixels. Send it back to
what's called the discriminator. Is it a
face? No, try again." And they work in
this adversarial loop. So it's like
somebody's checking your homework.
>> But it it it seems like it could get
stuck never getting to a face,
>> you would think. And that's what's
amazing about the GANs. the gen is that
they converge.
>> They converge.
>> And part of that is the way they they've
been trained. But that's what's the
genius of this is that the generator is
not very smart
>> because all it's doing is modifying
pixels. And the discriminator is
actually quite simple. It's simply
saying, does this thing look like these
things? And because you pit them against
each other in this adversarial game,
this sort of amazing thing happens out
the other side.
>> So here's the question. In on average,
how many iterations does it take? And
then how much time does that translate
to?
>> Yeah, that's a real great question. So
typically the time is in seconds. So
there's two phases. There's you train
the GANs. That's a really long process.
But then what we call inference, which
is that run this thing, it happens in
seconds. And the reason it happens in
seconds is by the way that is hundreds
of thousands of iterations. Wow. But
it's on a GPU, which is very powerful
and very fast. And then there's these
tricks to make it even faster. You start
with small images and then you make them
bigger over time. So there's these
tricks, but it is literally seconds to
make that image.
>> And what the brilliance of that is the
two systems are competing with each
other.
>> And then this thing that seems like
intelligence come out even though it's
not. If you think about those two
individual components, right?
>> They're pretty basic, but then you have
this like emergent behavior almost. It's
like you know how to generate images of
people. That's amazing.
>> So let's have a little fun. I understand
good
>> that you brought me some fakes
>> and some real images
>> to put to the test. Good
>> to see if I can discern the difference.
>> So before I I'm going to play for you a
couple of audios. Before I do this, let
me say I've been doing this for a long
time and I've been I'm pretty good at
it. I'm pretty good at what I do and I
created three audio samples. I'm going
to play them for you.
>> Wait, are you allowed to say that that
you're you're good at what you do? I'll
say that. Honey is really good.
>> I said pretty good, by the Yeah, he's
amazing.
>> But this is amazing. This is this is
this is a true story, by the way. So, I
made three audio clips for you of me
talking and you and I have been talking
for a little while, so you now know what
my voice sounds like.
>> And uh I got off the plane and I was in
the car coming over here and I wanted to
make sure they worked. And I played all
three of them and I couldn't tell which
one of me was real or fake. I wasn't
100% sure.
>> Wow.
>> And I do this for a living and it's my
voice,
>> right?
>> So, okay. So, that is Okay.
>> So, wait a minute. Which AI did you use?
Is this something that you created or
something generally available?
>> So here's the thing you have to
understand about AI is this is so
readily available. So here's what I did.
I went to a service. It's a commercial
service. Um I uploaded I think it was
about 3 minutes of my voice.
>> I said please um uh please clone my
voice. Um and it clones my voice. And by
what I mean by that is that it learns
the patterns of my voice, what I sound
like, the intonation, my cadence, how
fast I speak, where I put the pauses.
And then I can simply type and have it
say anything I want to say.
>> And so I'm going to I'm going to read
I'm going to have you play I'm going to
listen have you listen to three
sentences. Okay.
>> Um and one of them is fake. I'm going to
give you a hint. One of them is fake and
two are real.
>> Okay. And let's see what you we you can
do. Okay. Here we go.
>> And in fairness, this is not the best uh
speaker, but okay.
>> Are there guardrails in our law?
>> Ah, good. Uh, so first of all, when I
went to do this this service, um, I
uploaded my voice and there's a button
that says, "Do you have permission to
use this person's voice?" And and I did
because it was my voice, but I can
upload anybody's voice and click a
button.
>> The laws are very complicated and they
actually vary state-to-state and of
course internationally. Wow.
>> So, there are almost no guardrails on
grabbing people's likeness. And even if
there were,
>> there's,
>> you can still do it anyway.
>> There's there's no stopping this.
There's no stopping it. Okay. All right.
number one. Oh, and by the way, the the
three U this is part of a talk I gave
recently on deep fakes. So, you'll hear
a consecutive thing. Okay, ready?
>> And if you invite me back next year,
almost certainly everything will have
changed. Uh the nature of creation of
deep fakes, the risk of deep fakes.
>> That's the deep fake right there, man.
>> Is changing.
>> Hold on. Hold on. That was good.
>> It is a fastmoving field and we have to
start thinking seriously and carefully
about the threat of misinformation.
>> Okay, good. And one more. We are living
through an unprecedented time where we
are relying more and more on the
internet for information. For
information that affects our health, our
societies, our democracies, and our
economies.
>> Can I hear number one again?
>> Yep. You're a little less sure than you
were a minute ago.
>> Yeah.
>> And if you invite me back next year,
almost certainly everything will have
changed. Uh the nature of creation of
deep fakes, the risk of deep fakes, and
the detection of deep fakes is changing.
I think it's the first one still.
>> I got it right.
>> Yeah.
>> Yeah. I struggled with it, by the way.
Honestly, I couldn't remember. I'm from
the future.
>> You're the time traveler, it turns out.
>> Wow. Well, you know what? I So, I I
started my media work in audio, right?
Being a voice actor. And and very
quickly, I was able to pick up on music
and commercials and movies where they
were dropping in uh you know, pickups.
The the reason I figured it out is
there's a difference in the background
noise. Like one had more reverb than the
other. Um which is how I I I then
remembered it. But you got to admit all
three of them sound like me.
>> Oh, they all do. They all sound like
you.
>> Oh, and by the way, so not only
>> Let let me tell you what has gotten me
recently is I'll get these uh social
media announcements. Oh, there's a new
song by Tupac and Eminem. And I start
listening to it and halfway in I'm like,
no, this is this is Yeah. But at the
beginning,
>> it's coming from music. It's coming from
music as well, by the way. So, this is
one of my favorite videos, by the way.
Let me just show this to you.
>> And if you invite me back next year,
almost certainly everything will have
changed. The nature of the creation of
deep fakes, the risk of deep fakes,
real.
And your mouth is doing it. I don't
speak Japanese.
Doesn't
it sound like?
>> Yes, it does.
>> I know. So, now I can do full-blown
video.
>> Any language. Any language. By the way,
here's what's really cool about this.
Here's a really cool application. I like
foreign films a lot, but I can't stand
>> bad lip syncing. It makes me crazy. But
you don't need it anymore.
>> You don't need it.
>> We're now going to make videos in any
language you want, and it's going to be
perfect.
>> What? How did you do that? How
>> This is also a commercial software. Um,
you upload a video. say that you have
permission to do it and you say, "Please
translate this into Japanese, Korean,
Spanish, French, German, anything you
want."
>> It's amazing.
>> That is nuts. But the fact that the
mouth change to to voice the word,
>> by the way, the way this works, this is
really amazing, is you upload a video of
you talking. And what it does is it
takes the audio and transcribes it. So,
it goes from audio to words
>> and then it translates from English to
Spanish
>> and then it synthesizes a new audio in
Spanish and then it puts that audio back
into the video. Every one of those is an
AI system by the way and it does that in
about 3 minutes.
>> Wow.
>> And it's amazing. So, if you wanted to
take this podcast
>> and distribute it in Spanish, French,
German,
>> Yeah.
>> upload it.
>> Man, I'm just hitting India, China,
Southeast Asia.
>> Two and a half billion people. Done.
Done. 10 cents each, we're good to go.
>> This podcast is from the producers of
Nova. Nova is supported by Carile
Companies, a manufacturer of innovative
building envelope systems. With
buildings responsible for over a third
of total energy use and energy demand on
the rise, Carile's mission is to meet
the challenge head on. Carile's energy
efficient solutions are built to reduce
strain on the grid. For example,
Carile's Ultra Touch Denim Insulation
made from sustainable recycled cotton
fibers delivers high thermal performance
for improved energy efficiency while
being safe to handle and easy to
install. Made from 80% recycled denim,
Ultraouch diverts nearly 20 million
pounds of textile waste from landfills
each year. Operating nearly 100
manufacturing facilities across North
America, Carile is working towards a
more sustainable future. Learn more at
carile.com.
>> We have systems now to detect AI text,
AI audio, AI images, AI video. What is
the give us the nuts and bolts of how
you detect these fakes? My bread and
butter as an academic and also now as a
chief science officer of get real is to
build technology to distinguish what is
real from what is fake. Okay. And so
here there is some AI there. There are
some more classic techniques and I'm
going to I want to talk if I may about
my favorite one and I think this may
resonate with you as a physicist. So
what you have to understand about um
generative AI deep fakes is that it is
fundamentally learning how to generate
images, audio and video by looking at
patterns in billions and billions of
images, audio and video.
>> Okay?
>> But it doesn't know what a lens is. It
doesn't know what the physics of the
world is. It doesn't know about
geometry. It doesn't know about the
physical world. It's not recreating this
thing that we you and I are in right
now.
>> Right?
>> Um take any image outdoors. sunny day
here in uh in Virginia, go outdoors and
because the sun is shining, you will see
shadows all over the place.
>> Yeah.
>> And those shadows have to follow a very
specific law of physics, which is that
there's a single dominant light source,
the sun,
>> and it is giving rise to all those
shadows.
>> So, we have geometric techniques that
can say given a point on a shadow and
the part of the object that is casting
it, tell me where the light is that's
consistent with that. And we can do that
not once, not twice, not three times,
but as many times for shadows that we
find,
>> for every shadow in the
>> every shadow. And if we find that they
are not converging on a single light
source, the sun, then we have a
physically implausible scene.
>> Yeah,
>> it seems like that would be easy for AI
to figure out.
>> You would you would think, but here's
why I can't. Because what I described to
you is a three-dimensional process
that's happening in the
three-dimensional world, but the AI
lives in 2D.
>> It lives in a two-dimensional world. And
reasoning about the three-dimensional
world is not something it does. Now, we
can sort of fake it pretty well the way
artists fake it. Right.
>> Right. Lots of things in paintings are
not physically plausible, but our visual
system doesn't really care. We're
looking at a pretty picture.
>> So, that's one of my favorite
techniques. Um, here's another one that
I love.
>> Um, go outside and well, you shouldn't
do actually do this, but stand on the
railroad tracks. I don't actually advise
doing that. I did this the other day
with one of my students. I'm like, what
are you doing standing on the railroad
tracks? I wanted to take a picture of
railroad tracks. And the reason I want
to take a picture of railroad tracks is
that when you're standing on the
railroad tracks, those railroad tracks
of course are parallel in the physical
world and they are remain parallel as
long as the track continues going. But
if you take a picture of it, those uh
train tracks will converge to what's
called a vanishing point. This is a
notion that Renaissance painters have
understood for hundreds of years. And
why is that? It's because when you
photograph something, um the size on the
image sensor is inversely proportional
to how far it is from me. So as the
train tracks recede, it looks like
they're converging. Right? This is
called projective geometry, a vanishing
point. It's a very specific geometry.
And this is true of the parallel lines
on the top and bottom of a window, on
the sides of a building, on a sidewalk,
anything that you have a flat surface
like this table that we're at.
>> Take a photo of this of this table and
the all these parallel lines will
converge to a vanishing point.
>> Right?
>> So we can make those measurements in an
image. And when we find deviations of
that,
>> something is physically implausible.
Your the image is violating geometry.
>> Okay.
>> All right. Let me move to a sort of
different side of it. This is actually
one of my favorite techniques is when
you go to your favorite AI system and
you ask it to make an image, um, it will
create all the pixels, but then it has
to bundle it up into a JPEG image or a
PNG image or some format,
>> right?
>> And it actually does that in a very
specific way. And so here's an analogy.
When I buy something from an online
retailer, there's the product I get, but
that product is also packaged in a box.
Yes. And different retailers have
different ways of doing it. Apple has a
very specific way of doing beautiful
packaging, right? Other retailers, you
know, just shove it in a box and send it
off.
>> So the packaging when I create an image
on OpenAI or on Enthropic or on
MidJourney, all these different
generators, they package it up
differently. M
>> um and it's different than the way my
phone packages up the pixels and it's
different than the way Photoshop
packages up the pixels. So when we get
an image or an audio or video for that
matter, we can look at the underlying
package and saying is this a packaging
that is consistent with OpenAI or
Enthropic or a camera or whatever it is.
And so
>> so it doesn't have package emulators.
>> It does not. It doesn't know about
because it doesn't care. Why would you
care about it? I'm the only person in
the world who probably cares about this.
You certainly don't care how it's
packaged because what do you do? You
open the package, you throw it away and
you got your product, the image, right?
So, we can look at the packaging.
>> U there's a whole another set of
techniques um that so everything I've
described is sort of after the fact,
right? Right. You wait for the content
to land on your desk and you start doing
these analyses, right? There's a whole
another set of techniques that are what
are called active techniques. So, Google
recently announced that every single
piece of content that comes from their
generators, image, audio, or video will
have what's called an imperceptible
watermark. So, think we don't use
currency that much anymore, but take
your $20 bill out of your wallet and
hold it up to the light and you'll see
all kinds of watermarks that prevent or
make it very difficult to counterfeit.
>> So, what Google has done is they have
inserted an invisible watermark into
images, audio, and video at the point of
creation that says, "We made this."
>> Yeah.
>> And then when I get that piece of
content, I have a specialized piece of
software because we over at Get Real
have a relationship with Google that
says, "Is there a watermark in there?"
It's it's a signal and you can't see it,
>> right? And the adversary can't see it,
but I can see it. So, that's really
cool. And by the way, if this comes into
the phones, so if Apple decides, we're
going to watermark every single piece of
content that is natural,
>> I've got a signal that is built in,
right?
>> So, we've got lots of different
techniques from things that we rely on
third parties like the Googles of the
world to measurements that we can make
in an image, a video, or an audio. I'll
give you one of my favorite audio ones,
by the way. So, if you're listening to
this, you won't be able to see us, but
if you're watching this on YouTube, you
will know we're in a really nice studio
>> and there are soft walls around us and
we have really nice microphones. And so,
the amount of reverberation,
>> yeah,
>> that you hear is quite minimal. We're
This audio is going to sound really good
because you guys are pros here, right?
But the amount of reverberation is
dependent on the physical geometry
around us, how hard those surfaces are,
and that should be fairly consistent
over an audio,
>> right? But what you see with AI
generation is you see inconsistencies
in the microphone and the reverberation
because it doesn't it's not physically
recording these things.
>> So even in a single
>> recording you'll see modulation that are
quote unquote unnatural. Yeah. So what a
lot of what we do is look for patterns
you expect to see that mimic the
physical world. Okay. Now, and then I
talked about the active techniques, the
watermarking, and then there's a whole
another set of techniques that I'm going
to talk about a little, but not a lot,
and you'll understand in a minute why
not. So, the other side of of what we do
is we try to understand the tools that
our adversary uses. So, if you're using
an open AI or anthropic or some open
source code, we actually go into the
well, we can't do this for open AI, but
for anything that's open source. M
>> um so there's there are so-called um uh
um face swap deep fakes where you can
take somebody's face eyebrow to chin
cheek to cheek and replace it with
another face
>> and these are all open source libraries.
We can dig into the code and we can see
okay what are they doing? All right the
first thing they're doing is this and
then they do this and then they do this
and then they do this and then we'll say
ah that second step should introduce a
very specific artifact.
>> Um so I'll give you one example but not
more than one. Yeah.
>> So, one of the things that a lot of
these swap faces do is they put a a
square bounding box around the face.
>> They pull the face off, they synthesize
a new face, and then they put it back.
>> But when they put it back, it's with a
bounding box. And they do it very well.
>> Yeah.
>> You can't see it, but we know how to go
into the to the video and discover that
bounding box that was there.
>> Wow.
>> Right. So, that's an example where we
reverse engineering because we
understand how the adversary has made
something. Now we have lots of other
ones which I don't want to tell you
about.
>> Yes, I understand why.
>> You can see now cuz it's adversarial.
>> Exactly. Right. Right. Right. Man, it it
sounds very systematic. I I have a a
decent understanding now if I want to
make a lab to do this some techniques to
do it. But you know the average person
out there isn't a scientist. How can
people how can I how can my mother
identify real from the fake in the world
of AI?
>> Yeah. Yeah. They can't. This is the
reality of where we are right now. And
this is important to understand because
I don't want you to walk away from this
podcast thinking, "Okay, I understand a
little bit now. Now, when I'm scrolling
through, you know, X or Blue Sky or
Facebook or Instagram, I'm going to be
able to tell." You won't.
>> You won't be able to tell. And even if I
could tell you something today that was
reliable, six weeks from now, it will
not be reliable and you'll have a false
sense of security.
>> Right?
>> So, I get this question a lot and the
thing you have to understand is this is
a hard job. It is really hard to do this
and it's constantly changing and the
average person doom scrolling on social
media cannot do this reliably. You can't
do it reliably. I can barely do it
reliably and this is what I do for a
living.
>> So here's some some things.
>> Stop for the love of God getting your
news and information from social media.
Yeah, this is not what it was designed
for and it's not good. If you want to be
on social media for entertainment,
that's fine. I don't I don't care. I
don't think you should, by the way, but
I don't care.
>> But this is not where you get news
information. You know where you get it?
Things like this,
>> right?
>> Right. You get it from news outlets that
have standards, that have serious, smart
journalists who work hard to get you
information. And we have to come back to
some sense of reality of where we get
our news.
>> Man, rigor around determining truth and
measuring uncertainties is not something
that we're generally taught. And when
you become a scientist, like it it is
unnatural.
>> Yeah.
>> It's an unnatural way to think.
>> That's right. I agree. Yeah. And look,
you know, you and I have both fallen in.
You were telling me in the green room
before this, right? You heard the story,
you thought it was true, you assumed it
was true. Somebody called you out on it,
you went and figured out like, "Oh, God,
I was wrong."
>> Yeah. Yeah.
>> Right. And now imagine that at scale of
the thousands of posts you're seeing.
>> The reason why I thought it was true was
because everybody else was saying,
>> exactly. And that's what social media
is. Everybody is saying the same thing.
Millions of views. Oh, this must be
true. But that's the way social media
works. We have to get back to getting
reliable information from reliable
sources. That's number one.
>> Number two,
>> and I tell you the other thing about it
though is that even though I I I assumed
it was true because everyone else
assumed it was true and these others
were scientists just like me,
>> is I still knew that I didn't know
>> that I had not confirmed it for myself.
And I think that is where the average
person can can, you know, if we know the
difference between knowing and not
knowing,
>> then you can check yourself. Even if
everybody's saying it, you don't know
that that's the truth. I I agree and and
this gets to number two in a you said it
in a nicer way than I was going to, but
understand that the business model of
social media is to draw your time and
attention by feeding you content that
outrages you and engages you to deliver
ads to get you to buy stuff you don't
need.
>> Yeah.
>> And so understand that you're being
manipulated,
>> right? And that you are being fed
information that the social media
companies believe you are going to
engage with. And first of all, that
should make you angry that you are being
manipulated. But you're 100% right is
that, you know, we live in these
distorted bubbles when it comes to
social media.
>> And it's very easy to forget
>> that you actually don't know what is
happening in Gaza, in Ukraine, in Texas,
in Los Angeles. You just think you do.
>> And that is incredibly dangerous. It
gives you this false sense of security.
>> And so it comes back to what what I say
is you got to have some humility. M
>> you have to have humility that this is a
complicated world. It is fastm moving
and you have a responsibility to get
reliable information because not only
are you being lied to and being deceived
and making decisions on bad information,
you're also spreading that bad
information. So you are actually part of
the problem now because when you like
and share and send this to your friends,
>> you are now a carrier of dis of
disinformation. So, we have these AIs
that do things for us
>> and we're sort of managing them.
>> Yeah, that's right.
>> And you know, I might write something
and I like, "Oh, give me an edit on
this." Right. And then it comes back and
it'll tighten it or whatever.
>> But you can move to a point where you're
like, "Give me the first draft."
>> Yeah. That's right. Yeah. And it's
coming.
>> And you get to the point where it's
like, "Okay, do it."
>> Yeah. Start responding to my emails. I
don't even have to read my emails
anymore. Right. And by the way, there's
a really weird future that you could
imagine where emails are being sent by
each of our agents and we're hanging out
in the beach. I mean, what are we doing?
>> Exactly. That's the point. It's almost
as if we are making ourselves obsolete
in many, you know, we're not you need a
human being to do uh your um to build
your house, right? You need humans there
swinging hammers. AI can't do that. But
yeah,
>> well AI robots, right?
>> That's right. That's right. But this is
this is sort of the ultimate joke in
some ways or the the the irony of all
this is
>> I think if you go back uh 50 years, what
were people worried about? That we were
going to take blue collar jobs away,
manufacturing jobs, physical labor jobs,
that we were going to build robots to do
those jobs. What did we end up doing? We
took out the white collar jobs. We took
out those highpaying computer science
jobs. We took out those jobs that AI is
now doing better than most humans,
>> right?
>> And that is a that's a weird world. I
can tell you I I am on the campus almost
every day
>> and there is a lot of anxiety among
students about what the future holds for
them. Are there going to be jobs?
Because you're seeing unemployment go up
in these high historically
>> in my own house. My son, he's 20, he's a
senior in college this fall. Guess what
his major is? Computer science.
>> Yeah. And he's struggling. I'm I
Everybody is. But I can tell you, so I'm
at UC Berkeley, one of the top, let's
say, five CS uh programs in the world.
And our students typically had five
internship offers throughout their first
four years of college. They would
graduate with exceedingly high salaries,
multiple offers. They had to run the
place.
>> That is not happening today. They're
they're happy to get one job offer. So
something is happening in the industry.
I think it's a confluence of many
things. I think AI is part of it. I
think there's a thinning of the ranks
that's happening part of it. But
something is brewing and for for people
like your son by the way who four years
ago were promised
>> right
>> go study computer science it's going to
be a great career it is future proof
that changed in four years
>> that is astonishing and I get this
question by the way from students all
the time is
>> how do I prepare for this
>> yeah exactly
>> and honestly I'm sort of at a loss my
best advice and I don't know if it's
good advice by the way
>> is I used to tell people this is what I
used to tell people is you want a broad
education you should know about physics
and language and history and philosophy,
but then you have to go deep
>> like deep deep deep deep into one thing.
Become really really good at one thing.
>> Now, I think I'm telling people be good
at a lot of different things because we
don't know what the future holds.
>> You need options
>> and you need options. And so depth is in
some ways less relevant, particularly if
you know how to harness the power of AI
to get the depth that you need in a
particular area. But if you have a broad
knowledge, I think you're probably today
more futurep proof than if you're very
narrow in one area. And the best line I
heard about this was this is in the in
the the framing of the legal system is
that um I don't think AI is going to put
lawyers out of business, but I think
lawyers who use AI will put those who
don't use AI out of business. And I
think you can say that about every
profession.
>> So I think two things are going to
happen is that you're going to have to
learn like every technology how to
harness the power of AI. And that was
true of computers. That was true of the
internet. That was true of everything.
>> Yeah.
>> So, but
>> because of how powerful the AI systems
are, it is absolutely going to reduce
the workforce.
>> And I think the big question here is the
question always that happens with
disruption of technology is we are going
to eliminate and reduce certain jobs and
the question is do we create new jobs
and what do they look like? Exactly. And
I don't think anybody knows the answer
to that right now.
>> How is AI going to affect bluecollar
jobs? Is AI going to affect blue collar
jobs? We have this vision of white
collar jobs
>> but there could be an effect. What do
you think?
>> Yeah, it's a great question. So now
let's come to blue collar.
>> So where is it coming first? So here's
what I can tell you. It's coming for the
self-driving cars. It's coming for
drivers
>> taxis
>> cuz go to San Francisco.
>> Right.
>> Right. And you can get a a car that will
that is self-driving. It's weird. By the
way, you go to San Francisco and just
stand on a street corner and look at how
many cars drive by you without a driver.
>> What ab
Resume
Read
file updated 2026-02-13 12:57:35 UTC
Categories
Manage