Transcript
GMoOCKkcd_w • How to Detect Deepfakes: The Science of Recognizing AI Generated Content
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/novapbs/.shards/text-0001.zst#text/1087_GMoOCKkcd_w.txt
Kind: captions
Language: en
We have systems now to detect AI text,
AI audio, AI images, AI video. What is
the give us the nuts and bolts of how
you detect these fakes?
>> So, what you have to understand about um
generative AI deep fakes is that it is
fundamentally learning how to generate
images, audio, and video by looking at
patterns in billions and billions of
images, audio, and video.
>> Okay?
>> But it doesn't know what a lens is. It
doesn't know what the physics of the
world is. It doesn't know about
geometry. It doesn't know about the
physical world. It's not recreating this
thing that we you and I are in right
now.
>> Right.
>> Um take any image outdoors. Sunny day
here in uh in Virginia, go outdoors and
because the sun is shining, you will see
shadows all over the place.
>> Yeah.
>> And those shadows have to follow a very
specific law of physics which is that
there's a single dominant light source,
the sun
>> and it is giving rise to all those
shadows. So we have geometric techniques
that can say given a point on a shadow
and the part of the object that is
casting it tell me where the light is
consistent with that
>> and we can do that not once not twice
not three times but as many times for
shadows that we find
>> for every shadow in the
>> every shadow. And if we find that they
are not converging on a single light
source the sun then we have a physically
implausible scene.
>> Yeah.
>> It seems like that would be easy for AI
to figure out. You would think, but
here's why I can't. Because what I
described to you is a three-dimensional
process that's happening in the
three-dimensional world, but the AI
lives in 2D.
>> It lives in a two-dimensional world. And
reasoning about the three-dimensional
world is not something it does. Now, it
can sort of fake it pretty well. The way
artists fake it,
>> right? Lots of things in paintings are
not physically plausible, but our visual
system doesn't really care. We're
looking at a pretty picture.
>> So, that's one of my favorite
techniques. Um, here's another one that
I love. Um, go outside and well, you
shouldn't do actually do this, but stand
on the railroad tracks. I don't actually
advise doing that. I did this the other
day with one of my students. I'm like,
what are you doing standing on the
railroad tracks? I wanted to take a
picture of railroad tracks. And the
reason I wanted to take a picture of
railroad tracks is that when you're
standing on the railroad tracks, those
railroad tracks of course are parallel
in the physical world and they are
remain parallel
>> as long as the track continues going.
But if you take a picture of it, those
uh train tracks will converge to what's
called the vanishing point. This is a
notion that Renaissance painters have
understood for hundreds of years. And
why is that? It's because when you
photograph something, um, the size on
the image sensor is inversely
proportional to how far it is from me.
So, as the train tracks recede, it looks
like they're converging. Right? This is
called projective geometry, a vanishing
point. It's a very specific geometry.
And this is true of the parallel lines
on the top and bottom of a window, on
the sides of a building, on a sidewalk,
anything that you have a flat surface
like this table that we're at, right?
>> Take a photo of this of this table and
the all these parallel lines will
converge to a vanishing point,
>> right?
>> So, we can make those measurements in an
image and when we find deviations of
that,
>> something is physically implausible.
Your the image is violating geometry.
>> Okay?
>> All right. Let me move to a sort of
different side of it. This is actually
one of my favorite techniques is when
you go to your favorite AI system and
you ask it to make an image, um it will
create all the pixels, but then it has
to bundle it up into a JPEG image or a
PNG image or some format,
>> right?
>> And it actually does that in a very
specific way. And so here's an analogy.
When I buy something from an online
retailer, there's the product I get, but
that product is also packaged in a box.
Yes.
>> And different retailers have different
ways of doing it. Apple has a very
specific way of doing beautiful
packaging, right? Other retailers, you
know, just shove it in a box and send it
off.
>> So, the packaging when I create an image
on Open AI or on Enthropic or on
MidJourney, all these different
generators, they package it up
differently.
>> Um, and it's different than the way my
phone packages up the pixels and it's
different than the way Photoshop
packages up the pixels. So when we get
an image or an audio or video for that
matter, we can look at the underlying
package and saying is this a packaging
that is consistent with OpenAI or
Enthropic or a camera or whatever it is.
And so
>> so it doesn't have package emulators.
>> Yeah, it does not. It doesn't know about
because it doesn't care. Why would you
care about it? I'm the only person in
the world who probably cares about this.
You certainly don't care how it's
packaged because what do you do? You
open the package, you throw it away, and
you got your product, the image, right?
So we can look at the packaging. U
there's a whole another set of
techniques um that so everything I've
described is sort of after the fact
right you wait for the content to land
on your desk and you start doing these
analyses right there's a whole another
set of techniques that are what are
called active techniques so Google
recently announced that every single
piece of content that comes from their
generators image audio video will have
what's called an imperceptible watermark
>> so think we don't use currency that much
anymore but take your $20 bill out of
your wallet and hold it up to the light
and you'll see all kinds of watermarks
that prevent or make it very difficult
to counterfeit.
>> Yeah.
>> So what Google has done is they have
inserted an invisible watermark into
images, audio and video at the point of
creation that says we made this.
>> Yeah.
>> And then when I get that piece of
content, I have a specialized piece of
software because we over at Get Real
have a relationship with Google that
says
>> is there a watermark in there? It's it's
a signal and you can't see it.
>> Right? And the adversary can't see it,
but I can see it. So that's really cool.
And by the way, if this
decides we're going to watermark every
single piece of content that is natural,
I've got a signal that is built in,
right?
>> So, we've got lots of different
techniques from things that we rely on
third parties like the Googles of the
world to measurements that we can make
in an image, a video, or an audio. I'll
give you one of my favorite audio ones,
by the way.
>> So, if you're listening to this, you
won't be able to see us, but if you're
watching this on YouTube, you will know
we're in a really nice studio. Yeah. And
there are soft walls around us and we
have really nice microphones and so the
amount of reverberation
>> that you hear is quite minimal. We're
this audio is going to sound really good
because you guys are pros here, right?
But the amount of reverberation is
dependent on the physical geometry
around us, how hard those surfaces are
and that should be fairly consistent
over an audio,
>> right? But what you see with AI
generation is you see inconsistencies
in the microphone and the reverberation
because it doesn't it's not physically
recording these things.
>> So even in a single
>> recording you'll see modulations that
are quote unquote unnatural. Yeah. So
what a lot of what we do is look for
patterns you expect to see that mimic
the physical world. Okay. now and then I
talked about the active techniques the
watermarking and then there's a whole
another set of techniques that I'm going
to talk about a little but not a lot and
you'll understand in a minute why not so
the other side of of what we do is we
try to understand the tools that our
adversary uses so if you're using an
open AI or anthropic or some open source
code we actually go into the well we
can't do this for open AI but for
anything that's open source
>> um so there's there are so-called um uh
um face swap deep fakes where you can
take somebody's face eyebrow to chin
cheek to cheek and replace it with
another face. And these are all open
source libraries. We can dig into the
code
>> and we can see okay, what are they
doing? All right, the first thing
they're doing is this and then they do
this and then they do this and then they
do this and then we'll say ah that
second step should introduce a very
specific artifact.
>> Um so I'll give you one example but not
more than one. So one of the things that
a lot of these swap faces do is they put
a a a square bounding box around the
face.
>> They pull the face off. They synthesize
a new face and then they put it back.
>> But when they put it back, it's with a
bounding box and they do it very well.
>> Yeah.
>> You can't see it, but we know how to go
into the to the video and discover that
bounding box that was there.
>> Wow.
>> All right. So, that's an example where
we reverse engineering because we
understand how the adversaries made
something. Now, we have lots of other
ones which I don't want to tell you
about.
>> Yes. I understand why. You can see now
cuz it's adversarial.
>> Exactly. Right. Right. Right. Man, it it
sounds very systematic. I I have a a
decent understanding now if I want to
make a lab to do this some techniques to
do it. But you know the average person
out there isn't a scientist. How can
people how can I how can my mother
identify real from the fake in the world
of AI?
>> Yeah. Yeah. They can't. This is the
reality of where we are right now. And
this is important to understand because
I don't want you to walk away from this
podcast thinking, "Okay, I understand a
little bit now. Now, when I'm scrolling
through, you know, X or Blue Sky or
Facebook or Instagram, I'm going to be
able to tell. You won't.
>> You won't be able to tell. And even if I
could tell you something today that was
reliable, six weeks from now, it will
not be reliable and you'll have a false
sense of security.
>> Right?
>> So, I get this question a lot. And the
thing you have to understand is this is
a hard job.
>> It is really hard to do this. And it's
constantly changing. And the average
person doom scrolling on social media
cannot do this reliably. You can't do it
reliably. I can barely do it reliably
and this is what I do for a living.