Interview: Deepfake Detection and the Future of AI with Hany Farid | Particles of Thought
nG2_GhNdTek • 2025-08-26
Transcript preview
Open
Kind: captions Language: en We have to put our hands back on the steering wheel and we have to start getting serious about how do we put guard rails on the system because what we know is that if you unleash Silicon Valley they will burn the place to the ground to get to the finish line first and we've got to start putting guardrails in place and the thing is is that this is a borderless problem right we this can't be an only US or an only EU or an only we have got to start thinking about this globally >> and I don't think we are doing that very [Music] [Applause] [Music] Hello everybody. I sat down with Hani Farid, one of the leading voices in AI research. He's a professor at UC Berkeley where he works on digital forensics and misinformation, especially things like deep fakes, image analysis, and how we perceive fake content. He's also the chief science officer at GetRal Labs, a company that focuses on the authentication of digital media, telling us what's real out there and what's fake. So, yeah, he's busy, but he made time for us. This isn't your standard AI interview, cuz I know y'all have heard too much about AI and you're tired of that same old conversation. This time, we talk about how AI actually works and how he can detect the fake stuff. and of course what we all want to know, what he thinks the future holds. If you enjoy the show, I'd love your help spreading the word. So, take a moment to rate, review, or leave a comment. And don't forget to subscribe so you don't miss anything. Your support really helps us reach new audiences. So, let's go. Honey, welcome to Particles of Thought. >> It's great to be here, Hickey. >> All right, man. So, look, this isn't going to be like your normal interview because this first question I got for you >> Yeah. >> is something that the people need to know. >> Good. >> All right. So, >> I can't tell you I'm a little nervous already, but go ahead. >> All right. So, listen. You're you're an expert on AI >> and you specialize in identifying deep fakes. >> So, there's been three occurrences in recent history that has everybody in a tizzy. >> Yeah. >> The first one was the movie The Matrix. >> Yeah. The second one is when physicists came up with their holographic theory which seems to indicate that life could be a simulation. >> Yeah. >> All right. So, and the third one now is AI. >> Yeah. >> And the question everyone has is >> is reality a deep fake? >> And since you are an expert in uncovering deep fakes, are we cold, man? Is reality what we think it is? And if it is a deep fake, would a deep fake expert have an idea of how to uncover that? >> No. I mean, if this was all a simulation, my job would be a lot easier, honestly, because then you sort of give up, right? There's there's no more reality anymore, I think. But this is sort of where we are, Hakee. >> Yeah. >> Is we are now questioning everything, not just what I see on social media, but our our existence. we are starting to question and our existence today and yesterday and in the future >> and it does feel to me I've been I've been thinking about these problems for 25 years that we went from if I was having a podcast with you 25 years ago your questions would have been of the form hey let's talk about Photoshop and how people splice together two images and maybe we would talk about the Lee Harvey Oswald photo or the moonlanding photo and today a perfectly reasonable question is let's talk about the nature of reality and that's happened in 25 years. Wow. Imagine the next 25 years, the kind of conversations we're going to be having. So, to get back to your question, I don't know. I honestly don't know. And I also don't know where this latest AI boom is taking us. And I I don't think anybody knows honestly. But here's what I can tell you. >> If you look at the personal computer revolution, that took roughly 50 years to unfold, which at the time felt really fast. >> Where do you start from? like >> I'm going to start in 19 let's start in 1950 1945 and let's go to about 2000 when more than half of US um homes had uh a personal computer >> then you look at the internet revolution right from the beginning of the htt http or html protocol Tim Berners Lee to about half the world's population being online that was 25 years the mobile revolution was less than 10 years and the AI revolution has been well the AI revolution of course started in 1950 but the one we are experiencing now was about 2 to 3 years we have gone from 0 to 100 miles an hour where we are now talking about existential threats of AI. We're talking about 50% of jobs being eliminated in the next 5 years. We are talking about general artificial intelligence. We're talking about the Terminator. And we would not have had this conversation 5 years ago. And by the way, on top of all of that, we are don't even know what's real anymore because we consume all of our content from online sources. Online sources have been polluted for a while. out there getting more polluted thanks to AI and suddenly our whole notion of reality is up in the air and that is it's unsettling. I think people feel unanchored and I don't >> I don't know how to help everybody with that. >> Yeah. All right. Well, listen man, let's step back >> because everybody listening may not know what's even meant by AI, what's meant by artificial general intelligence, what's meant by deep fakes. So give us some bearing and foundation and define AI and I will tell you I started in >> big data >> back in 2008. >> Yeah. >> And I was doing what was called machine learning, right? Classified I mean supervised and unsupervised learning and I was working on data sets in astronomy and astrophysics. And it was for the Vera Rubin telescope which has just debuted its its images this year. >> And I was being told, hey, you're building the software infrastructure for analyzing that data now. >> Yeah. >> No, I wasn't. >> No, I wasn't cuz AI didn't exist. So >> good. >> I think of machine learning as just statistics. Define AI for us. Define deep learning. I mean deep fakes and define general intelligence. >> All right. Let's start with AI because I think you raised a really good point which is that everything we are talking about today is in fact not AI. It is machine learning. It is statistics. Almost everything we see from the chat GPTs of the world to deep fakes which I'll define in a minute to all of the things that we are seeing impacting our day-to-day lives is machine learning. So let me rewind to 1950. 1950 was when the term give or take when Alan Turing and John McCarthy conceived of the term or the concept of AI and the idea was then quite bold given where computers were can we imbue intelligence the way we conceive of intelligence with humans into machines >> and the idea was yes we should be able to do this and I think it's because when we think about our own brain and how we operate it seems like it should be straightforward but of course it's not talk to any neuroscientist and they'll tell you that but for a long time for about 30 years the field of AI struggled to be relevant >> because there was no path to creating human intelligence in a machine >> and then around the 1980s the field splintered and the field of machine learning came about which was what you were just referring to and here the idea is we are not going to imbue intelligence into machines we are going to learn it from data So we are simply going to present to a machine a bunch of data and have it infer the rules and the logic that we think that we have and frankly it failed for about 20 years. So 1980 to about 2000 the field really struggled for relevance and it struggled for relevance for two reasons. One is there was not enough data because there was no internet and there was not enough computing power because there was no Nvidia. >> So what happened at the turn of the millennium right the internet rose. What did the internet give rise to? lots of things, but one of the things it gave rise to is a boatload of data >> from us. Right? The ultimate irony, by the way, is if AI comes for us, we created it by giving it all of our our data to learn from. >> So around the turn of the millennia, around 2000, huge rise in data being pushed online and then of course a huge rise in computing power with things like Nvidia. >> So define Nvidia for the audience for people who don't know. >> Good. So Nvidia is the company of course that makes GPUs, graphical processing units. This is the core computip. Yeah. That is particularly good at the types of computation you need to do machine learning. >> And suddenly we saw this explosion. It really started around 2015 2016 in the last decade has really exploded with phenomenal amounts of data, phenomenal amounts of computing power. And you got to give credit to the statistics and machine learning community. Some real insights on how to do machine learning and how to do the types of things that you did. you were way ahead of the curve in the in the early knots and then we got better at doing that because we had some insights from statistics and physics and mathematics and computer science and engineering. So that is so everything you hear today in my opinion that we talk about AI is machine learning. We've sort of conceded that we're just going to call it AI because that's just the term that has gotten adopted. But underneath it, understand that everything that is happening is pattern matching. >> That you take a bunch of data and you learn the rules implicitly, not explicitly from the data. >> And that's both very powerful and very dumb. Right? It's powerful because you can learn complicated patterns, but it's dumb because you don't know what the rules are. Like it's not learning that gravity is - 9.8 m/s squared. It's simply learning that if you film something falling, this is how it will fall. But it doesn't know what the physics of it are. >> Right? It simply knows this is where the ball will be at any given moment. Okay? >> Now, so that's AI/Machine Learning. >> Um, artificial general intelligence is a term that I don't think anybody knows how to define, but I'm going to give you my definition of it, or at least we don't agree on a definition. So, typically, historically, when you've looked at machine learning and AI systems, they've been bespoke. So, you'll have a system that does medical imaging. You'll have a system that does drug discovery. You'll have a system that drives um self-driving cars. >> The idea of AGI, artificial general intelligence, is that it does everything. >> What humans do, right? >> Um >> or stem cells >> or that's right. Exactly. That's a good example of it. Right. So, so is that a year away, 5 years, 10 years? I don't know. I don't think anybody knows. But there's a lot of speculation about that. I would argue by the way that we have some notion of AGI because if you go over to your favorite large language model like chat GPT yeah or claude you can ask it a lot of different questions about physics about medicine about computer science about televisions anything and so is that AGI I don't know I suppose it depends on how you define it >> that that takes me to the touring test right that was the big so so so what was the touring test is it if you can't tell the difference between when you're talking to a machine or >> Yeah. So you you got to by the way, you got to have respect for Alan Turing. >> Oh, absolutely. >> In the 1950s, this guy was thinking about things that were really outrageous to be thinking about at the time. >> Maybe he was a time travel. >> That's right. That's really funny. Don't don't give anybody any ideas by the way. >> Okay. So what is the Turing test? So when Alan Turing first conceived of this notion of imbuing intelligence in machines, he came up with a mechanism to determine if you have solved that problem. And the idea was that you would have two computer screens uh and behind one computer screen was a human and behind another computer screen was a AI system. You of course couldn't see what was behind it. And you were allowed to interact with it with a keyboard. Okay? And if you and you can ask it questions, you can have a conversation the way you and I are having right now. And if you could not tell the difference, then that machine has passed the so-called touring test and it has what probably today we would call AGI. Okay. So now we've done AI and we've done AGI, right? Okay. So let's talk about deep fakes which is this sort of sliver of all of this. >> Yeah. >> So deep fakes is an umbrella term for using machine learning AI to whole cloth create images, audio and video of things um that have never existed or happened. So for example, I can go to my favorite deep fake generator and say give me an image of Hakee in a studio doing a podcast with Professor Hani Fared >> and actually it would do a pretty good job because you have a presence online. I have somewhat of a presence online. It knows what we look like and it would generate an image that's not exactly this but something like that. Or I can say please I by the way I still say please when I ask AI for for things. One of my students told me that this is a good idea because when the AI overlords come they're going to remember you were polite to them. Ah, >> I actually really like this advice. >> Wait a minute. So, I read an article. >> Yes. It cost tens of millions of dollars. >> The energy ultimate. Yes. Just saying please and thank you. I still do it by the way. And even in my head right there when I was asked when I was I I still in my head say please. >> Well, listen. I have AI connected to my AI, right? And so my AI corrects my AI prompts >> to proper grammar and it's like >> please. It puts please in there. >> I know. And it does cost tens of millions of dollars for that extra token. Okay. >> So, I will ask it for an image of a um of a unicorn wearing a red clown hat um walking down the street at Times Square and it will generate that image. Um I can ask uh generate an audio uh of Professor Hani Fared saying the following, right? >> Um I can generate a video of me saying and doing things I never did. And you can clearly see the power of that technology from a creative perspective. If you and I are having a conversation and in post we said something we didn't mean to, we can just fill it in with AI now. >> Well, here here's the thing that makes me you just mentioned how we're only two three years into this. So, however good it is now, you know, this is the worst it will ever be, >> right? >> So, if you look at the So, I can tell you by the way how good it is. >> So, in addition to being trained as a computer scientist and applied mathematician, I've been somewhat trained as a as a cognitive neuroscientist. And we do perceptual studies. So what we do is we recruit participants. We show them images, audio clips and video. And we tell them half of the things you're going to look at are real. Half of the things are AI generated. We explain to them what AI generated is. We give them examples of that. >> And for images as of last year, people are roughly at chance at distinguishing a real photo from an AI generated photo. >> So what you mean by that is if they were just if you had a a monkey behind a keyboard, >> flipping a coin. >> Flipping a coin. >> Yeah. Yeah. The monkeyy's probably better than you. By the way, I'm I'm going to go off and guess. Um, so with audio, so we play a clip of somebody speaking like you and then we play an AI generated version. They're slightly above chance, not like 65%. >> On image at chance, at audio, slightly better than chance. >> And video, they're a little bit better, but all of those trends are going towards chance. So here's what we know. everything in the next 12 months, 18 months, 24 months, I don't know what the number is, >> it will be indistinguishable to the average person online, right? And that is >> that is a weird world we're living in because think about how much in first of all, the vast majority of Americans now get the the the majority of their information from online sources and unfortunately from social media too. >> And that and because it is so easy to create this content, understand all this is is a text prompt away. I type, "Please give me an image of this, generate this audio, generate this video." There are dozens of services that will do this extremely inexpensive or for free. And you can carpet bomb the internet with fake images of the conflict in uh Gaza. >> Fake images. >> I have seen them too. Fake images of the flood in Texas, fake images and video of the fires in name it across the boards, right? Fake images of people stuffing ballot boxes. Now we have a threat to our democracy. >> Wow. So suddenly our sense of reality coming back to your first very good question is up in the air because I can create whatever reality I want and understand that there's sort of three things happening here when we talk about deep fakes. There's the creation of it. That's what we've been talking about. >> There's the distribution which we democratized 20 years ago. So anybody can >> publish to the world and that's very powerful and very terrifying because there's no editorial standards on social media. And then there's the amplification that we have become so polarized as a society that when you see things that conform to your world view, you are more than happy to click like, reshare. And now you have creation, distribution, amplification. >> Wow. >> That's the ball game, >> right? That's the ball game for spreading massive lies, conspiracies, and disinformation campaigns that affect our global health, our planet's health, our democracy, our economy, everything. Everything. So let's get into how these fakes are generated. So start with images. >> Good. So let's start with images because in some ways it's the easiest one, but all of these have a similar theme. And one of my favorite techniques for generating images called a generative adversarial network or a GAN. And here's how it works. >> Wait a minute. Wait a minute. Adversarial. >> Adversarial. >> So that means that you're fighting your computer. >> Two two computer systems are fighting each other. And this is sort of the genius of this technique. So here's how it works. >> You have two systems. One system's job is to make an image of a person or a landscape or whatever you want. Yeah. And so what it does, it starts by, this is literally true, it just splats down a bunch of random pixels. So I say, generate an image of a of a person and it says, "Okay, here's a bunch of so so think uh the monkeys at the keyboard typing randomly. Let's see if this is Shakespeare, >> right? >> And then it takes that image and it hands it to a second system and it says, is this a face?" And that system has access to millions and millions of images that it scraped from the internet that are faces. >> I see. >> And that system says that thing that you generated doesn't look like these things over here. >> And it gives the feedback to the generator and it says, "Nope, try again. >> Modify some pixels. Send it back to what's called the discriminator. Is it a face? No, try again." And they work in this adversarial loop. So it's like somebody's checking your homework. >> But it it it seems like it could get stuck never getting to a face, >> you would think. And that's what's amazing about the GANs. the gen is that they converge. >> They converge. >> And part of that is the way they they've been trained. But that's what's the genius of this is that the generator is not very smart >> because all it's doing is modifying pixels. And the discriminator is actually quite simple. It's simply saying, does this thing look like these things? And because you pit them against each other in this adversarial game, this sort of amazing thing happens out the other side. >> So here's the question. In on average, how many iterations does it take? And then how much time does that translate to? >> Yeah, that's a real great question. So typically the time is in seconds. So there's two phases. There's you train the GANs. That's a really long process. But then what we call inference, which is that run this thing, it happens in seconds. And the reason it happens in seconds is by the way that is hundreds of thousands of iterations. Wow. But it's on a GPU, which is very powerful and very fast. And then there's these tricks to make it even faster. You start with small images and then you make them bigger over time. So there's these tricks, but it is literally seconds to make that image. >> And what the brilliance of that is the two systems are competing with each other. >> And then this thing that seems like intelligence come out even though it's not. If you think about those two individual components, right? >> They're pretty basic, but then you have this like emergent behavior almost. It's like you know how to generate images of people. That's amazing. >> So let's have a little fun. I understand good >> that you brought me some fakes >> and some real images >> to put to the test. Good >> to see if I can discern the difference. >> So before I I'm going to play for you a couple of audios. Before I do this, let me say I've been doing this for a long time and I've been I'm pretty good at it. I'm pretty good at what I do and I created three audio samples. I'm going to play them for you. >> Wait, are you allowed to say that that you're you're good at what you do? I'll say that. Honey is really good. >> I said pretty good, by the Yeah, he's amazing. >> But this is amazing. This is this is this is a true story, by the way. So, I made three audio clips for you of me talking and you and I have been talking for a little while, so you now know what my voice sounds like. >> And uh I got off the plane and I was in the car coming over here and I wanted to make sure they worked. And I played all three of them and I couldn't tell which one of me was real or fake. I wasn't 100% sure. >> Wow. >> And I do this for a living and it's my voice, >> right? >> So, okay. So, that is Okay. >> So, wait a minute. Which AI did you use? Is this something that you created or something generally available? >> So here's the thing you have to understand about AI is this is so readily available. So here's what I did. I went to a service. It's a commercial service. Um I uploaded I think it was about 3 minutes of my voice. >> I said please um uh please clone my voice. Um and it clones my voice. And by what I mean by that is that it learns the patterns of my voice, what I sound like, the intonation, my cadence, how fast I speak, where I put the pauses. And then I can simply type and have it say anything I want to say. >> And so I'm going to I'm going to read I'm going to have you play I'm going to listen have you listen to three sentences. Okay. >> Um and one of them is fake. I'm going to give you a hint. One of them is fake and two are real. >> Okay. And let's see what you we you can do. Okay. Here we go. >> And in fairness, this is not the best uh speaker, but okay. >> Are there guardrails in our law? >> Ah, good. Uh, so first of all, when I went to do this this service, um, I uploaded my voice and there's a button that says, "Do you have permission to use this person's voice?" And and I did because it was my voice, but I can upload anybody's voice and click a button. >> The laws are very complicated and they actually vary state-to-state and of course internationally. Wow. >> So, there are almost no guardrails on grabbing people's likeness. And even if there were, >> there's, >> you can still do it anyway. >> There's there's no stopping this. There's no stopping it. Okay. All right. number one. Oh, and by the way, the the three U this is part of a talk I gave recently on deep fakes. So, you'll hear a consecutive thing. Okay, ready? >> And if you invite me back next year, almost certainly everything will have changed. Uh the nature of creation of deep fakes, the risk of deep fakes. >> That's the deep fake right there, man. >> Is changing. >> Hold on. Hold on. That was good. >> It is a fastmoving field and we have to start thinking seriously and carefully about the threat of misinformation. >> Okay, good. And one more. We are living through an unprecedented time where we are relying more and more on the internet for information. For information that affects our health, our societies, our democracies, and our economies. >> Can I hear number one again? >> Yep. You're a little less sure than you were a minute ago. >> Yeah. >> And if you invite me back next year, almost certainly everything will have changed. Uh the nature of creation of deep fakes, the risk of deep fakes, and the detection of deep fakes is changing. I think it's the first one still. >> I got it right. >> Yeah. >> Yeah. I struggled with it, by the way. Honestly, I couldn't remember. I'm from the future. >> You're the time traveler, it turns out. >> Wow. Well, you know what? I So, I I started my media work in audio, right? Being a voice actor. And and very quickly, I was able to pick up on music and commercials and movies where they were dropping in uh you know, pickups. The the reason I figured it out is there's a difference in the background noise. Like one had more reverb than the other. Um which is how I I I then remembered it. But you got to admit all three of them sound like me. >> Oh, they all do. They all sound like you. >> Oh, and by the way, so not only >> Let let me tell you what has gotten me recently is I'll get these uh social media announcements. Oh, there's a new song by Tupac and Eminem. And I start listening to it and halfway in I'm like, no, this is this is Yeah. But at the beginning, >> it's coming from music. It's coming from music as well, by the way. So, this is one of my favorite videos, by the way. Let me just show this to you. >> And if you invite me back next year, almost certainly everything will have changed. The nature of the creation of deep fakes, the risk of deep fakes, real. And your mouth is doing it. I don't speak Japanese. Doesn't it sound like? >> Yes, it does. >> I know. So, now I can do full-blown video. >> Any language. Any language. By the way, here's what's really cool about this. Here's a really cool application. I like foreign films a lot, but I can't stand >> bad lip syncing. It makes me crazy. But you don't need it anymore. >> You don't need it. >> We're now going to make videos in any language you want, and it's going to be perfect. >> What? How did you do that? How >> This is also a commercial software. Um, you upload a video. say that you have permission to do it and you say, "Please translate this into Japanese, Korean, Spanish, French, German, anything you want." >> It's amazing. >> That is nuts. But the fact that the mouth change to to voice the word, >> by the way, the way this works, this is really amazing, is you upload a video of you talking. And what it does is it takes the audio and transcribes it. So, it goes from audio to words >> and then it translates from English to Spanish >> and then it synthesizes a new audio in Spanish and then it puts that audio back into the video. Every one of those is an AI system by the way and it does that in about 3 minutes. >> Wow. >> And it's amazing. So, if you wanted to take this podcast >> and distribute it in Spanish, French, German, >> Yeah. >> upload it. >> Man, I'm just hitting India, China, Southeast Asia. >> Two and a half billion people. Done. Done. 10 cents each, we're good to go. >> This podcast is from the producers of Nova. Nova is supported by Carile Companies, a manufacturer of innovative building envelope systems. With buildings responsible for over a third of total energy use and energy demand on the rise, Carile's mission is to meet the challenge head on. Carile's energy efficient solutions are built to reduce strain on the grid. For example, Carile's Ultra Touch Denim Insulation made from sustainable recycled cotton fibers delivers high thermal performance for improved energy efficiency while being safe to handle and easy to install. Made from 80% recycled denim, Ultraouch diverts nearly 20 million pounds of textile waste from landfills each year. Operating nearly 100 manufacturing facilities across North America, Carile is working towards a more sustainable future. Learn more at carile.com. >> We have systems now to detect AI text, AI audio, AI images, AI video. What is the give us the nuts and bolts of how you detect these fakes? My bread and butter as an academic and also now as a chief science officer of get real is to build technology to distinguish what is real from what is fake. Okay. And so here there is some AI there. There are some more classic techniques and I'm going to I want to talk if I may about my favorite one and I think this may resonate with you as a physicist. So what you have to understand about um generative AI deep fakes is that it is fundamentally learning how to generate images, audio and video by looking at patterns in billions and billions of images, audio and video. >> Okay? >> But it doesn't know what a lens is. It doesn't know what the physics of the world is. It doesn't know about geometry. It doesn't know about the physical world. It's not recreating this thing that we you and I are in right now. >> Right? >> Um take any image outdoors. sunny day here in uh in Virginia, go outdoors and because the sun is shining, you will see shadows all over the place. >> Yeah. >> And those shadows have to follow a very specific law of physics, which is that there's a single dominant light source, the sun, >> and it is giving rise to all those shadows. >> So, we have geometric techniques that can say given a point on a shadow and the part of the object that is casting it, tell me where the light is that's consistent with that. And we can do that not once, not twice, not three times, but as many times for shadows that we find, >> for every shadow in the >> every shadow. And if we find that they are not converging on a single light source, the sun, then we have a physically implausible scene. >> Yeah, >> it seems like that would be easy for AI to figure out. >> You would you would think, but here's why I can't. Because what I described to you is a three-dimensional process that's happening in the three-dimensional world, but the AI lives in 2D. >> It lives in a two-dimensional world. And reasoning about the three-dimensional world is not something it does. Now, we can sort of fake it pretty well the way artists fake it. Right. >> Right. Lots of things in paintings are not physically plausible, but our visual system doesn't really care. We're looking at a pretty picture. >> So, that's one of my favorite techniques. Um, here's another one that I love. >> Um, go outside and well, you shouldn't do actually do this, but stand on the railroad tracks. I don't actually advise doing that. I did this the other day with one of my students. I'm like, what are you doing standing on the railroad tracks? I wanted to take a picture of railroad tracks. And the reason I want to take a picture of railroad tracks is that when you're standing on the railroad tracks, those railroad tracks of course are parallel in the physical world and they are remain parallel as long as the track continues going. But if you take a picture of it, those uh train tracks will converge to what's called a vanishing point. This is a notion that Renaissance painters have understood for hundreds of years. And why is that? It's because when you photograph something, um the size on the image sensor is inversely proportional to how far it is from me. So as the train tracks recede, it looks like they're converging. Right? This is called projective geometry, a vanishing point. It's a very specific geometry. And this is true of the parallel lines on the top and bottom of a window, on the sides of a building, on a sidewalk, anything that you have a flat surface like this table that we're at. >> Take a photo of this of this table and the all these parallel lines will converge to a vanishing point. >> Right? >> So we can make those measurements in an image. And when we find deviations of that, >> something is physically implausible. Your the image is violating geometry. >> Okay. >> All right. Let me move to a sort of different side of it. This is actually one of my favorite techniques is when you go to your favorite AI system and you ask it to make an image, um, it will create all the pixels, but then it has to bundle it up into a JPEG image or a PNG image or some format, >> right? >> And it actually does that in a very specific way. And so here's an analogy. When I buy something from an online retailer, there's the product I get, but that product is also packaged in a box. Yes. And different retailers have different ways of doing it. Apple has a very specific way of doing beautiful packaging, right? Other retailers, you know, just shove it in a box and send it off. >> So the packaging when I create an image on OpenAI or on Enthropic or on MidJourney, all these different generators, they package it up differently. M >> um and it's different than the way my phone packages up the pixels and it's different than the way Photoshop packages up the pixels. So when we get an image or an audio or video for that matter, we can look at the underlying package and saying is this a packaging that is consistent with OpenAI or Enthropic or a camera or whatever it is. And so >> so it doesn't have package emulators. >> It does not. It doesn't know about because it doesn't care. Why would you care about it? I'm the only person in the world who probably cares about this. You certainly don't care how it's packaged because what do you do? You open the package, you throw it away and you got your product, the image, right? So, we can look at the packaging. >> U there's a whole another set of techniques um that so everything I've described is sort of after the fact, right? Right. You wait for the content to land on your desk and you start doing these analyses, right? There's a whole another set of techniques that are what are called active techniques. So, Google recently announced that every single piece of content that comes from their generators, image, audio, or video will have what's called an imperceptible watermark. So, think we don't use currency that much anymore, but take your $20 bill out of your wallet and hold it up to the light and you'll see all kinds of watermarks that prevent or make it very difficult to counterfeit. >> So, what Google has done is they have inserted an invisible watermark into images, audio, and video at the point of creation that says, "We made this." >> Yeah. >> And then when I get that piece of content, I have a specialized piece of software because we over at Get Real have a relationship with Google that says, "Is there a watermark in there?" It's it's a signal and you can't see it, >> right? And the adversary can't see it, but I can see it. So, that's really cool. And by the way, if this comes into the phones, so if Apple decides, we're going to watermark every single piece of content that is natural, >> I've got a signal that is built in, right? >> So, we've got lots of different techniques from things that we rely on third parties like the Googles of the world to measurements that we can make in an image, a video, or an audio. I'll give you one of my favorite audio ones, by the way. So, if you're listening to this, you won't be able to see us, but if you're watching this on YouTube, you will know we're in a really nice studio >> and there are soft walls around us and we have really nice microphones. And so, the amount of reverberation, >> yeah, >> that you hear is quite minimal. We're This audio is going to sound really good because you guys are pros here, right? But the amount of reverberation is dependent on the physical geometry around us, how hard those surfaces are, and that should be fairly consistent over an audio, >> right? But what you see with AI generation is you see inconsistencies in the microphone and the reverberation because it doesn't it's not physically recording these things. >> So even in a single >> recording you'll see modulation that are quote unquote unnatural. Yeah. So what a lot of what we do is look for patterns you expect to see that mimic the physical world. Okay. Now, and then I talked about the active techniques, the watermarking, and then there's a whole another set of techniques that I'm going to talk about a little, but not a lot, and you'll understand in a minute why not. So, the other side of of what we do is we try to understand the tools that our adversary uses. So, if you're using an open AI or anthropic or some open source code, we actually go into the well, we can't do this for open AI, but for anything that's open source. M >> um so there's there are so-called um uh um face swap deep fakes where you can take somebody's face eyebrow to chin cheek to cheek and replace it with another face >> and these are all open source libraries. We can dig into the code and we can see okay what are they doing? All right the first thing they're doing is this and then they do this and then they do this and then they do this and then we'll say ah that second step should introduce a very specific artifact. >> Um so I'll give you one example but not more than one. Yeah. >> So, one of the things that a lot of these swap faces do is they put a a square bounding box around the face. >> They pull the face off, they synthesize a new face, and then they put it back. >> But when they put it back, it's with a bounding box. And they do it very well. >> Yeah. >> You can't see it, but we know how to go into the to the video and discover that bounding box that was there. >> Wow. >> Right. So, that's an example where we reverse engineering because we understand how the adversary has made something. Now we have lots of other ones which I don't want to tell you about. >> Yes, I understand why. >> You can see now cuz it's adversarial. >> Exactly. Right. Right. Right. Man, it it sounds very systematic. I I have a a decent understanding now if I want to make a lab to do this some techniques to do it. But you know the average person out there isn't a scientist. How can people how can I how can my mother identify real from the fake in the world of AI? >> Yeah. Yeah. They can't. This is the reality of where we are right now. And this is important to understand because I don't want you to walk away from this podcast thinking, "Okay, I understand a little bit now. Now, when I'm scrolling through, you know, X or Blue Sky or Facebook or Instagram, I'm going to be able to tell." You won't. >> You won't be able to tell. And even if I could tell you something today that was reliable, six weeks from now, it will not be reliable and you'll have a false sense of security. >> Right? >> So, I get this question a lot and the thing you have to understand is this is a hard job. It is really hard to do this and it's constantly changing and the average person doom scrolling on social media cannot do this reliably. You can't do it reliably. I can barely do it reliably and this is what I do for a living. >> So here's some some things. >> Stop for the love of God getting your news and information from social media. Yeah, this is not what it was designed for and it's not good. If you want to be on social media for entertainment, that's fine. I don't I don't care. I don't think you should, by the way, but I don't care. >> But this is not where you get news information. You know where you get it? Things like this, >> right? >> Right. You get it from news outlets that have standards, that have serious, smart journalists who work hard to get you information. And we have to come back to some sense of reality of where we get our news. >> Man, rigor around determining truth and measuring uncertainties is not something that we're generally taught. And when you become a scientist, like it it is unnatural. >> Yeah. >> It's an unnatural way to think. >> That's right. I agree. Yeah. And look, you know, you and I have both fallen in. You were telling me in the green room before this, right? You heard the story, you thought it was true, you assumed it was true. Somebody called you out on it, you went and figured out like, "Oh, God, I was wrong." >> Yeah. Yeah. >> Right. And now imagine that at scale of the thousands of posts you're seeing. >> The reason why I thought it was true was because everybody else was saying, >> exactly. And that's what social media is. Everybody is saying the same thing. Millions of views. Oh, this must be true. But that's the way social media works. We have to get back to getting reliable information from reliable sources. That's number one. >> Number two, >> and I tell you the other thing about it though is that even though I I I assumed it was true because everyone else assumed it was true and these others were scientists just like me, >> is I still knew that I didn't know >> that I had not confirmed it for myself. And I think that is where the average person can can, you know, if we know the difference between knowing and not knowing, >> then you can check yourself. Even if everybody's saying it, you don't know that that's the truth. I I agree and and this gets to number two in a you said it in a nicer way than I was going to, but understand that the business model of social media is to draw your time and attention by feeding you content that outrages you and engages you to deliver ads to get you to buy stuff you don't need. >> Yeah. >> And so understand that you're being manipulated, >> right? And that you are being fed information that the social media companies believe you are going to engage with. And first of all, that should make you angry that you are being manipulated. But you're 100% right is that, you know, we live in these distorted bubbles when it comes to social media. >> And it's very easy to forget >> that you actually don't know what is happening in Gaza, in Ukraine, in Texas, in Los Angeles. You just think you do. >> And that is incredibly dangerous. It gives you this false sense of security. >> And so it comes back to what what I say is you got to have some humility. M >> you have to have humility that this is a complicated world. It is fastm moving and you have a responsibility to get reliable information because not only are you being lied to and being deceived and making decisions on bad information, you're also spreading that bad information. So you are actually part of the problem now because when you like and share and send this to your friends, >> you are now a carrier of dis of disinformation. So, we have these AIs that do things for us >> and we're sort of managing them. >> Yeah, that's right. >> And you know, I might write something and I like, "Oh, give me an edit on this." Right. And then it comes back and it'll tighten it or whatever. >> But you can move to a point where you're like, "Give me the first draft." >> Yeah. That's right. Yeah. And it's coming. >> And you get to the point where it's like, "Okay, do it." >> Yeah. Start responding to my emails. I don't even have to read my emails anymore. Right. And by the way, there's a really weird future that you could imagine where emails are being sent by each of our agents and we're hanging out in the beach. I mean, what are we doing? >> Exactly. That's the point. It's almost as if we are making ourselves obsolete in many, you know, we're not you need a human being to do uh your um to build your house, right? You need humans there swinging hammers. AI can't do that. But yeah, >> well AI robots, right? >> That's right. That's right. But this is this is sort of the ultimate joke in some ways or the the the irony of all this is >> I think if you go back uh 50 years, what were people worried about? That we were going to take blue collar jobs away, manufacturing jobs, physical labor jobs, that we were going to build robots to do those jobs. What did we end up doing? We took out the white collar jobs. We took out those highpaying computer science jobs. We took out those jobs that AI is now doing better than most humans, >> right? >> And that is a that's a weird world. I can tell you I I am on the campus almost every day >> and there is a lot of anxiety among students about what the future holds for them. Are there going to be jobs? Because you're seeing unemployment go up in these high historically >> in my own house. My son, he's 20, he's a senior in college this fall. Guess what his major is? Computer science. >> Yeah. And he's struggling. I'm I Everybody is. But I can tell you, so I'm at UC Berkeley, one of the top, let's say, five CS uh programs in the world. And our students typically had five internship offers throughout their first four years of college. They would graduate with exceedingly high salaries, multiple offers. They had to run the place. >> That is not happening today. They're they're happy to get one job offer. So something is happening in the industry. I think it's a confluence of many things. I think AI is part of it. I think there's a thinning of the ranks that's happening part of it. But something is brewing and for for people like your son by the way who four years ago were promised >> right >> go study computer science it's going to be a great career it is future proof that changed in four years >> that is astonishing and I get this question by the way from students all the time is >> how do I prepare for this >> yeah exactly >> and honestly I'm sort of at a loss my best advice and I don't know if it's good advice by the way >> is I used to tell people this is what I used to tell people is you want a broad education you should know about physics and language and history and philosophy, but then you have to go deep >> like deep deep deep deep into one thing. Become really really good at one thing. >> Now, I think I'm telling people be good at a lot of different things because we don't know what the future holds. >> You need options >> and you need options. And so depth is in some ways less relevant, particularly if you know how to harness the power of AI to get the depth that you need in a particular area. But if you have a broad knowledge, I think you're probably today more futurep proof than if you're very narrow in one area. And the best line I heard about this was this is in the in the the framing of the legal system is that um I don't think AI is going to put lawyers out of business, but I think lawyers who use AI will put those who don't use AI out of business. And I think you can say that about every profession. >> So I think two things are going to happen is that you're going to have to learn like every technology how to harness the power of AI. And that was true of computers. That was true of the internet. That was true of everything. >> Yeah. >> So, but >> because of how powerful the AI systems are, it is absolutely going to reduce the workforce. >> And I think the big question here is the question always that happens with disruption of technology is we are going to eliminate and reduce certain jobs and the question is do we create new jobs and what do they look like? Exactly. And I don't think anybody knows the answer to that right now. >> How is AI going to affect bluecollar jobs? Is AI going to affect blue collar jobs? We have this vision of white collar jobs >> but there could be an effect. What do you think? >> Yeah, it's a great question. So now let's come to blue collar. >> So where is it coming first? So here's what I can tell you. It's coming for the self-driving cars. It's coming for drivers >> taxis >> cuz go to San Francisco. >> Right. >> Right. And you can get a a car that will that is self-driving. It's weird. By the way, you go to San Francisco and just stand on a street corner and look at how many cars drive by you without a driver. >> What ab
Resume
Categories