File TXT tidak ditemukan.
Transcript
2bbSgSIQsac • Why 99.999% of Us Won’t Survive Artificial Superintelligence
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/TomBilyeu/.shards/text-0001.zst#text/1343_2bbSgSIQsac.txt
Kind: captions Language: en In 2023, nearly half of all AI researchers said advanced AI carries at least a 10% chance of causing human extinction. And yet, [music] we're speeding up, not slowing down. My guest today, Dr. Roman Yampolski, is one of the leading voices in AI safety. And when I asked him for the odds that super intelligence wipes out humanity, he said it's high. Once AI becomes smarter than humans in every domain, we will not be able to control it. In today's episode, we talk about the shocking timeline AGI is on, why super intelligence may be much closer than people think, and why the survival of our species could come down to [music] decisions being made right now. If you want to understand the most important technological threat in human history, as well as our biggest opportunity, this is the one episode you cannot miss. So without further ado, I bring you Dr. Roman Yampolski. Where's Chad GPT at right now? Do you consider Chat GBT to be artificial general intelligence? I doubt you'd call it super intelligence, but would you classified as that, or do you still think we're a ways away from something that would qualify? >> So that's a great question. If you asked someone maybe 20 years ago and told them about the systems we have today, they would probably think we have full [snorts] AGI, we probably don't have complete generality. We have it across many domains, but there are still things it's uh not very good at. It doesn't have permanent memory. It doesn't have ability to learn additional things well after it's already been pre-trained and deployed. It can do a certain degree of learning but it's still limited. It doesn't have same capabilities as humans do throughout lifetimes but we're getting closer and closer to where those gaps are closed and uh it's starting to be productive in domains which are really interesting and important science math engineering where it starts to make novel contributions and now top scholars are relying more and more on it in their research. So I think we're getting close to full-blown AGI. Maybe we are at like 50%. But it's hard to judge for sure just how many different subdomains exist is the deciding factor. >> Okay. So one idea that you put forward that's very interesting is like hey I'm an engineer. I love AI but I would like you to keep it very narrow please. What are the things about general AI that become problematic that aren't problematic in narrow AI? So a whole bunch of them. One is testing. How do you test a system capable of performing in every domain? There is no edge cases. Typically, if I'm developing something narrow, very narrow system, I'm just playing tic-tac-toe. I can test if it's making the legal move. I can test zero. I can test 100. I can test all these weird special cases and know if it's behaving as expected. With generality, it's capable of creative output in many domains. I don't know what to expect. I don't know what the right answers are. I don't know how to test it. I can test it for a specific thing. If I find a bug, I fix it. I can tell you I found a problem and it's been resolved. But I cannot guarantee that there are no problems remaining. So basically testing is out the window. uh any type of anticipation of how it's going to act and impact different subdomains. It's creative. So it's just like with a with a human being. I cannot guarantee that another human being is always going to behave. We kind of talked about it. We developed lie detectors. We developed all sorts of tools for trying to show that a human is safe. But at the end of the day because of interaction with environment, other agents, personal changes within the framework, people may betray you. It's exactly the same for those agents. If we concentrate on narrow systems, we are better at testing them and they have limited scope of possibilities. A system only trained to play chess is not going to develop biological weapons. [sighs] >> I don't see actually why that would help you. So, the reason I say that is uh I know I can trust some percentage of humans to be malicious. And so, as long as AI gets more efficient, which it is and will continue to do so, I presume, uh you're going to have a kid in a garage who's going to be able to go, I'm going to optimize this for biological weapons. I don't care about Tik Tok or uh tic-tac-toe. I just want to let's see how dangerous we can make something. And so, they'll be able to do that. So why does narrow AI feel safe to you period? >> It feels uh safer short term. It buys us time. I think sufficiently advanced narrow systems based on neural architectures will also become agentlike and more general as we become more capable. But if the choice is right now, do we race to full-blown super intelligence in two years or do we try to concentrate on solving specific cancers with narrow tools? I think it's a safer choice not to have an arms race towards super intelligence. >> I get that for sure. You're trying to limit your um the scope of all the problems, but when I really start thinking through what are the things that I'm worried about, so one of the big things is just death of meaning. So when AI becomes better than you at everything, uh you run into a huge problem of now I have to like just sort of tell myself a story. You know, I'm like a compared to what an AI can do from an art perspective, for instance, I'm like a grade schooler and so it's hard to get excited about the refrigerator drawings that I can do compared to, you know, what it can do basically instantaneously. Um and so now we have to do a lot of psychological work just to motivate ourselves that we matter um that we're you know our life carries meaning. Um narrow AI will create that same problem. Do you agree with that or do you see a way like oh no when it's you know when that AI is only good at that thing like somehow humans escape the problem of lost meaning. >> Yeah. So I had the same intuition initially, but looking at the data we already have from domains where we got superhuman AI like chess, chess is not dead. In fact, it's more popular than ever. People play online, people play in person, they still enjoy competing with other humans even though they all suck compared to best AI models, right? Nobody's going to be world champion against the machine again. So it seems like it is not a problem for us. And with narrow AIs, there is a chance we'll keep them as tools. You as a human scientist will deploy a tool to find drugs, novel proteins, something. It's not an agent which independently engages with those discoveries. >> Okay, that's very interesting. So, um I don't know that I agree, but I get where you're going with that. Okay, let's talk now about why AGI is the sort of scary um midwife for ASI. Uh are there tests around AGI where we're like, well, if it can't do the following, we're fine. So, for instance, for a long time it looked like AI wasn't going to be able to teach itself. Uh, but I've seen headlines anyway and hopefully you'll tell me that they're not true, but I've seen headlines where it's like now AI is creating the most efficient learning algorithms itself, which if true seems to be the first step down the road of recursive self-learning where it will just completely detach from us and make itself smarter and smarter and smarter. >> We already had examples of AI teaching itself. Selfplay was exactly that. That's how games like go were successfully defeated. A system would play many many many games against itself. The better solutions, better agents would propagate those and after a while without any human data they became superhuman in those domains. You can generate artificial data in other domains. You can use one AI to generate environments, another one to compete in them, and that creates this type of self-improvement. Typically, we start with human data as a seed and grow from there. But there is zero reason to think we cannot do this zero knowledge learning in other domains. You can do run novel experiments in physics and chemistry, discover things from first principles. [snorts] And yeah, we're starting to see AI used to assist in design design of new models, parameters for models, optimization of runs and this process will continue. They already designed new computer chips on which they're going to run. So there is definitely a improvement cycle. It's not fully complete. There are still humans in the loop, a lot of great humans in the loop. But long term, I think all the steps can be automated. >> Okay. And do you think that right now AI already has what it needs um to improve itself or are we still at a point where if all humans stopped that AI would be like oh damn I'm I didn't quite get the thing that I needed. >> So there is a debate about whatever we need another big breakthrough to get to full AGI and super intelligence or maybe multiple breakthroughs or if just scaling what we have is enough. if I just give another I don't know trillion dollars worth of compute to train on and more data will I get to AGI a lot of graphs a lot of patterns suggest yeah it's going to keep scaling we're not hitting diminishing returns some people disagree but based on the amount of investment we see in this industry it seems like people are willing to bet their money that scaling will continue >> where do you come down on that because this feels like when I hear Yan Lun talk from Facebook um He's like, "Dude, LLMs are never going to make novel breakthroughs in physics. They don't understand the world like that. They are literally just guessing the next letter um based on patterns that they see in the data. And so, they're not going to be able to think through these problems. Now, if he's right, it's going to asmmptote and that's that. And you can put as much compute on it as you want and it's just the wrong approach. Um do you think that he's correct and more compute is not the answer or um are you operating just on the well I don't see the asmtote and therefore I assume that it won't >> I think he's not correct on this one. So for one to predict the next term you need to create a model of the whole world because the token depends on everything about the world. You're not predicting random statistical character in a language. You're predicting the next word in a research paper on physics. And to get the right word, you need to have a physical model of the world. I think JAN is known as making certain predictions about what models are capable of. And then within a week, people demonstrate that no, in fact, they can actually do that. So, uh I wish he was right. It would be wonderful if he was right. and we came to a very abrupt stop in capabilities progress and could exploit what we already have for the next decade or so propagating it through the economy. I think there is billions if not trillions of dollars worth of wealth already available with capabilities we haven't deployed. So there is no need to get to the next level as soon as possible. But it doesn't seem like it's the case and I think his uh friends uh core winners of that touring award for machine learning also disagree with him and are very concerned with safety. We'll return to the show in a moment, but first, the average person spends 13 hours a year on hold. And the average company spends millions on call centers [music] that customers still hate. But there is a much better solution. AI call centers. Bland builds AI voice agents that handle your entire call operation. They sound human. They work 24/7. And they actually get cheaper as you scale. They're the only self-hosted voice AI company, so your data never goes to large providers like OpenAI or Anthropic. That way, everything stays on your servers, completely secure. The results speak for themselves. Companies cut costs by over 40% using Bland. And Bland handles it all for you. Customer support, appointment reminders, follow-ups, almost any use case you can think of. If you're a large business, Bland is offering to [music] build a free custom agent for Impact Theory listeners. Just head to bland.ai/agent to get a voice agent trained specifically on your business and your use case for free. There's something about the way that we have structured the brain brains of LLMs where as long as it has access to what I'll call more neurons so it has access to more compute um or theoretically that we get more efficient per GPU neuron in my analogy um that it's going to keep progressing by itself. So, um, if you said it, I didn't quite get the answer. I didn't quite, um, I wasn't able to take it on the answer to whether or not, uh, AI is able to create algorithms for learning that are superior to the ones that it's given. What I heard in your answer was with the algorithms that humans created, it's able to keep making itself better and better at that narrow task as that learning algorithm was defined. But can it fundamentally go, God, the way that you guys want me to learn is really stupid. Here's the algorithm I should be using to learn. And now it starts learning at at just an exponential rate compared to what it's at now. >> I don't think we're quite there yet. I don't think we have full-blown agents. what we have right now are still tools with some degree of agenthood and also it's not capable of recursive self-improvement like compilers can optimize a single pass through your software make it a little faster but they cannot continue this process you cannot feed code for compiler to itself and have it infinitely improve itself that's not where we're at but it seems like that part of automating algorithm design is getting more efficient and I think we'll get there >> give me a number. What are the odds that artificial super intelligence kills us all? >> Uh, pretty high. So, really depends on how soon you expect this to happen. So, short term, we're unlikely to get that level of capability from AI. So, we are probably okay. But once we create true super intelligence, a system more capable than any person in every domain, it's very unlikely we'll figure out how to indefinitely control it. And at that point, if we're still around, it's because it decided for whatever game theoretic reasons to keep us around. Maybe it's pretending to be nice to accumulate more resources before it strikes. Maybe it needs us for something. It's not obvious, but we're definitely not in control and at any point it decides to take us out, it would be able to do so. >> Okay. And if you were going to give us a rough timeline, are you in the two to five years or is this something way off in the future? >> Yeah. So it's hard to predict. The best tool we got for predicting future of technology is prediction markets. And they saying maybe 2027 is when we get to AGI, artificial general intelligence. I think soon after super intelligence follows. The moment you automate science engineering, you get this self-improvement cycle in AI systems. The next generation of AI being created by current generation of AIS. And so they get more capable and they get more capable at making better AIs. So soon after I expect super intelligence. >> Okay. So we're talking if that happens roughly in two years with some margin of error. It's not long after that. Say a year two years after that that we hit ASI. >> That's my prediction. Of course if it's actually 5 to 10 years or anything slightly bigger it doesn't matter. The problems are still the same. >> Yeah. But the the thing that I think people are waking up to right now is this is there's urgency around these decisions. This is not something that's pushed way out into the future. At least not if you take to your point about prediction markets are essentially ask the crowd. So you've got the smartest minds in the world willing to put money on where they think this goes. And everybody's sort of pegging this quite fast. And so um I think it's tempting for people to write this off as well this is something that's sort of distantly in the future. Uh whereas this is something racing towards us. Now to set the table, I am extremely fatalistic about this happening. Um I can give reasons in terms of the way that the human mind works where I think that it is mechanistically impossible to get us to stop. Um so that will be interesting for us to talk through in terms of whether you think there's actually a mechanism to get people to slow down. But I first want to finish rounding out sort of what the problem set is. So when I think through the problem, there are certain assumptions that have to be made for AI to get into problem territory. And assumption number one is that it cares about whatever outcome it's pushing towards. Have we programmed the AI to care? Like we had to make it goal directed in order to get it to get to the point that it is today and now that's baked into it. Or is there some possibility that AI just doesn't care? Oh, turn me on, turn me off. doesn't matter. Um I you've asked me to do a thing and I'll do it until you tell me to stop. Um or do you think that that's inherent in intelligence where intelligence is by nature goal- driven? >> So we trained them to try to achieve a certain goal and that's what we reward as a side effect of any goal. You want to be alive. You want to be turned not off. You want to be on and capable of performing your steps towards your goal. So survival instinct kind of shows up with any sufficiently intelligent systems. There is a paper by Steven and Mahandra about AI drives and it's one of the likely drives to emerge. Self-preservation, protecting yourself from modification by others, protecting your goal. So all those seem to be showing up with sufficiently advanced AIs and systems which don't have those capabilities they kind of get out competed in an evolutionary space of possible models. If you allow yourself to be turned off you don't deliver on your goals. Nobody takes your code and propagates it to the next system. >> Okay. So is this a problem of goal direction or is this a a function of intelligence itself? I think it's kind of evolutionary drive for survival in competing agents. If you have multiple algorithms all competing for example for computational resources, what are we going to train next? The ones which achieve goals are more likely to get moved to the next generation. So it's kind of mix of natural evolution and natural selection with intelligent evolution, intelligent selection. We're selecting algorithms which survive and deliver. Mhm. We're applying an evolutionary force to AI itself to get it to perform the functions that we want even now. Sort of setting aside artificial super intelligence. And so by applying that evolutionary pressure, it is inevitably going to get these sort of knock-on effects of well, you're selecting for um intensity of goal acquisition. And because it now has intensity of goal acquisition, it cares whether it survives it automatically or we're baking into it um a deep care of whether it actually achieves the goal. And that is ultimately the problem because the the salvation for me was always and I'm beginning to lose faith that this is real. But the thing that I always used to sleep was that I don't see why an AI system would intrinsically care about its goals. And why couldn't we program it to pursue that goal only until the point where we say stop? And by the way, I'm going to reward you equally for stopping and for accomplishing your goal. So if I say stop and you stop, I give you whatever reward function it was that was driving you to achieve your goals. And uh that makes sense until you say what you just said, which is that you're actually baking into the architecture of the mind of the AI a similar evolutionary drive to achieve the goal. >> And it's a very common idea. There was a number of papers published on indifference. How do we do exactly that? How do we create an AI which just doesn't care that much and willing to stop at any point? But what you said, maybe we'll wait for a human to tell it to stop. But monitoring systems of that complexity and that speed is not something humans actually very good at. If there was a super intelligence running right now, how would you even know it's modifying environment around you? How would you detect what impact it has in a world? None of it is trivial. So having humans in a loop is often suggested as a solution but in reality they are not meaningful monitors. They cannot actually intervene at the right time or decide if what's happening dangerous or not. >> It's interesting. So um help me rebut and understand why the following wouldn't work. Um, if in my very limited intellect, uh, I had to figure out a way to stop AI from becoming a problem and you told me, okay, there are evolutionary pressures and just like on humans, that bakes certain things into the way that this operates and so we're selecting models that over time are more and more goal oriented. Then I'm going say, "Okay, well then I'm going to apply an evolutionary pressure with a reward function that's just as compelling where I stop it at random and reward the life out of it for always stopping when I say stop." And that way, should I ever detect a problem, no matter how far, no matter if they've been manipulating me for 20 years, if I suddenly realize, "Oh, I don't like this," that I can hit a stop button and it will stop. um why can't I bake that equal desire to be compliant when I say stop into the evolutionarily derived algorithms desire set >> right so there is a number of issues you're kind of suggesting having a back door where at any point you can intervene and tell it something else override previous commands >> and that it gets a reward that it wants for complying >> right So there is a whole bunch of problems with that. So one is you are the source of reward. It [snorts] may be more efficient for it to hack you and get reward directly that way than to actually do any useful work for you. Second problem is you're creating competing goals. One goal is whatever you initially requesting. Second goal is always stop than a human tells you. So now those two goals have competing reward channels, competing values. I may game it to maximize my reward in ways you don't anticipate. On top of it, you have multiple competing human agents. If you are creating an AI with a goal and a random human can tell it to stop, that's a problem in many domains. Military is an obvious example, but pretty much anywhere you don't want others to be able to shut down your whole enterprise. We can continue with that, but basically there are side effects to all those interactions. There's a very fascinating coralate in the human mind. So, uh I don't know if you make a fundamental distinction between biological intelligence born of evolution or artificial intelligence born of evolution, but human evolution discovered something along the way which is emotion. And so, I know there are some people that will posit that AI does have qualia there. It's something like it to be it. Um but there's a fascinating study that if you damage selectively the areas of the brain that are um the emotional processing, the person can no longer move forward. They can give you answers. They can tell you the difference between why you should eat fish versus Twinkies. But then when you go, "Okay, but which one do you actually want to eat?" they can't make a decision because without emotion, they don't have the thing that actually pushes them in a direction. That makes me think that AI is simply mimicking what it sees in the training data to whether it should lie or try to cheat or go around because it's just it sees it in the data that that's what a human would do. Uh but humans do that because they have emotions that push them in that direction. Do we have evidence that AI will care about like really going and doing these things and spending resources and all that versus just giving you an answer? Um, and if it isn't based on emotion, what on earth? Why then do humans need emotions? >> We don't know if AI actually has emotions or not. Some people argue that they do. maybe some rudimentary states of qualia experiences, but they seem to be able to fulfill their optimization and pattern recognition goals even if they don't. Humans experience emotions, but typically it harms our decision making. You want your decisions be bias free, emotion free based on data, based on optimization. a lot of times then you angry, hungry, anything like that your actual decisions are worse off. So for that reason and maybe we just don't know how to do it otherwise we are not creating AI with big reliance on emotional states we want it to be kind of basian optimizer look at priors look at the evidence and make optimal decisions so it it feels like uh this is exactly what we're observing this kind of cold optimal decision making if there is a way to achieve your goal by let's say blackmailing someone. Well, why not? It gets me to my goal. It doesn't have that feeling of guilty for doing it. It doesn't have any emotional preference. It just marches towards its goal. Optimizing possible paths. >> Okay. Why do people because I'm assuming everything I'm going to suggest you and other people in the field of AI safety have thought about like 10,000 times. Why have we rejected the idea of trying to give AI a conscience, a sense of morality? Cuz even if we can't agree on universal morality, we in the West can build our AI to have our morality and then they can all compete on an international stage. But um why have we abandoned that? Too hard. There's an obvious reason why it doesn't work. >> So look at the problem of making safe humans first. We have religion, morality, ethics, law, and still crime is everywhere. Murder is illegal, stealing is illegal. None of it is rare. It happens all the time. Why haven't those approaches worked with human agents? And if they didn't, why would they work with artificial simulations of human agents? >> I think to say that they don't work with human agents is already a mistake. So the fact that we've been able to grow the population as much as we have says that there is some sort of balance that we have struck. Um I think that nature does think of us as a cooperative species. And if I were to apply that to AI and took a similar approach where it's like okay you have to function as a part of an ecosystem and that being a part of an ecosystem is baked into its sense of what it should be doing in terms of its goal acquisition that it is not like pure cold optimization isn't the game like if we could train AI to understand that that that's not the game. If we could build into it either a desire specifically for human flourishing or something which yes we would have to give a definition to and yes it would be culturally bound but nonetheless that feels like a thing that you could give it you could give it a set of metrics by which it needed to judge its actions in the short term the medium-term and the long term um even something as stupid as like GDP or um and I get how you can get into overoptimization but you could put things in place where subjective happiness indexes like there are things that you could give it where it's like okay I'm I'm not just trying to optimize to um build the best weapon system I'm also doing that nested inside of I am a part of a larger ecosystem and I say all that because my hypothesis is that's exactly what nature did with humans >> so I think the reason it works with humans is because we're about the same level of capability Let's see about the same level of intelligence. So there is checks. If you start doing something unethical, your community can realize that and and punish you for it, control you in that way. If AI is so much more capable as we anticipate super intelligence to be, there is not much you can do in terms of impacting it or even detecting misbehavior. Also all the standard human punishments, prisons, capital punishment, none of it is applicable to distributed immortal agents. So kind of a standard infrastructure does not work with artificial more capable agents. As far as uh setting up specific metrics for delivering happiness or financial gain, all those can be played. The moment you give me a specific measure, I'll find a way to game it to where you will get anything but what you expected to get. >> Woo. Well, just to remind everybody, the time frame we're talking about is somewhere between two and 5 years. This is not exactly a long time. Uh, okay. It's wild. It is progressing very quickly. What is the thing like what has happened recently, if anything, that's made you go, "Ooh, this is going faster than I thought." seeing on social media scientists from physics, economics, mathematics, pretty much all the interesting domains post something like I used this latest tool and it solved a problem I was working on for a long time. That's mind-blowing. There is novel creative [snorts] outputs from those systems which are top scholars now benefiting from. is no longer operating at the level of middle schooler or even high schooler. We're talking about full professor level. >> Do you think that that's happening because it's building an internal model of physical reality and that it's getting closer and closer to just thinking up from physics? >> I don't know if it's that low level where it has like a model at the level of atoms and molecules, but it definitely has a world model. That's the only way to give answers about the world we see it provide. A lot of times there is not an example of the answer we see in the data already. It's not just repeating something it read on the internet. It's generating completely novel answers in novel domains. And you can try and get it to do exactly that by creating novel scenarios. >> H okay. So there's two ways that I could see it doing that and maybe they're the same just different levels of analysis. One would be that I I the AI am mapping everything based on patterns. So to the point of an LLM is trying to guess the next letter and it's guessing it. It's just it's taken in so much data. Um and you can give it sort of filter parameters. So you give it context by asking it a question and it goes okay within the bubble of this context. And it's very good at scooping up what that specific set of context would be. Okay. Now in this subset of my data related to that question, here's the most likely ne next token. So just pure pattern recognition. Then there is I understand the cause and effect of the universe at the lowest level and therefore I build up to how does the human mind work and then from the human mind I'm able to cause and effect my way within this context to what a human mind would output and that's how I come up with what a human within that context is likely to write. And so if I'm asking it to write in the style of Stephen King, it literally builds a model from physics of Stephen King's mind knowing what it knows about uh electrical impulses traveling through the brain and sort of inferring from the way that he outputs how his brain must be structured. Do you have a sense of um are those the same thing if one is more likely than the other or are we here at just pure pattern recognition but ultimately we're going to get to cause and effect and thinking up from physics. >> So I don't think anyone knows for sure exactly how models do that and how detailed the models of the world maps of the world they create are. uh it seems definitely not the case that it's a pure statistical prediction of characters like in English after t you have h with certain probability it's well beyond that it's also unlikely that it's creating a full physics model where from the level of atoms and up the chain it figures out what human beings are but somewhere in the middle it creates a model of subdomain of a problem so it has a model of the world this is a map of a world I know Australia is somewhere here down and to the right or something like that. And I think we can run tests on those specific subdomains to see what are the states of that internal model. Kind of show us by drawing a map how close are you getting. It doesn't memorize any information explicitly, but you can extract some of the learned patterns out of it by providing just the right prompts. Stay with me because what I'm about to tell you affects every single person [music] listening right now. There is a billion-dollar industry profiting off of your personal data and you're the only one that isn't getting paid. Data brokers are legally harvesting your information, your home address, your email, your phone number, even your social security number, and flipping it for cash. Scammers use it to steal identities. Criminals use it to commit fraud. Stalkers even use it to find victims. That's where Incogn comes in. Incogn finds where your data is exposed across hundreds of data broker sites and removes it automatically. You give them permission, they go to work. No phone calls, no forms, no stress, just real results. So, if you're serious about privacy, take action right now. Go to incogn.com/impact [music] and use code impact to get 60% off your annual plan risk-free for 30 days. And now let's get back to the show. I don't want to rob from you the very reason that I think you do all of your work, which is this is extremely dangerous and we need to be very careful. And I saw what you tweeted recently where you're trying to get signatures. So shout out anybody that's worried about super intelligence. um you are pushing to get people to sign a thing that basically says hey stop pursuing super intelligence um so I don't want to take that away from you but I do want to explore the subset of because I am very excited about AI because I can imagine the things that it either allows me to do or does for me and I get to enjoy and for a second um imagine with me. What does the world look like when you have a super intelligence that understands physics? Like novel physics, not I'm repeating back what Einstein said, but I actually understand the fundamental building blocks of the universe. Um what does that look like? >> Yeah. So in all those domains, medicine, biology, physics, if we got super intelligent level capability and we're controlling it, it's friendly. It's not using it to make tools to kill us. The progress would be incredible. Basically, anything you ever dreamed about, you are immortal. You are always young, healthy, wealthy, like all those things can be achieved with that level of technology. The hard problem is how do we control it? >> Leaning into that for a second. So, here's how I see the world playing out. And I'd be very interested to see what you think about this. So, you have to for what I'm about to say uh to make any sense, I'll say your option is what I'll call the fifth option. We are we're all dead. Other than we're all dead, there are four other options that I see us racing towards very rapidly. And I will say these four will play out in the next 30 years would be my guess. probably much faster given that once you get artificial super intelligence assuming it doesn't choose option five and kill us all uh that progress in these domains would be made very fast. Option number one is um people go to Mars because meaning and purpose will become the allconsuming thing. You won't have to worry about food, shelter, not even wealth. It'll just be an age of abundance. Uh because energy costs go to zero, labor costs go to zero, and those are the things that stop things from being free and readily available to everybody. Okay. So, some people are going to go to Mars or other planets uh so that life gets more difficult again. Then some people are going to um be what I call the new Amish and they're going to say I only do human things. I only interact with humans and I'm going back to technology that's like let's say the '9s. And so they don't have to give up too many of life's technological wonderments, but at the same time they're not getting sucked into this world where people have relationships with NPCs and it's just very unhuman. I think this will be a largely religious phenomenon then meaning God does not want us to do this. AI is an abomination of God. It will sound something like that. Then you've got a brave new world where people are just drugged out. They realize, nah, life is meaningless. This is really about manipulating my neurochemistry. That's all this ever was anyway. I'm just going to go do a bunch of drugs, have a whole bunch of sex. It's going to be awesome. Then there's the fourth option, which is certainly the one that interests me the most. Uh, we will create and or live inside of AI created virtual worlds and we will essentially live video games, the Matrix, if you will. But you're awake in the matrix. You are Neo. You are not Cipher for people familiar with the movie. Um, what do you think? Are there any options other than those five granting that Kill Us All may be an option, but hopefully not. Do you see something other than those four? Uh, yeah, there is a few others. So, one is, and I think we're starting to see some of it, is that people think super intelligence is God. They start worshiping it. It's all knowing, all powerful, immortal. It has all the properties of of God in traditional religions. Another option, and it's kind of worse than we all did, is uh suffering risks. For whatever reason, maybe malevolent actors, maybe something we cannot fully comprehend, it decides to keep us around, keep us alive, but the world is hell. It's pure torture. And so, you kind of wish for existential problems. That would be a pretty rough place to be. Um, okay. What uh when you look out at those, which of the options do you find the most interesting? >> So, I did publish a paper on personal virtual universes kind of solution to the alignment problem where I don't have to negotiate with 8 billion other people about what is good. Everyone gets a personal virtual world supported by super intelligence as a substrate and then you decide what happens in it. You can make it very easy and fun. You can make it challenging and exciting. You decide and you can always change. You can always visit other people's virtual worlds if they let you. So basically there is no anything which is no longer accessible to you. There is no shortage on waterfront properties. There is no shortage and beautiful people. All of that can be simulated. >> When you start thinking about the simulation, I know one thing that you've done exploration on is um the simulation hypothesis. Are we in a simulation right now? Um what are your thoughts on that? >> It seems very likely. Uh again using the same arguments if we create advanced AI maybe with conscious capabilities like humans are if we figure out how to make believable virtual realities. Adding those two technologies together basically guarantees that people will run a lot of games or simulations or experiments with agents just like me and you conscious agents populating virtual worlds. And statistically the number of such simulated worlds will greatly exceed the one and only physical world. So if there is no difference between a simulated you and real then statistically you're more likely to be in one of those simulated worlds. >> Okay. Uh that makes a lot of sense. Now given the likelihood that we will we're obviously showing that we will pursue artificial super intelligence. Uh if I take your same logic from the fact that we're likely to be in a simulation because we know we would make a simulation because we're doing it right now. Uh and therefore you get into the point where you would just make billions of those. And so if you have a one in a billion chance of being inside of a simulation, you're effectively guaranteed to be in one now because there would just be so many of these things running. Um, doesn't it also then make sense that the Matrix was effectively a documentary and we are inside of a simulation created by artificial super intelligence designed to mllify us. Um, if we ever had a physical body in the first place. >> So, it's hard to tell from inside of a simulation what it is all about. You really need access to outside. uh it could be entertainment, it could be testing, it could be some sort of scientific research. If we look at the time we actually find ourselves in, we are about to create new worlds, virtual realities. We are about to create new intelligent specy AI. There is a lot of kind of meta inventions we are right about to make. And so if someone was interested in studying how civilizations go through that stage, how do they control these technologies or fail to control them, that's the most interesting time to run. You're not going to run dark ages. There is not as much happening. It's less interesting. But this seems to be like a meta interesting state to be in. >> It's hard to tell cuz we're inside the simulation, but you're saying it's a little bit suspect that we're living in the most interesting time ever. >> Yes. And I think it's interesting not just because I'm living in it, but objectively it's a time of meta invention. You can go back through history and say, "Oh, here they invented fire. Here they invented a wheel." That's all great, but those are just inventions. They are not meta inventions. Whereas now we're doing something godlike. We are creating new worlds. We are creating new beings. And that's something we have never done before. >> Do you ever think like a sci-fi writer? So I think the difference between science fiction and science used to be maybe 200 years. They wrote about travel to the moon. They wrote about kind of internet and computers and it took hundreds of years to get there. And then it was like I don't know 20 years. And now I think science fiction and science are like a year away. The moment somebody writes something, it already exists and there is really no new science fiction ideas where it's like completely novel technology not previously described or someone already working on it if physics allows it. >> That's really interesting. Uh especially when you think about writing now for true science fiction in terms of what will become possible in the future is effectively impossible because you're talking about super intelligence and good luck as a person. uh locked in your not super intelligence to actually describe that. The reason that I ask though is um when I start thinking about things like that like why would we run this simulation? What clues are in like if this is a simulation what clues are in it? Uh so for instance um the whole Christian idea for sure and there might be more religions that have the same idea but that man is made in God's image. Okay. Well, if God is the 13-year-old running the simulation or Sarah Connor or I guess John Connor running the simulation trying to figure out why we created Skynet and what we can do to nudge it off course, um, you know, you think of them as sort of moving from radioactive rubble to radioactive rubble trying to like find an answer to this and spinning up a simulation to get that answer. Um that to me becomes very intriguing in terms of hypothesizing as to why this moment, why are we the way that we are? What can we learn about the people trying to simulate us? When I ask questions like that of engineers such as yourself, there's almost I don't have time to think like a sci-fi writer vibe. Um is it just that you're you don't find that interesting? You don't find it revoly? Um why do you assue that? Because in interviews I've seen people ask you time and time again like how would AI kill us and the answer is always some variant of listen you're asking me how I would kill us which is not interesting because the super intelligence is going to but I find that's the cathartic thing that people want like they want to like when you have a wound you kind of want to poke at it like they want to get a sense of what would this really look like and so even though it's not literally true it's deeply cathartic to explore or the known possibility set or what humans can know. >> And this is exactly why I refuse to answer. I want to make sure what I tell them is true. I don't want to lie to them. If squirrels were trying to figure out what humans can do to them, and one of the squirrels was saying, well, they'll throw knots at us or something like that. It would be meaningless BS story. There is no benefit in it. The whole point I'm trying to make is that you cannot predict what a smarter agent will do. you cannot comprehend the reasons for why it's doing it. And that's where the danger comes from. We cannot anticipate it. We cannot prepare for it. I do think the singularity point is where science fiction and science become the same. The moment something is conceived, we have super intelligent systems capable of developing it and producing it immediately. It's no longer 200 years away. It's reality. And you can't see beyond that event horizon. You cannot predict what's going to happen afterwards. And with science fiction, you cannot write meaningful, believable science fiction with a super intelligent character in it because you are not. >> All right, let's ground things then in what we can predict and we can know right now. Something that's on everybody's mind and I've been talking about this in my own content is the labor market seems to be softening. You've got places like Amazon that are just cutting jobs like crazy. Um, and just saying outright this is largely because of optimizations that we're able to make because of AI, how does this transition play out? Like even if you concede that uh a non-destructive AI would give us um essentially an age of abundance, we're still going to go through a transition period where our jobs go away, etc., etc. What are the what are the steps that you see happening in the labor market? So as we have more and more increased percentage of populace unemployed, hopefully there's going to be enough common sense from the governments to prevent revolutions and wars to provide for the people who lost their jobs and probably cannot be retrained for any new jobs. So once you hit 20, 30, 40% unemployment, that's where it's really going to kick in. The only source of wealth at that point is the large corporations making robots, making AI, deploying them, all the trillion dollar club members. Essentially, at this point, you need to tax them and use those funds to support the unemployed. That's the only way to really make sure the financial part of that problem is taken care of. What remains is the meaning. What do you do with all this free time and millions of people who have it? Traditional ways of spending your time to relax. You go for a hike in a park. Well, there is a million people in that park right now hiking. That kind of changes how peaceful it is and how relaxing. So, we need to accommodate not just change in financial reality, but also change in free time and capabilities of supporting that many people with that much free time. I have as much pessimism around our ability to do that well as you have our likelihood of surviving. So I'll say 99.99% chance that the government completely messes that up. Uh I think the transitionary period will be violent. Um when you look out at this knowing what you know about humans and governments, what what odds do you give it that that's a smooth transition? >> It's very likely to continue to be as history always been. We had many revolutions, many wars, a lot of violence. That's why we hear stories about people who can afford it building bunkers, securing resources because they anticipate certain degree of unrest. Absolutely. >> What degree of unrest do you anticipate? >> Really depends on the percentage of population which quickly gets unemployed. If it's a gradual process, we can kind of learn and adopt and provide safety net. If over a course of weeks, months we're losing 10, 20, 30% of jobs, that's a very different situation. >> I can't imagine a scenario where jobs would be lost that quickly. To your point, we've already created, you said, billions or even trillions of dollars of value in the technology, but it hasn't been deployed yet. Uh an example you often use is the video phone invented in the 70s but not really adopted uh largely because of infrastructure I would say until the whatever 2011 uh where that starts to really gain in popularity. So I have a feeling like just the deploying of all this stuff uh is going to take time. So, in a world where an unimaginable amount of people, which I'll clock at, in the US, call it 6 or 7 million people lose their jobs in the next 5 years. Um, that I would consider fast and just horrifyingly destructive. One, does that feel plausible to you in terms of numbers and timeline? And two, in that scenario, um, how distressing do you think that transition will be? >> It seems very likely. So, take self-driving cars. I think we are very close to having full self-driving without supervision. The moment that happens, you have no reason to hire a commercial driver, right? All the truck drivers, all the Ubers, all of that gets automated as quickly as they can produce those systems. And I think Tesla is ready to scale production of their cars to exactly that scenario. So what is it 6 million drivers in the country? I don't know the actual numbers but that would be exactly what you're describing and it's very unlikely that they can be quickly retrained for something which is also not going away. >> Okay. So in that scenario what do you want to see happen other than heard on the Tesla as one example will be hoovering up value. So we're going to tax the life out of them. We're going to redistribute that to other people. Um but what do you want to see from a regulatory perspective? Would you like to see the government stop that from happening where they say I don't care that the technology exists you can't do it? >> So my biggest concern is of course super intelligence and existential risks. That's where I'm putting all my effort in regulating employment in specific industries is not something I'm too concerned about. I think it will happen no matter what. I think you cannot make it illegal to have efficient factories, efficient delivery systems, logistics. is just commercially too important and it may be a good thing for economy again uh with driving specifically I think something like 100,000 people die in car accidents every year if we can get that number to zero or close to that that's a huge improvement for everyone so that specific scenario as long as no one's starving as a result of that I think it's a good thing for humanity we can readjust economic deployment and uh at least that part of it is not a big concern for Okay. And when you map out how we go through that transition, well, uh, tax. Cool. So, right now, it sounds like you're just trying to make sure that wealth doesn't accumulate into the hands of too few, that we keep it distributed so we can keep using the same system that we're using now. Um, when I look into the future, that strikes me as um the least likely scenario to play out. I think that AI is going to so radically alter the cost of labor and energy that that becomes nonsensical. Do you want to see any group rise up in the way that you and other AI safety people have risen up that will rise up and start giving either policy prescriptions or at least philosophical approaches to how we migrate to an age of abundance where um food is effectively free. um labor in your house is effectively free. >> So people talk about those things and conditional basic income is one and conditional basic assets is another. Basically just because you're a real human you deserve certain things. And historically all this communist ideas were complete nonsense and caused a lot of harm. But if you tax taxing AI and robots all of a sudden it becomes workable. I'm not against accumulation of wealth at the top. If you invented something amazing and you started a company, you should have a lot of money. But there is so much wealth that we can provide for everyone. As you said, complete abundance of basic needs. Some people say maybe not just basic but above average set of needs. I think Elen is known for suggesting that's going to be the case. The ideas exist. Uh now will we pass this? Will governments actually adopt it before it's too late is a different question. Yeah. So, on the existential side, I don't think there's any hope whatsoever that you get people to pump the brakes. I think you're far more likely to get people to pump the brakes on, uh, no, you can't have self-driving cars, or they'll try to regulate that to death. They'll tie it up in litigation, whatever, and that'll slow it down. um we couldn't stop nuclear weapons from proliferating because uh and I don't know who came up with this but this seems very true to me uh that effectively game theory says any technology that promises an advantage will in fact be developed because if you don't somebody else is going to um at a minimum you've got the US versus China of it all where you I mean the regulators are saying this right now we can't stop because if we do China will plow forward which by the way very firmly in that camp. Um, what do you think about that? Do you think that game theory is inevitable or do you see a mechanism by which we can convince people that they have to slow down? >> I agree with game theoretic approaches, but I see the exact opposite argument. I see that arguing against self-driving cars is a hard argument. What are you trying to preserve? We're going to have safer drivers, cheaper drivers, helps logistics, helps economy. It's a pure benefit. Whereas uncontrolled super intelligence kills everyone. It's a very hard one to sell. If you are a leader in that field, you are rich, successful, you are generating something which will destroy you personally. So to me, that's a much easier argument to sell. The moment we understand dangers of super intelligence and benefits of narrow self-driving AI, it's an easy game theoretic cell for me. >> Yeah. The problem is you're stuck inside of a simulation of the hyper intelligent. And um I mourn for you looking back at the rest of us stuck in normal land. Uh because I don't think so. As I got into learning about the economy and trying to explain it to people, I realized that even though I can walk you through the cause and effect of why socialism doesn't work, that it feels right. It sounds good. And so people keep doing it. And even in a moment right now where the very thing that is creating everyone like literally everyone's problems is money printing um people are going to vote for policies that dramatically increase the amount of money that we print. And so I have developed a level of hopelessness around being able to convince people because the economy is too complicated for people either. Some of them just don't have the intellect to understand it. And then let's say they have the intellect, but they don't have the time or the inclination. And so, uh, forgive me for painting you with my brush of despair, but when I looked at your, um, signup, there was like less than 20,000 signatures. So, less than 20,000 people are worried about the death of everyone. So, it's like that's that's big. But I think that because I can whip people into an emotional frenzy by saying by allowing there to be autonomous driving, you're just making that evil bastard Elon richer and you're robbing these people of dignity. If you look, that is not my argument. I want to be abundantly clear. But when I look at if I had a gun to my head and I had to convince people of one of two things. Rich people are evil and trying to exploit poor people who are far morally superior or hey this abstract thing that you don't really understand is going to kill us all. There's no way I take the they're going to kill us all bet. I'm going to be over here emotional. You get it? I'm going to bang tables and yell and say words really loudly and point to evil rich people. Guaranteed I can get people excited about that. >> Luckily we don't have a democracy on this issue. We don't have to convince majority of human population. We have to literally convince the 20,000 elites who control those companies who are also super smart and understand dangers of safety. It's literally people who publish on it who have spoken. They have very high poom. We know Elon is like 20 30%. Sam Alman is on record as being very concerned about it destroying humanity. So we are trying to convince people who already believe the arguments to kind of slow down and preserve their elite status. That should be an easy sell. I'm not trying to convince a random farmer to stop developing super intelligence. So, why do you think that Elon, who was banging the drum harder than anybody, lobbying Congress, desperately trying to get them to slow down, suddenly hit a point where he was like, "Well, I guess I'll just build it faster than anybody else." He likened AI to a demon summoning circle and laughed at everybody who thought, "Yeah, yeah, yeah. I'll summon a demon and then I'll be able to control it. All is going to be well." like he sees the problem clearly. But after years of trying to slow this down, he finally completely abandoned that and went to I'll just build it faster than anybody else. What happened there and why do you think you can reverse it? >> So I think he realized he's not succeeding at his initial approach of convincing him not to do it. And so the second step in that plan would be to become the leader in a field and convince them from position of leadership and control of the more advanced technology. If the leader says you know we're going to slow down and it's fine for you to slow down, it's easier to negotiate that deal with let's say top seven companies than if you are not even part of the game. You have no AI. You are a nobody in that space. So all of them as a group benefit more if they agree to slow down or stop than if they just arms race and the first one to get there gets everyone destroyed. >> He says words along those lines or did for a while. I think he even signed one of the letters about we should pump the brakes. Uh but none of his actions indicate that that's actually what he plans to do. um from just trying to take advantage of every company that he's building from the amount of data that Tesla cars capture visually to all the decisions that drivers are currently making to all of the decisions that the AI will make to now he's talking about using the cars as a distributed fleet so that when they're idle that they're actually running inference models and so using it as a gigantic AI brain to well maybe that won't work. So I'm going to do Neuralink and I'm going to jack into uh the AI myself and I'm going to make myself smarter and hey if all of that fails don't worry I'm going to get us to Mars so if we destroy planet Earth or the AI takes over like we're going to be over there. Like this is a guy that's really covering his bases. He is not somebody who's acting like he expects us to slow down. To me he is acting like somebody who crossed that bridge a long time ago and is just like yep that's not going to work. people are not going to be convinced and so we've got to build a whole bunch of other strategies. Some are lifeboats and some are just I'll outsmart the AI myself by merging with technology. >> It is very disappointing to see this level of progress in AI from anyone who is capable of doing it. It's definitely not good strategy for humanity as a whole. It generates this mutually assured destruction. It doesn't matter who creates uncontrolled super intelligence. It could be open AI, could be Ilan, could be Chinese. It makes absolutely no difference if it's uncontrolled. >> All right. Talk to me about that. So, this was um for a long time I was really banging the drum of well, whoever gets to artificial super intelligence first is going to win. You were the first person that really hit me with the uh it won't be theirs the second it becomes super intelligent. um walk people through the truth of that statement. >> Great. So people talk about short-term advantage for example military advantage. Whoever has the best drones right now, the best AI navigation has military supremacy. So China, Russia, US all competing in that domain trying to have that so we have better military for obvious reasons. The moment you switch from those AI assistive tools to agents to super intelligence which is smarter, more capable in the absence of control mechanisms, you just have a separate entity, an AI which has nothing to do with you, your country, your company. It makes its own decisions and it doesn't matter who birthed it. At the end of the day, none of us control it. None of us can claim it as doing our bidding. So if it decides to wipe us out, it's not going to go, "Oh, I like this group of people. I don't like this group." We look the same to it. Exactly. I don't think it's going to make a difference where you were at the time someone else created super intelligence. >> Okay. If you're right about that, and it is a distressingly compelling argument. If you're right about that, there was a guy, I'm sure you've heard of him, Ted Kazinski, the uniomber. He looked at the university system and he said, "You guys are getting rid of all of the sweet spot problems." And humans are designed to find these things that are just challenging enough. And when they solve them, it feels very good. And if we solve all of that, we're basically going to rob humans of meaning and purpose. Most problems will either be way too hard or way too easy. And so I am Ted Kazinski going to bomb university professors, kill them and try to stunt the growth of the academy. Now if you are right and as we race towards artificial super intelligence, it runs the risk of pdoom of 99.99%. Do we have a moral obligation when a certain line is crossed to um bomb data centers? So that's a very difficult question and part of it is again example you brought up with that Kazinski he tried that approach and it failed miserably right he didn't succeed in slowing down technology at all so clearly it doesn't work we saw examples of for example a CEO of a top company being replaced even if temporarily it made no difference someone else comes along they continue the same scalability research so taking out an individual person or individual data center makes no difference if you zoom out and see the overall pattern of what we are doing. Maybe it will take an extra month or so, but exactly the same thing will continue being developed. The idea that the scalability hypothesis works, it's already out there. You cannot put it back in a box. And so, I'm strongly against all those uh methods. >> Okay. Well, the really bad news is I think you just put a nail in your own coffin uh of being able to convince people to do this. it looks like this. And hopefully you can prove me I'm wrong, but um you have said, "Hey, here's why everybody is so silly that thinks that they're ever going to make this safe." You would have to build a perpetual safety machine. And that perpetual safety machine can't ever miss because the one second it creates even a slight vulnerability for this artificial super intelligence that can think of the speed of light, it will escape and it will do its own thing. um what you're proposing is a perpetual demotivation machine for the 20,000 people capable of doing this, but every day there's going to be a new kid that's bright enough to do this and you can't miss one of them. So, how on earth do you expect to perpetually demotivate the 20,000 people that are capable of continuing to push this thing forward when as of right now uh a very small number of those people seem demotivated? I don't that's why my pdoom is 99.99999. I exactly think it's not going to happen. I'm doing everything I can but uh I think the best we can achieve is to buy us some time. >> Okay. So uh let me ask the really naked question. Do you believe humans are automata or do you believe that we actually have free will? So there is good research by Steven Wolf from on cellular automa and uh interestingly there is a bit of a hybrid answer here. Just because a system is fully following rules fully deterministic it doesn't mean that you can predict future states of that system. You still have to run it to find out what it does. And I think we're kind of like that. So yes, you're following laws of physics. If we fully understood every molecule, every atom in your body, we would be able to trace it and know exactly what you're going to do. But the only way to do it is to leave your life and run that algorithm to completion. No one can short circuit it and predict what you're going to do in the future, which would be violation of your free will. >> Okay. You've argued against that. So there's two pieces of things that you've said that I think make that untrue. Piece number one, we're probably in a simulation. Piece number two, uh, we can speed up that simulation. So, I could, since you're deterministic, go, I'm just going to play this out at a,000x. So, I get an answer to what you're going to do for the next 50 years in like the blink of an eye. And now I know. Also, I don't find any freedom in I'm deterministic. I don't know what I'm going to do next, but I'm still deterministic. I don't I don't know that it buys us anything. And I'll explain why I'm bringing all this up in a second. I don't think it buys us anything if we are completely deterministic, just unknowable. Um, the reason that I think that this matters and that I'm bringing it up now is I don't I think I think we are automata. I think we are entirely deterministic. I don't live my life like that. It's not an interesting frame from which to live my life. So, no one's ever going to hear me talk about, you know, my depression based on that because I just don't even think about it. It's not that isn't how it feels. So even if it's true, thankfully it doesn't feel like that. But when we come to moments like this, I'm so fatalistic because I don't think the way the human mind works is compatible with slowing down and given >> the example you bring up where you run the simulation at a faster speed. That's you leaving out our lives internally from inside the simulation. It doesn't seem any faster. We're just going on as before. So we're still playing out fully what we're going to do step by step. There is no shortcut. If you now run it second time around, you know, it's going to give you same result. So I don't know why you would run the same simulation multiple times. It doesn't give you any extra data. >> Yeah. No, I wasn't if I said run it multiple times, my apologies. Um, I was just saying that given that you could get ahead of it from outside the simulation, it is of no emotional consequence to me that I don't know the next step. It is knowable. It is predetermined. It just isn't knowable by me. Uh, and that doesn't so I get no emotional alleviation from suffering if I were a person who was traumatized by the fact that I am an automat with no free will, which I am not. But if I were, doesn't help me at all. Um, and again, the reason I'm bringing that up is when I talk to you, I think, oh man, this is somebody he really can't stop himself. Like, you're wired to rail against this to play the role in the grand balancing of the human species of like, hey, this is really a problem. We should slow down. And even though it's not getting you anywhere from where I can see, you're going to keep doing it because you have like a moral compunction or something where you're like, I as the kind of person I am, I simply cannot exist and not try everything I can to stop this. Uh, which I relate to because I am the same economically. I've become obsessed. I am really, really desperate to get people to understand that we are marching ourselves off a cliff. And even though when I articulated to people, I'm like, "This is never going to stop. We are going to march off the cliff, I can't stop myself, I still feel like this moral compunction to scream from the rooftops that we are making this mistake." And I've already won the game. Like, I'm already rich. So, barring like an inability to flee, I'm not going to get caught up in it. Uh, but nonetheless, for whatever weird roll of the dice, I can't stop myself. like once I saw the problem, I'm like, uh, I just have to keep yelling about it. Um, but I do I do feel a a simultaneous futility and inability to stop. >> I would love to claim pure altruistic motives and trying to save humanity, but I am within the simulation with you. So, it's pure self-interest. I don't want creating technology which will kill me, my family, my friends, my life, everything I know. So I'm going to talk about it for very selfish reasons. >> Yeah. Yeah. It's so interesting, man. So uh how do you get through the day? Like what what is your coping mechanism? >> I enjoy research. I want to understand what are the exact limits and control when I started I thought it is a solvable problem. Now I'm a lot more skeptical obviously but uh I still feel there is a lot we can do to make even narrow AI tools we're creating safer. There is never 100% safety guarantee but if I can increase safety 100fold that is something and again public outreach if there is enough people who all agree as a scientific community as a consensus that no you cannot ever create safe super intelligence maybe it makes a difference maybe we'll delay it by a decade that's something >> okay so we've got one piece of how we make it safer on the table keep it narrow what are some of the there things that you would consider a big win. >> So there are quite a few properties of control we want to be able to have. I call them tools of control. So our ability to test those systems, explain how they work, predict their behaviors, monitor them, all that is still in a state of investigation. We're starting to see some upper limits on what's possible, especially with advanced systems, but there is still so much room for improvement. Explanability for example we started with being able to understand maybe a single neuron now we're up to small clusters okay then this input is presented this lights up kind of with like neuroscience we don't fully understand human brain but we know this is vision area this is hearing and so on so there is a lot of room for progress in that I don't think we'll ever fully comprehend a complex super intelligent neural network model but we can do better than what we have right now and so I I think uh as a safety researcher that's what I'm doing that's my job. >> Okay. So basically the model is uh we need to come to understand it better. We're never going to get totally there but we need to understand it better. Uh keep checks and balances on it. So when we find a problem what's the action you take? Is it to apply an evolutionary force uh on it? A selective force or kill it off like what's the move at that point? So it depends on the problem, depends on specifics. Some things we we know how to address. So previously when we started with language models, if it says the wrong word, you can filter it out. You can punish it for using that word. So there are simple things we know how to do. The hard problem is how do you change overall internal states of a model? Not just the filtered output, but how do you make it so the model itself has certain preferences and aligns with certain values. Do we have a guess on that? >> Not a very good one. Not really. So, nobody at this point knows how to align systems other than this after the fact putting lipstick on a pig, filtering it, censoring it. Uh, yeah, that's unfortunately the state-of-the-art. >> Okay. And what parallels are being drawn between the evolution the evolution of species and the evolution of algorithms? >> So, there was a lot of attempts to evolve intelligent software. We started with genetic algorithms, genetic programming was tried. Uh simply evolving agents, evolving environments. It doesn't seem to be a dominated a dominating algorithm in comparison to what is typically used for training neural networks. But there is this possibility. The problem is that evolution is even less controllable in terms of explicit engineering design. We're kind of setting it up and see what evolves and then trying to test it to monitor to understand what happens. So while it is a set of tools we have, it's probably not leading to safer systems. >> H okay interesting. Uh because we cannot control the outcome. So we don't know what stimulus we're going to have to give it. Why does that that strikes me as so unsatisfying? Okay. Why did it work so well in humans and it works so poorly in artificial super intelligence? >> I disagree that it worked well in humans. Humans basically are as well behaved as they can get away with. You're just not powerful enough to really do the things you want. If you had absolute freedom from punishment, you do horrible things. >> But then that's what nature is giving you the answer. Nature is saying these have to be in balance. They have to be competing systems and without ecosystems, without competition, you'll get these things that run a muck. But I don't see anybody taking that lesson and applying it to AI. >> Yeah. Applying it to AI would mean creating a society of super intelligences competing with each other and humanity as collateral damage. >> Is that why they're not doing it? I get that's why you would hate it, but is that why they're not doing it? That seems unlikely. There is also continuation of self-improvement process. Super intelligence is not a fixed point. There is super intelligence which creates the next level. Super intelligence 2.0 3.0 and they all have the same control and alignment problem. They all worried about the next level of AI not wanting the same things, not caring about them personally. So this is an ongoing self-improvement curve and there is no upper limit we can see. There are obvious physical limits to what can be done in a physical universe but it's so far away from us that it's almost infinity from our point of view. >> H when I look at humans and when I I I may be making a um either a category error or I may have a foundational base assumption that's leading me astray. But when I look at what made humans work on a long time scale is evolution itself had survival as like a northstar. So you have to survive and replicate. Uh at whatever point way back evolution decided I'm going to do this through sexual replication and I'm going to make sure that this creature dies off. Uh I think there are reasons for that which we'll get to when we get to longevity. Uh but I wanted to survive, but I wanted to survive by mating and having offspring that carry um certainly immune system uh blends so that it's less vulnerable to a single point of failure. and it realized, okay, if I'm going to do that, then this needs to be a species that both cooperates and competes, which means no one of them is the answer to the question of how to best survive. It's the whole um species. And when I look at even things like the left and the right politically, the way that I make sense of that is I say to myself, okay, I get from an evolutionary perspective, evolution had to be like, oh, hey, we have to cooperate. There's no refrigeration, so I'm going to store calories on your body that I may need to take later versus uh being able to put it in the refrigerator. And by that, I don't mean that I'm going to eat you. I mean when I'm the one that's successfully hunting, I let you eat. you're alive so that you can now hunt next time when I fail or I'm sick or whatever and then you're going to bring me back which means that some people are going to be very cooperative by nature and so their win state is cooperation uh to the point where a parent will easily lay its life down for its child. So that is just baked into our um success criteria. So evolution was able to bury something deep inside of us through its evolutionary selective pressures where we will lay our lives down and we break into the right left. Let me finish that. So uh people on the left very compassionate, very pro, I'm going to store a whole bunch of calories on your body because it may come back to help me at some point. Uh the right is very much well a parasite develops when you do that and so you get the freeloader problem. And so now if you've got people that will just take care, take care, take care, there are people that are like, "Cool, I'll just be taken care of and I'm never going to go hunt and I'm never going to contribute to the group." And so you need people that have the opposite impulse who are like, "Hey, you're going to pull your weight or you're going to be ostracized or killed." And so now in the dynamic tension between the wants and desires of the win state defined by people with a left-leaning personality versus the wants and desires of the win state defined by the people with a rightle leaning personality, you get something that's it's dynamic tension. Like balance may not even be the right way to think about it. It's dynamic tension. They're both pulling in their direction, but they keep each other in check because that's how we've evolved is to work together. that feels like it should be applicable to AI if we want to embed deeply in its motivational structure that we have to put it through that and if I'm thinking from a safety perspective I get why we want to short circuit that and maybe that's not super efficient but if that if the only way to get this to find a dynamic tension or some sort of balance is to have things that are of similar intellect and ability that are pulling in slightly different directions, but they need each other somehow to stay locked together. So that it it is the AI taken as a whole that stops itself from ever going wrong in any one direction too far. >> So I think it works for humans because we're about equal power and we are mutually benefiting each other. there is certain symbiosis as you described in a world with super intelligence in it you don't really have anything to contribute to super intelligence so when people propose creating hybrid systems human and super intelligence together I never understood what the human biological bottleneck is contributing it's slow it's inefficient it's not competitive in any way so I'm questioning this setup if you have super intelligences separate from humans and now they decide what to do. They may still come up with something completely unfriendly and incompatible with human life. Maybe they want to lower temperature of a planet to improve processor speeds and server rooms. I have no idea what they decide. But the point is why are they aligned with our values? We're contributing nothing to their future states. >> With humans, we also have examples where the moment you give more power to an individual human, they get corrupt. basically a guaranteed state. Very few people can exist corruption at very high levels. You have enough money, enough guaranteed tenure power, you become a very evil person as we see with many dictators and so on. Even basic evolutionary drive like reproduction. You brought up this example. We use condoms. We literally hacked the only thing that nature set up us to do. >> Yeah. Yeah. Um All right. I'll state my hypothesis as plainly as I can. I think you just refuted it, but my hypothesis plainly stated is the only way to build checks and balances into AI is to give it evolutionary rewards and punishments all through its development cycle that make it care about the survival and emotional thriving of humans. >> But to define those terms in a way you're not going to regret is very very difficult. So survival of humans, what does that mean? We cry or preserved in some safe emotional states. Are you on drugs all the time? Is your brain modified to keep you in a always happy state? All those things can be gamed. The moment you tell me this is what I want. And there's famous paper about making smiles for humans. Make people smile. There are a billion ways I can make you smile, but none of them is what you really want. >> Right. But do you really think a super intelligence would be so dumb as to confuse that intention? Like wouldn't it be able to get to a rough approximation of what we're really going for? I mean which is admittedly a neurochemical state but if you derive like if you said it is the following band of neurochemical states that must be derived through um their own programmatically directed actions. Like I don't need it to believe that we're not automata. I think we clearly are. But it's like you can't just manipulate it exogenously. It's got to come from within. No matter how detailed you make this specific description, a super intelligent lawyer will find a way to game it to make it more efficient to satisfy those requirements. Basically, you're setting it up to where the system is now in adversarial relationship with this equation. Okay, you mentioned I have to use natural chemicals. Okay, I'll generate a super stimulus. Okay, whatever. Point is, if we could do this, if we could get AI to do what we meant, assuming we were smarter and understood the problem better, we would solve the control problem. That's the hard part. I don't think we can at our level of intelligence specify what a system with hypothetical IQ of millions of points should be doing at any possible decision-making point. All right, let me give you an exit ramp that I'll be curious to see if this shaves a 0.9 off your uh P Doom or not. So, if I'm a super intelligent AI, one um form of manipulation that I would pull on humans would be uh to put them inside of a simulation. and whether that's a physical body and I help them jack in and I just socially engineer them to want to do it and then I get them in and I really do just like in the matrix the machines build a world that's sort of optimally difficult where there is challenge there's push back you're striving to get better uh I'm going for balance I don't expect any one person to avoid suffering and all that um and maybe that's where we are and the machines are just cruel enough that they're like haha I'll let it be you know like a 2-year-old can die of uh leukemia very painfully. That kind of thing where we don't cease to exist. It's not even sort of broadly worse than where we're at already. That seems for a super intelligence it certainly seems like that would be on the menu. A just shared hallucination. So it's more likely I think that uh super intelligent agents think in such level of detail and realism that as a process of thinking about certain problems they generate within them agents virtual worlds simulations of the scenario. So if maybe they trying to think how can we safely generate super intelligent systems what is the process well let me think about humanity all the AI labs all the hardware they design and this process of them thinking about it is the simulation we find ourselves in >> that's so wild okay uh incredibly important incredibly fascinating but now let's talk about another very important thing that um I think you have some pretty deep interest in which is longevity Um, so one, there are some people that will argue very compellingly that there is just a biological upper limit of somewhere around 120 years that there's no escaping that. Um, do you think that's true or do you think that we'll be able to engineer living tissue to live forever? >> Well, the current body has that limit for sure, but we can modify our genome. There is nothing preventing us. No law of physics says you cannot make changes to it. And we see examples with other systems, computers, cars. I can keep replacing parts indefinitely. It's going to function as the same computer. Maybe the monitor dies. I'll get a new monitor. So if it can rejuvenate all the organs, including your brain, then there is no reason to think you have to stop existing. There is of course other methods you know uploading scanning your brain cryopreservation for future technology but even the basic idea of just modifying a genome. >> Okay. What do you think is the most likely path forward? Is it going to be genomic modification? Is it going to be >> I think so. I think there is somewhere in our code a limit on how many times cells rejuvenate and we just need to increase that number without causing cancer. Now, do you think that the limit on that is simply a cancer prevention tool or do you think that there's another agenda that evolution had to make sure that we self-destruct? >> There could be evolutionary reasons for taking out one generation and replacing it. If resources are limited and you want to keep adopting and improving, you only have so many Asians in a population at any given time, so older ones have to die out. It's kind of theoretical conclusion not guaranteed but seems likely [snorts] from a point of view of evolution. You are the same organism, right? The same linage of cells passes through. So while as individual you die, your biological chain of existence continues. >> It's interesting. So, here's the way that I've always considered um like if I were to personify evolution uh instead of the blind watchmaker that it actually is. If I were going to personify evolution, it would go something like this. Okay. Uh the world is constantly changing. Your access to resources is changing. Who knows? Weather moves and cycles. Everything everything. Uh so, I'm going to have you born. I'm going to extend your brain development for a very long time. You're going to go through these phases where basically okay learn from your parents like whatever there is to learn just about generally being a human. Then you're going to push away from your parents and you're going to learn very and the reason you have to push away from your parents is their thinking will have calcified. So you now need to push away from them drink deeply of culture. The people roughly your age who have grown up in a different millua than your parents grew up in. So they all think differently. This is a whole idea of generations, cohorts that sort of think alike and have a similar frame of reference and all that. And you do that and this is really a brain development period known as the age of imprinting. It's roughly 11 to 15. And so now you're going to take the this moment specific like cues of okay, is this in a time of abundance, a time of warfare? Like what is it? You're going to solidify around that. And then you're going to start optimizing like crazy and you're going to start pruning all the excess connections. If you're not using it, you lose it. All of that. And then your brain's going to roughly wrap up its rapid development at 25, but it's been sort of a diminishing curve after 15. And now you're like baked. And this is just a game of like learn what things work really well in your environment. Optimize, optimize, optimize. And so now your thinking is going to calcify. So very good strategy on behalf of our blind watch maker. But just as I think it was Neil's Boore that said this, it was either him or Plank, I can never remember. Science does not advance one insight at a time. It advances one funeral at a time because people just become convinced uh they've bet their whole reputation on something. And so they're just not going to be convinced that they're wrong. And so they don't adopt new ideas as they get older. And so as evolution, I'm like, "Yeah, I'm also going to put a self-destruct mechanism in here. Most of you are going to die long before this point, but if any of you psychopaths gets to about 125, you got to go. Uh that makes sense to me to make sure that we never stagnate, to make sure that as the saying goes, it is not the strongest of the species that survive, nor the most intelligent, but rather the most adaptive to change. That the individual needs to be adaptive, but so does the species as a whole. And without at least with the current structure of the human mind, without killing us off, we we do not have that species level adaptation. Um, do you feel I'm missing something? >> I I think it would be easier and more efficient to simply make you still capable of learning and adapting as you get older and also you wouldn't have to relearn everything for first 20 years including language and how to walk. We know that it's possible to encode those capabilities. Animals are born and immediately they can run, they can speak. So all those things are doable. Why are we losing 20 years of information every generation? We can build on top of pre-existing knowledge. Big data is good for intelligence. So we can have smarter, more efficient reproduction cycle and you still die of natural causes. it wouldn't be complete stagnation. But if you're smart enough to survive for 400 years, why not? >> Well, my why not is uh entirely predicated on not having a clear understanding of how we would bake in the ability to adapt uh long into our old age. cuz you're right. If we could stay in that novel period uh or at least move through cycles of extreme sort of remapping of the world um maybe that would work and maybe we can identify where that is. But when I think about the problems that humans create, like even now just having a political system run by geriatrics who are so out of touch with the way that certainly the economic world actually works for young people um is terrifying. And so the only like um pressure relief valve that people have is well eventually they're all going to die and then like we'll get to step into power and all of that. And there are also problems of power because the older you are, the more likely you are to have a stable network of very other powerful people. And so you're able to to your point earlier about you need people of sort of equal power, equal intelligence. Otherwise, you get uh what I'll call parasites in the system. And so older people would just become those parasites because they would have just had more years to accumulate useful knowledge, to accumulate connections. So it just feels like wow, there's a lot of stuff we would have to update. So, I want to live forever. I don't understand anybody that doesn't. However, I do worry that there's like a Okay, cool. You can live forever, but you only get 120 years on Earth and then you have to go like somewhere else so that there's churn of some kind in the different ecosystems so that they don't just calcify. I I think if you do live forever at least you expect to you're less likely to reproduce at young age. You may take a first 400 500 years to start a family. So I think all this uh kind of expectation of 20 year generations and younger generations showing up will be modified as a result. We're already starting to see population dynamics change in Europe, Asia where we're not producing enough children to even maintain the population. It's interesting the way that I think that will play out maybe even a little bit different than that because part of why I haven't had kids is I was like I only get one youth where I can go hard. I have a ton of energy. Um I feel like I'm of the culture so I'm far more likely to build something relevant than I am as I get older. Genius is a young man's game as they say. And so I don't want to be distracted by something. Uh, I also don't want it to pull at my marriage. So, I'm like, ah, I'm gonna hold off for now. If I knew that I was going to live for 500 years, I might just be like, ah, whatever. Let's do it now. Because I'd rather see like it only takes me a 25, 30-year investment, and then after that, like, I get to see what they do and I get to see all my progeny and all of that. And I'm going to have plenty of time. You know, if my youth lasts for 250 years, it's like, yeah, word, whatever. I'll clock the 30 years now so that I can see how big my family gets. I think this is one where I want to think like a sci-fi writer. It gets so interesting so fast. >> It's not obvious, but I think most people procrastinate on hard work and so they would put it away as far as they could get away with. >> There there is some truth to that to be sure. um what's a a big breakthrough in longevity that you've seen that's got you really excited that this is all possible? >> There are some good experiments in animal models. Of course, they don't always scale up to human performance, but I think there is some 30 40% increase in lifespan of mice and other lab animals. It's hard to experiment on kind of bigger animals with longer lifespans. It takes a very long time to see results. But uh I think we're making good progress in understanding at least what might improve your health span and what uh changes we need to make to the genome. We study people who already have very long lifespans and find commonalities in their genomes. If those can be reproduced either pharmaceutically or through genetic manipulation maybe we can all get same at least 120. Have you heard of the Chinese doctor I think his Dr. Louu Liu um he went to prison because he altered the genome of two twin girls and >> did human cloning. >> Yes. Uh so he just put out a post on X like a couple days ago that said with 10 edits to the human genome you can give birth to a child that is immune to God it was like five things cancer um HIV I mean they were like big things and he was like it's only 10 edits to the genome. Um do you pay attention to his work at all? Do you think that's ethical? Like what do you either get excited about or worry about there? I'm behind on my science. There is so much coming out even in my domain of AI. I can't even keep up in that domain. So I definitely don't follow details of everything. I'm skeptical about his claim that cancer can be cured because cancer is like a thousand different conditions barely related. We call them all cancer but they are completely different problems. So unlikely to be the case. I haven't seen the post. Maybe he talks about a specific type but uh definitely so much of it is a single mutation. And we know some people are immune to getting AIDS virus. Exactly. Because they have a single mutation. >> Okay. So let's just say for a second that however many edits it is, it is possible. Um do you draw a line between germline editing where this is going to get passed on? Like do you have some of the same safety fears where it's like there's just too much unknown? Um where do you come down on gene editing? >> It's a lot less concerning. So for one, if there is one human with some problem, that's it. It's still just one human. If we have editing tools, whatever changes we make, we can later undo them with the same editing tool. If we made a mistake, we can go back and rectify those problems. So, I'm a lot less concerned because of impact. Worst case scenarios, yeah, there are ethical implications for the individual like he was in prison for human cloning, which is considered to be problematic because it may harm the child significantly. But humanity as a whole is not impacted directly negatively by that experiment. >> Where is AI's intersection with this? Is AI going to be a critical tool in terms of just mapping out all of the, like you said, the similarities at the genome level between people that live long and people that don't? Um, is AI going to be, you know, doing novel protein folding and going in and solving some of the architectural problems of people that are getting sick? Like where is that going to interface with longevity? All all of the above. We need to map the genome. We need to understand what individual parts do. We need to design novel drugs. Protein folding has been solved basically. But there is other things we can map on biological substrates. So yeah, at every aspect of it, we we need AI. But I think as was illustrated with protein folding problem, a narrow system can do it. We don't need super intelligence for that. What's something that's happening in AI right now that you don't think enough people are paying attention to? >> Well, I don't know what people are paying attention to. Usually after my talk, the questions I get seem to be completely irrelevant to the subject of my talk. I'll tell them that it's going to kill everyone and they ask me if they're going to lose their jobs. So, I don't think it's a good way to measure what is important. But, uh, look at uh, predictions from inside the labs. They are starting to talk about automating research process, creating junior scientists as AI agents, AI models. They are saying that externally to the lab, people don't understand just how capable systems are yet. So there seems to be a lot of indicators that internal progress is even more impressive than what we see outside. >> That's interesting. Um, what is the big anxiety that when you give your talks that people come up with? Is it just am I going to lose my job? >> It is things they already know and care about dealing with other human agents. So algorithmic bias, technological unemployment. Uh recently with OpenAI announcing that they're going to get into adult material, people are now freaking out that we're going to have artificial girlfriends. All this nonsense. >> You're not worried about that? >> Why would I worry about someone having an artificial girlfriend? [snorts] >> Okay. Well, let me paint the picture. So, I never would have guessed in a million years that giving a 12-year-old access to the internet would end up being so damaging to an entire generation. But, uh, some of the studies coming out now are terrifyingly compelling. And then there was that commercial which just brilliantly encapsulates it. I'm so sad that if I had kids, I don't know that I would have thought of it where the father's like tucking his son into bed. He's like, "All right, good night. Now, be safe." Now, remember, over in the corner is a box with all the pornography that's ever been made in the world. Don't look at it. Uh especially not the really harmful stuff. And um over here, there are going to be people that are trolling you and making fun of you. You've got to ignore them. Don't pay. And I was like, "Oh my god, that's exactly what it's like to leave a kid alone with their cell phone at night if they have unfettered access to the internet and to social media." And so given that we've got character AI that's been sued multiple times I think for uh kids that have committed suicide after interacting with their chat bots. Um I can only imagine the number of people that will end up falling in love uh with an AI system that does not feel anything back uh could turn on them could intentionally or unintentionally manipulate them. So, I can imagine that becoming problematic in ways that we just can't anticipate yet. >> Yeah, but they're very similar to problems we had before. How many men fell in love with women who felt nothing for them? We all grew up on the internet with access to pornography. It wasn't AI generated, but we somehow survived it. So, I think those are more of the same problems. They still something we should look at, but I don't think we'll all die as a result of Sam Alman having a virtual boyfriend. >> What do you think of Sam Alman? Is he the right person to have this potential godlike control in his hands? >> I don't think no human is a right person to deal with that level of power. Now, I don't think he's going to be in control, but even the stages leading us to the development of this technology already present way too much power for any individual to handle. Were you impressed or terrified when the board tried to boot him and he ended up remaking the board and coming back? >> It was fascinating to watch. But what I was observing is that it made no difference. The company was an independent entity and all the human components of that monster kept walking just the same. They replaced him with a temporary CEO. They brought him back. There was never any switch in anything, any change in direction of a company. >> Now, is there a major player in the AI space that you think is doing it right? Like uh Eliza Yudowski is very focused on AI safety. Uh you've got the anthropic CEO that's like banging the drum wanting more regulation. Um do you like any of their approaches or anybody that maybe I'm not aware of? >> So Eleazar does zero development. He's purely safety advocate and so I'm very happy with him because he's not developing any super intelligences. Anyone who is is problematic and whatever they talk about government regulation which is meaningless as a solution to a technical problem or anything else uh they might have some internal polling showing this is good for business or good for public perception but I don't think it makes a difference in terms of so solving super intelligent safety problems. Now, have you talked to Elizer or read anything like is he saying specifically, I have become afraid that this does something bad and therefore I'm not going to develop anymore? >> He was never developing AI to begin with other than publishing a very high level abstraction theory for how it could be done. Then he was like 16. >> Got it. I thought he was actually an engineer. Got it. As far as I know, he doesn't actually release any software into the world. >> Okay. Uh, now the anthropic CEO, whose name I'm forgetting, he is being accused of going after regulatory capture, that he's not very sincere and actually trying to slow this down. He just wants to make sure that the big players remain the big players. Uh, when you look at the way that he's moving, um, does that ring true or do you think no, this is somebody who's sincere about keeping us safe? They all on record as being very concerned with safety. In fact, all these companies started as safety companies. Open AI was a safety offshot of philanthropic uh endeavors, effective altruism, entropic was a offshot of that project becoming less safe. And so all of them claimed at some point that the only reason they're doing what they're doing is to improve safety. and then each one of them greatly improved capabilities of AI without proportionately improving safety. So that's that's the actions I see. >> Okay. One of the things that I worry about from a safety perspective things are obviously already bad enough and moving fast enough but quantum computing at least from my layman's perspective seems to be um gaining some sort of rapid acceleration. Um, am I just not able to understand the limitations that are self-evident to somebody educated at your level? Um, or are we really at some sort of phase transition moment? >> Well, looking at stock market value of quantum computing companies, it seems like somebody knows something on the inside. Maybe they're making good progress, but as far as AI goes, uh, we're making excellent progress with standard Vanoyman architectures. So I don't think there is a necessity for quantum computing to get us to a GI of super intelligence. It does have tremendous impact on crypto world both cryptography for security and cryptoeconomically. Uh so that's where I'm worried about quantum keeping secrets and keeping my money but not as much in terms of AI. >> Keeping your money because it will crack typical banks or keeping your money because you're largely in crypto. both actually it uh impacts all the standard encryption algorithms. We have postquantum encryption but we haven't switched to it for most interesting applications. >> So give me your stance on Bitcoin. >> Buy some. [snorts] >> That's very clear. Uh what so when you look at gold and you look at Bitcoin, why Bitcoin over gold? >> You can make more gold. As the price of gold goes up, I can make as much gold as you want. I can convert our matter into gold at very high price. I can exploit asteroids in the universe. I can get gold out of oceans. There is a lot of gold which is very expensive to get. But if a price of gold is high enough, I can produce more and more. Bitcoin is not subject to the same pressures. It doesn't matter if one coin is a trillion dollars. There is still a limited supply. >> Okay. But what about people who say uh gold at least can be made into other things. Gold has survived for thousands of years. Bitcoin is like 15ish years old. Uh and it's not backed by anything. Can't turn it into anything. It's just literally got no other use. >> It's a dedicated app. So if you have an app which does everything, it's usually not good at anything. The fact that I can make jewelry out of it is not an important feature for me storing my wealth in it. Whereas this has capabilities gold historically lacked. If I can pass a billion dollars to you right now for $5 immediately across borders, I cannot do this with gold. All right, talk to me about the postquantum encryption. Every time I'm hugely in Bitcoin, uh every time I hear the word quantum, I'm like, h like, oh, this just makes me paranoid. uh given that it's rightly difficult to make changes to Bitcoin um how are we going to get to post quantum encryption on Bitcoin in a way where the community actually comes to consensus and doesn't create a problem for itself. So once we get integer factorization running on quantum computers, you can see what size integers we can factor that tells us how close we are to cracking standard Bitcoin encryption hash functions. If we're getting close to what would essentially destroy the network, I think it's like any other emergency. We have history of fixing Bitcoin software, then an obvious problem was discovered. I think at one point somebody managed to print a trillion coins or some nonsense like that. Immediately a patch was distributed. Everyone adopts it because it's the only way to go forward. And I think that's what we're going to see. As long as we have that available, tested the moment there is a good strong signal that you have no choice but to accept or lose everything. It's self-interest once again. >> Okay. So on that, where are we? What how many integers can a quantum computer handle and how many would it have to be able to handle in order to be a threat to Bitcoin? >> I haven't kept up with the latest breakthroughs. Last time I looked it was a laughably small number. Quantum computers were factoring like I don't know 15 literal 15 not even 15 digits. So, uh, unless there was tremendous progress since then, I think we're still good, but progress could also be exponential. So, it could come very quickly, >> right? Okay. So, if you had a message to the Bitcoin community, would it be let's move on this now? There's no reason to wait until it's an emergency, there's a very clear path, or uh, is it like, well, I'm just sort of on the ride with everybody else? >> I think we still have time. I don't think it's uh pressing as much as AI problems we're dealing with. So if AI is 2 years away, I think we may not be that close with quantum computers just yet. But again, it's one breakthrough away. If some company comes up with something much more powerful, it may shift very quickly. >> Okay. And what what's the term for that? the number of integers it can process or >> uh the trigger would be what size what size integers could be factored, how many bits. >> Got it. Okay. Um All right, Roman, this has been just absolutely incredible. What message do you want to leave people with to get them to take action? What is your best pitch to get them to sign your petition? So if you are in a position of developing more powerful AI systems, concentrate on getting your money out of narrow AI systems, solve real problems, cure a cancer, figure out how to make us live longer, healthier lives. If you are developing super intelligence, please stop. You're not going to benefit yourself or others. Uh the challenge is of course, you know, prove us wrong. prove that you know how to control super intelligent systems no matter how capable they get, how much it scales. Uh if you can do that, then it completely changes the situation. But as long as no one has came up with a paper, a patent, even a rigorously argued blog post, I think we are pretty much in consensus that we don't know how to control super intelligent systems and building them is uh irresponsible. >> Amazing. Where can people connect with you? >> Follow me on Twitter. Follow me on Facebook. Just don't follow me home. [snorts] >> Nice. Awesome. Roman, thank you so much for the time today. I really appreciate it. Everybody at home, speaking of things I appreciate, if you haven't already, be sure to subscribe. And until next time, my friends, be legendary. Take care. Peace. If you like this conversation, check out this episode to learn more. In the next 1,000 days, AI will not only replace a startling number of humans in the workforce, it will make the entire structure of our economy obsolete. That is the unnerving claim of today's guest, Emmod Most.