Transcript
DWKIKf3MAV0 • The AI Scaling Problem: Why Current LLMs Aren’t Truly Intelligent
/home/itcorpmy/itcorp.my.id/harry/yt_channel/out/FoundationModelsForRobotics/.shards/text-0001.zst#text/0070_DWKIKf3MAV0.txt
Kind: captions Language: en If you follow tech at all, or you know, frankly, if you just exist in the world today, it is absolutely impossible to escape the constant drum beat of AI progress. I mean, every single day, right, there's another headline, another mind-blowing demo that just seems to defy what's possible. It can honestly feel like we're all strapped into this unstoppable high-speed train just rocketing towards a future where AI does everything. The pace is, it's just dizzying. And it's not just the headlines. We're hearing these incredibly bold claims from the biggest names in the game. You've got Mark Zuckerberg saying meta will have AI that can do the work of mid-level engineers in like a year. Tech leaders, influencers, they're all saying every new model means AGI, artificial general intelligence is right around the corner. We're talking months, not years. It all feeds this feeling that human level AI isn't some sci-fi dream anymore. It's an imminent reality. And what really cements this idea, what makes it so hard to argue with are the benchmarks. These are the hard data points. First, it was passing the bar exam, then the SATs, then suddenly it was acing coding challenges that would take a human engineer hours to solve. And now, now we're seeing these models solve PhD level math and physics problems that would stump most human experts. I mean, when you see a machine ASA test designed to measure the absolute peak of human intellect, you can't help but ask, are we about to be replaced? But what if that whole story, the unstoppable train, AGI being just around the corner, is just fundamentally wrong? What if the very yard stick we're using to measure all this progress is completely flawed? Let's really dig into that. Because according to AI researcher Aiden Meyer, whose work we're diving into today, this whole idea of an all- knowing AGI isn't just hype. He says it's a lie. It's A fundamental misdirection that's leading the entire field and all of us down the completely wrong path. So, how on earth could that be true with everything we're seeing? Well, in this deep dive, we're going to break down this really powerful counterintuitive argument. We're going to start by deconstructing that myth of unstoppable progress and show why we might be measuring the wrong thing entirely. Then, we'll explore a new and I think much more exciting vision for what AGI could be. unpack the three essential keys to get there and finally rethink what the future of AI should even look like. Okay, so this brings us right to the core of the whole argument. We are measuring the wrong thing. How can an AI ace all these impossible exams and still not be on the path to real general intelligence? Well, it all comes down to a basic misunderstanding of what we're actually testing. Our benchmarks from the LSAT to coding competitions are missing half of the equation. This slide splits the very idea of intelligence in two. And this is so key. On one side, you have applying knowledge. This is all about performance on a fixed task. It's what every single AI benchmark today tests. Can you answer this question? Can you solve this problem? Can you pass this exam? And yeah, our current models are amazing at this. But on the other side is the much much more important skill, acquiring knowledge. the ability to learn something genuinely new, to adapt, to grow. And that is the part we completely ignore. And look, this isn't some new definition of the word. Just look up intelligence in the dictionary. It's the ability to acquire and apply knowledge and skills. It's two parts, both equally important. By focusing only on the apply part, we've basically created the world's most sophisticated parrots. They are incredible at repeating and remixing the vast ocean of human knowledge they were trained on, but they're almost completely incapable of learning anything new after that training is done. Okay, so if the goal of building this all- knowing digital oracle is basically a dead end, what's the alternative? Well, the vision Meer proposes is a pretty radical shift. Instead of building an agent that knows everything, he says we should be building an agent that can learn anything, a lifelong learner. So, what does that actually look like? This new AGI wouldn't be some static database you just ask questions. It would be a dynamic agent. It would have its own goals and it would decide what it needs to pay attention to in order to achieve them. It would teach itself, learn from interacting with you and then use that new knowledge to keep learning on its own. So think less of a tool and more of a partner that you actually grow with. And this is a perfect example of the realworld limits of what we have now. Meyer talks about asking ChatGpt for ideas for his YouTube channel, and the suggestions are always just terrible. They're bland. They're generic, and they have nothing to do with his unique style. And that's because the AI can't actually learn him. It can't look at their past conversations, get his context, and create something new. It can only give him a remix of what's already popular on YouTube from its training data. It can apply knowledge about YouTube video ideas, but it can't acquire new knowledge about Eden Meyers video ideas. You know what's wild? This whole idea of focusing on learning isn't some futuristic concept. It's actually going back to the very roots of AI. Early research way back in the 30s and 50s was obsessed with creating systems that could learn and adapt, like a little robotic rat figuring out a maze through trial and error. The goal was to copy how animals learn continuously. It's only in our modern era of big data and giant computers that we've shifted to this train once on the entire internet then deploy a frozen model approach. We've traded the ability to learn for the ability to know. Okay, so if we accept this new vision of AGI as a lifelong learner, how do we actually build it? Well, Meyer lays out three critical and mostly neglected areas of research that are absolutely essential to make this happen. These are the three keys that could unlock a very different and honestly much more powerful future for AI. And hey, if you're finding this kind of deep dive valuable, make sure you're subscribed for more explainers that cut through the hype just like this one. The first key is the most direct one, continual learning. It's simple, really. The agent should never stop learning. Right now, AI models have two separate phases. A massive offline training phase and then an online deployment phase where its knowledge is basically frozen in time. Continual learning says those should be the same thing. Learning and doing should happen all at once forever. Now, you might be thinking, hold on, but current models can learn through things like fine-tuning or in context learning. But these are more like band-aids than real solutions. When you try to fine-tune a model with new data, you run into this huge problem called catastrophic forgetting. So, imagine you teach a model to be an expert on Shakespeare. Then you try to teach it about Jane Austin. In the process, it will literally overwrite and forget everything it knew about Shakespeare. It's not adding knowledge, it's replacing it. And over time, this also leads to a loss of plasticity where the model just gets worse and worse at learning anything new. As for in context learning, that's just cramming for a test. The information disappears the second the conversation moves on. These are not paths to lifelong learning. All right, this brings us to key number two, which is a little more abstract, but just as important for building something that thinks like a human. To really learn about the world, an AI has to experience it not as a jumbled mess of data, but as a single continuous, unbroken stream, one moment flowing into the next, creating a personal timeline of cause and effect. Now, let's contrast that with how we actually train large language models today. It is literally the exact opposite. We scrape billions of random disconnected bits of text from all over the internet. A paragraph from a poem, a form post about fixing a car, a cooking show transcript. Then we toss all of it into a giant digital blender, shuffle it up, and feed it to the model in these random huge batches. There is no timeline, no narrative, no journey of discovery. In training this way has huge consequences. It's why a concept like episodic memory, your memory of the events in your own life, is basically meaningless for an AI. To understand why that matters, Meer explains how the very idea for his research came to be. It wasn't a generic prompt. It was his unique single stream of experiences. A specific interest he had led to a certain grad program which led to specific conversations which were influenced by a startup he worked on years ago. That chain of events, his personal episodic memory, created something new. An AI trained on a random batch of the internet could never come up with that because it has no personal history to pull. And that brings us to the third and final key. This is all about changing how our models get smarter. Right now, they get smarter with more and more data. What we need are algorithms that get better with more compute. Basically, more thinking time. What's so interesting here is the difference in philosophy. Current scaling only works if you have more compute and way more data. It's incredibly inefficient. I mean, we've scripted basically the whole public internet and big tech companies are already worried about running out of new data. That's not a data problem. It's an algorithm problem. The real goal should be algorithms that scale with compute alone. Think of it like this. Current AI is a student who can only get a better grade by reading more textbooks. The goal is a student who can get a better grade by thinking more deeply about the same textbook. Given more processing power, this ideal agent could pull more wisdom and new ideas from the exact same set of experiences. It's the difference between thinking deeper, not just reading fast. So, when you put all these pieces together, the bad benchmarks, the forgotten history of AI, these three keys to learning, it paints a totally different picture of the future. The current north star of the field, this chase for an all- knowing oracle trained on all human data, it's not just ambitious, it's the wrong goal entirely. We need a new north star. So what's the big takeaway here? It's that our current benchmarks are just measuring static performance, not the dynamic process of intelligence. True intelligence at its core is about the ability to acquire new knowledge. So our goal shouldn't be an all- knowing agent, but a general purpose learning agent. And to get there, the field has to shift its focus to solving these hard fundamental problems. Continual learning, single stream experience, and scaling with comput. This really leaves us with a fundamental choice about the kind of future we want to build. Are we chasing this sci-fi fantasy of an all- knowing oracle that just spits out answers? A tool that kind of replaces thought, or are we trying to create a true lifelong learning partner, an agent that can grow, adapt, and discover new things right alongside us? The path we choose from here is going to define the next era of AI. If this kind of analysis, the kind that cuts through the noise and questions the core assumptions of a field, is something you find valuable, make sure you subscribe. We'll be bringing you a lot more explainers that go beyond the headlines to give you a deeper, more nuanced view of the tech that's shaping our world. Thanks for joining this deep dive.