🎓 Learn AI In 10 Minutes A Day - https://www.skool.com/theaigridacademy
Get your Free AGI Preparedness Guide - https://theaigrid.kit.com/agi
🐤 Follow Me on Twitter https://twitter.com/TheAiGrid
🌐 Learn AI Business For Free AI https://www.youtube.com/@TheAIGRIDAcademy
Links From Todays Video:
Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos.
Was there anything i missed?
(For Business Enquiries) contact@theaigrid.com
Music Used
LEMMiNO - Cipher
https://www.youtube.com/watch?v=b0q5PR1xpA0
CC BY-SA 4.0
LEMMiNO - Encounters
https://www.youtube.com/watch?v=xdwWCl_5x2s
#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience
Оглавление (3 сегментов)
Segment 1 (00:00 - 05:00)
So, has Elon Musk's XAI created a recursively self-improving AI? Let's talk about it. So, with the latest update from Gro 420, many individuals were wondering, what is this model? What is it doing? And I think I found something pretty interesting. So, if you take a look at this post here, some people are saying that Elon Musk is sharing new delusions, but after some digging, that could not be further from the truth. So, if you were aware, Elon Musk released Grock 4. 2 agents. It's around four different agents that work together to come to a result. And what's super interesting is that you have you know Elon Musk saying that the foundations of this model are such that it is able to improve every week. So the recursive intelligence growth will be strong. Now what's interesting about this is that you can see here and so Elon Musk's tweet says that the foundations of 4. 2 are such that it is able to improve every week. Now you can see here the quote tweet says that I'm not fully sure what he meant but probably not continual learning in the strict sense where deployed model updates its weights live over time. More likely it's fast iteration loops. Now you can see that the actual, you know, Reddit post says that Elon Musk is showing you delusions and it's completely understandable because Elon Musk will frequently kind of exaggerate a little bit certain timelines. I mean, he said that a lot of things are going to happen and they haven't just yet. So when he says this about AI, it may be no different. But if you actually look at all of the information around this statement, it actually points to this statement being more true rather than false. So, we need to take a look at exactly what it said here. And you can see here that Jimmy Bar has said that it's his last day at XAI. And Jimmy Bar is a guy who used to work at, you know, XAI. And I believe that he was the co-founder. So, the most interesting thing comes from this. Okay. He says that, you know, he was excited to work there, pushed humanity forward. But you can see what I've highlighted. He said, "We are heading to an age of 100x productivity with the right tools. " and that recursive self-improvement loops are likely to go live in the next 12 months. Okay, that is what he said just as he was leaving XAI. So remember, recursive improvement loops are basically where an AI system that can improve itself and then use that improved version to improve itself again, thus creating a loop that compounds. And so the basic cycle will be an AI model helps design/train a better version of itself and that better version is smarter at designing/training. So the next iteration improves even faster. And when you repeat that, you're potentially accelerating beyond human ability to keep up. Now, right now, you know, humans are still the primary bottleneck in AI development. Researchers design experiments, interpret the results, write the code, and then run the evaluations. And the claim here is that within 12 months, these AI systems will be autonomously running that entire loop without needing humans at all. AI doing research essentially. Now, I think this is going to be pretty interesting because of course he said that this is going to be one of the most busiest and consequential years. And remember, this is someone who has incredible credibility. He's not just anyone. He's one of the co-authors of the original Atom Optimizer, one of the most cited machine learning papers ever. And of course, like I said, he was a founding member at XAI. So, so that is why this tweak or the coverage it did. It's not like Elon is trying to hype us up some timeline here. You can see that Jimmy Bar is actually stating the look recursive improvement loops go live likely in the next 12 months. now actually decided to take a look at Elon Musk's other employees, see what they were tweeting, and decipher the information. Now, other employees from XAI are essentially talking about the continual learning problem and how recursive self-improvement is not far away. And all of this leads me to believe that Elon Musk's team somehow somewhere, I'm not sure how they've done it, but have made some kind of improvement that allows them to self-improve. Now essentially you can see this is the Shant Patel someone who works at XAI or someone who used to work at XAI but he no longer works there and he actually said and this was before he left that continual learning is largely a context compression problem. Compressing endless multimodal bits streams into dense reusable learning representations is crucial to deploying such systems in the real world. Building a unified learning representation will be necessary towards creating an AI that can continuously act and learn from the real world in a safe and reliable manner. So this is super interesting because the problem that he's describing is that you know current AI models are essentially frozen in time. Once training ends the weights are fully locked. They can't learn from new experiences without a full expensive retraining run from scratch. Every
Segment 2 (05:00 - 10:00)
interaction you have with an AI it forgets it immediately after. No accumulation of knowledge, no growth. So what context compression actually means here is that every second in the real world is a fire hose of data. Video, audio, text, sensors, all simultaneously. And that's the multimodal bitstreams. The problem is that you can't just dump all of that into an AI forever. It's computationally impossible. So the real question is and that the real question becomes how do you take that flood and squeeze it down into a compact learn representation that actually sticks in the model's weights. Think of it like the difference between a human who experiences something and genuinely learns from it versus one who just reads a transcript of what happens and then immediately loses the notes. And the unified learning representation matters because current models have separate systems for handling images, text, and videos. For true continuous learning, you need one unified representation that can absorb all of it together because the real world doesn't separate all of those things. And of course, Elon Musk basically says bullseye. And that's what XAI was trying to build. An AI that lives continuously and learns from reality. Tesla's Optimus, autonomous systems, Grock integrated with X realtime data stream. All of that requires that problem to be solved first. Now, take a look at this. Okay, someone else who worked at XAI said that they left and they're building something with the others that left XAI. So, Roland who left actually started a new company. And I bet you will not guess what they're working on. Take a look here. the guys who used to work XAI, you can clearly see and I mean if you were skeptical that XAI made some kind of breakthrough, clear advancement in AI capabilities, this is the statement from Roland who used to work there and he says, "During my time at XAI, I got to see a clear path towards hill climbing any problem that can be defined in a measurable way. And at the same time, I've seen how raw intelligence can be lobbomized by the finest human errors introduced in the agent harnessed. The gradients must flow. Learning shouldn't stop at the model weights, but continue to improve every part of an AI system. Today, I'm launching Neuroline, the missing infrastructure that enables AI native software to continuously self-improve. I mean, if this isn't the biggest, okay, clearly they saw some kind of way to self-improve with AI and they're probably going to be implementing that, I don't know what else is. I mean, of course, Elon Musk is, you know, pretty excited about AI, but when you've got three to four employees all talking about recursive self-improvement, I think it's very clear that they may have figured something out. Now, if we're going to look at other companies as well, and the only reason I'm including this is because it's pretty interesting, is that GPT 5. 3 Codeex, you can see here OpenAI basically says that this was their first model that was instrumental in creating itself. The Codeex team used early versions to debug its own training, manage its own deployment, and diagnose test results and evaluations. And their team was blown away by how much codeex was able to accelerate its own development. And this is what people are talking about. when the code gets good enough, the models are going to be able to improve themselves. — At the end of 2025, so just a few months ago, I think we've really begun um an entirely new paradigm or era of the technology, which is um really beginning the era of recursive self-improvement. And um I think we're really starting to see this in the development of our models. I think you know most of the major labs are seeing this in the development of their models which is that the models themselves have now become instrumental in accelerating the process of producing the next AIs. And so you know the sort of from the outside what you see is maybe an acceleration of overall development velocity an acceleration in at which the speed at which the new models are being released. Um internally what we see is just dramatic speedups. I think the sort of um the productivity of an individual researcher has grown dramatically. Um and we don't expect that to stop. I think this era of recursive self-improvement will uh feel like everything is speeding up and I think that um the technology will continue developing very quickly. — Take a look at what Peter Diamandis said. If the predictions for recursive self-improvement in 2026 are true, every prediction curve we have accelerates dramatically and every governance framework, every safety protocol and regulatory approach is already obsolete. We're building bricks for a car that is about to become a rocket. Take a look at what you can see here. Gary Tan says, "It's amazing to see the cutting edge jump to recursive self-improvement when the majority of developers in the real world are still mostly skeptical of vibe coding for production at all. " What he's saying here is that it's crazy how we live in two parallels. On one side, you have individuals who are saying
Segment 3 (10:00 - 14:00)
recursive self-improvement, but on the other side, you've got people who are even skeptical to use these coding tools because of bugs and all the other issues mentioned with vibe coding. It truly is a very interesting time. Of course, Dario Abday at the World Economic Forum essentially said that in 6 to 12 months, it's probably going to be Claude doing all of what software engineers do end to end. Of course, if that is true, it's going to be super interesting to see how that affects AI development. — We are now in terms of, you know, the models that write code. I have engineers within Enthropic who say, I don't write any code anymore. I just let the model write the code. I edit it. I do the things around it. I think I don't know we might be 6 to 12 months away from when the model is doing most maybe all of what sues do end to end and then it's a question of how fast does that loop close. — Take a look at another set of tweets that I found and slow developer Haidider tweeted about the fact that a deep mind researcher was predicting that 2026 was going to be the year of continual learning. You can see he tweeted about this on four separate occasions. Um, and it was super interesting. He said that we're going to make massive progress on continual learning. Maybe he actually just tweeted about it twice, but it's very clear that, you know, when you look at the fact that Google also released a paper called introducing nested learning, a new machine learning paradigm for continual learning, that was something that was also super duper interesting. So it's pretty interesting to see you know nested learning which is the new approach to completely avoiding the issue of catastrophic forgetting where when you learn new tasks you essentially forget the old ones and then of course not just Google are tackling that front. So we got XAI, Google. Now of course we do have Meta which is you know on their paper continual learning continual learning by sparse memory fine-tuning which is where it you know combats this catastrophic forgetting by selectively targeting updates to memory slots that are highly activated by a new piece of knowledge. Um and it is just really more effective. And overall we're basically seeing that this thing is pretty crazy. And what I want to show you guys as well, I found a clip of uh the Lex Freedman podcast where they were talking about the state of AI and they mentioned that cursor on their blog post. They actually did mention something briefly about continual learning. I don't think most people saw this, but take a look. It was super interesting, but just take a look cuz it's pretty crazy. — Yeah. — And we talked about continual learning and stuff. They had one of the most interesting like two sentences in a blog post which is that they had their new composer model which was a fine tune of one of these large mixture of expert models from China. You can know that by asking gossip or because the model sometimes responds in Chinese which none of the American models do. And they had a blog post where they're like we're updating the model weights every 90 minutes based on real world feedback from people using it which is like the closest thing to real world RL happening on a model. So you can see here that is Grog 5 going to be something that is going to learn almost immediately. We do know that Elon Musk is working on this model. I say Elon Musk but of course Elon Musk XAI of course not trying to discredit the researchers. This is another tweet from earlier last or later last year. And you can see he says dynamic reinforcement learning is important. Rock 5 like smart humans will learn almost immediately. I think it's very clear that at XAI they may have figured something out that is on the verge of continual learning. We've seen different tweets. We've seen those who've left AI actually go to a startup that is working on self-improving AI agents. We've seen Elon Musk tweet about this several different times. And not only that, we've seen new research papers published that actively talk about this as well. So, if that is true, what does Grock 5 look like? I mean, it's going to be super interesting to see across the stack how they manage to integrate their AI and the kind of results they do get. So whether you think it's delusions or truth, one thing is for certain, the future is very, very interesting. With that being said, I'll see you guys in the next