# Big AI News: Google Gemini 2.0 Date LEAKED! OpenAis New AI Project, New Robots, And More

## Метаданные

- **Канал:** TheAIGRID
- **YouTube:** https://www.youtube.com/watch?v=rDooa0wJhhw
- **Дата:** 24.11.2024
- **Длительность:** 31:55
- **Просмотры:** 37,713
- **Источник:** https://ekstraktznaniy.ru/video/13687

## Описание

Prepare for AGI with me - https://www.skool.com/postagiprepardness 
🐤 Follow Me on Twitter https://twitter.com/TheAiGrid
🌐 Checkout My website - https://theaigrid.com/


Links From Todays Video:
https://x.com/tsarnick/status/1860071730389352939
https://x.com/tsarnick/status/1860066656682492239 
Expand video - https://x.com/runwayml/status/1860074950167818617 
New AI Agent - https://x.com/hcompany_ai/status/1858907025436205278 
Text to 3d - https://www.nvidia.com/en-gb/gpu-cloud/edify/ 
Chatgpt creative writing - https://x.com/kyleshannon/status/1859751854562750829 
Demis hasabis on discovery - https://x.com/slow_developer/status/1860012871209603468 
Proxie cobot  - https://x.com/BradPorter_/status/1859265814337331709 
Brett - https://x.com/adcock_brett/status/1860527623908253729  


Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and

## Транскрипт

### New AI Model []

so one of the first stories in AI wasn't in fact the new R1 model there is another strange competitor that has seemingly come out of nowhere to challenge 01 previews so someone tweeted yet another seventh grade math problem that only 01 mini can solve and they basically showed chat logs where 01 mini was able to be the only model that was able to get the answer correctly but if we look at this model now this model is one that somehow manages to get the right and it says check reasoning process and the possible values that X can take to have exactly 30 zeros in the first 100 terms of the sequence are minus 3 2 6 and 7 and I'm not going to go into all of this reasoning details but what I will say is that this is super interesting because the kpu is essentially MSA kpu I'm not sure if that's how you say it but this is a model that was actually announced quite a long time ago now I'm going to show you guys exactly what I mean by that so I think it was around five months ago or so five six months ago we basically got this bombshell screenshot from this lab that apparently managed to develop some kind of I wouldn't say I don't even know if it's an algorithm I don't know if it's a way to use AI identically they haven't exactly revealed how they are using the model but they seem to have some kind of framework that allows to prompt the model in such a way that it is able to get far better reasoning outputs than anything we've ever seen before for example you can see on the GSM 8K you got like 97% on the math benchmark at the time that was state-ofthe-art at the BBH all of those ones they around to 95 to 100% And I remember at the time of this being released there was currently a lot of skepticism because you know you can't just come into the AI space drop a screenshot saying look my model beats yours on several benchmarks and then just disappear into the Wilderness nobody's going to believe that thing even exists but that's exactly what they did but now they seem to have come up with something that actually works and is actually a real product so a simpler question that they did ask this model was how can a person with no hands clean their ring finger and then you can see that this model reasons with itself and then it says this question contains a logical impossibility a person with no hands would not have ring fingers to clean since ring fingers are part of hands it's similar to asking how a person with no feet tie their shoelaces the premise is self-contradicting because the absence of hands means the ring fingers do not exist so I think this is the kind of model that trick questions simply don't get the model because I think that this is the response that we wouldn't really get from other models I tried to use this question in Google Gemini and you can see that the response I managed to get I can also show you guys this one here where I actually decided to put this exact query into Claude from anthropic and you can see right here that the model didn't respond right now I personally do think that yes this model does use some Advanced reasoning if it's able to solve problems are only 01 is solved but I do think one of the key things a lot of people are now you know going into that gray area with is missing the true intelligence of models due to I wouldn't say flimsy benchmarks but I would say due to vag benchmarks because for example right here what exactly is this question testing is it testing you know empathy is it testing reasoning because some people would argue that if a model says you can't wash your hands if you have no hands yada y yada some people would say that's completely right but other people would say that is completely unempathetic and I'm going to show you guys why so I don't know if you guys know about this model this one is the R1 light model this model is basically quite like opening eyes1 model and essentially we're able to see what the model is thinking before it manages to respond so you can see right here that it literally states that how can a person with no hands clean their ring finger and at first glance it seems pretty straightforward but when you think about it it's actually quite a complex problem and it says what does it mean to have no hands and does it mean that this person has no hands at all or perhaps they have some Prosthetics or some sort of assistive devices the answer might vary depending on the specifics but for the sake of this discussion I'll assume this person doesn't have hands and is looking for a way to clean their ring finger independently so basically what I'm trying to say here is that if this question allows the model to basically assume that the person has some kind of prosthetic I don't think that the model should be graded wrongly on that answer I think we should have you know questions that are really granular in terms of the preciseness and thus the expected output so this is something that I mean I'm really excited for this because if the model is you know Advanced at reasoning and is on 0 one's level and if they do manage to get to even a fraction of these capabilities I think this is going to completely change the game because they've got an agentic framework wrapped around AI and it's quite likely that is going to boost model performance even more now in terms of model performance going even higher something that was rather crazy was of course this so this is a LinkedIn post from dber Ren the regional head of Google Cloud AI customer engineering and he actually made a mistake that is now pretty viral on social media so the crazy thing is that he said Gemini experimental version 1121 has been gaining great momentum leading ranking on chat Arena and please stay tuned for Gemini 2. 0 next release during the second week of December now I do know that this information is quite likely true because this information was changed so this original post that was posted on I think LinkedIn was actually edited to say stay tuned for Gemini soon um this was the original tweet so this is something that is probably going to be real and I'm not going to lie we do know that Google Gemini is just around the corner because they have been teasing it for quite some time and we've seen their recent models managed to surpass open AI on certain benchmarks so I wouldn't be surprised if we do get Gemini 2. 0 during the second week of December meaning that would be between the 10th of December to the 13th of December now of course there is the possibility that they could potentially change this due to the fact that with AI releases there is a lot of competition for the spotlight a lot of the times when we see AI models being released there is open AI trying to you know push their models on the same exact date so it's quite likely that if Google are going to be releasing Gemini 2 open AI might release something just to try and upstage Google as they have in the past like literally the last time open openai upst Google was when Google released search grounding it was something that I was supposed to cover but due to all of the openai news that day I literally didn't even have time to so this is something that can happen again which means that during the week of December I think that it's quite likely we get Gemini 2 and it's also quite likely that we get a major opening ey release because they usually try to upstage Google releases now of course Google might actually change the date to throw off openingi so that is something to pay attention to because we know that these you know the these companies are really Fierce and they're really competing for a lot of stuff right now so it's going to be super interesting to see how that release is but if you haven't I would recommend trying out Gemini experimental because it is definitely one of the models that is a step ahead of other models and for those of you that say it can't reason that well I've actually changed my prompt structures when using Gemini and I've gotten better responses out of it so I think a lot of the times when we think a model isn't that good we just need to try different methods of prompting and then we can see that it usually manages to give us what we want so what could

### OpenAI Browser [7:52]

opening ey be doing to upstage Gemini well apparently they are going to be launching a browser very soon I don't know about you guys but I think you guys might be paying no attention to the AI browser space I've tried to talk to various individuals that are not in the AI space about the types of things that you can do with AI browsers and they just haven't been paying attention so essentially what this article States here is that openai is considering developing a web browser that it could combine with its chatbot and has separately discussed or struck deals to power search features for travel food and Retail websites according to people who have seen prototypes or designs of the product so essentially what we're looking at here is a situation where open AI are looking at once again to upstage Google by launching their own browser now this couldn't come at a worst time because I want to show you guys a piece of news that is pretty crazy so one of the things that Google is unfortunately suffering from is apparently it's Monopoly so there was this recent case about Google's Monopoly and apparently they might have to sell Google Chrome to end this search Monopoly so it will be interesting to see if Google has to sell Chrome and then open AI steps in at that moment because that would be absolutely incredible so things in the search space are really changing so I think that this is of course something to pay attention to and there's also a decent amount of information that is with this so one of the moves that openi has recently taken was that they've actually hired two people that were instrumental to Google success and I think this is rather interesting considering the fact that there are some reports that they aren't even close to a browser so I mean it's going to be super interesting to see exactly what goes on here I mean when taking a look at this I am really surprised that opening ey is trying to really upstage them on like just every single matric that there is and I mean it's so surprising because when we look at Google's market cap it's like something I don't even know if it's a trillion dollars I think last time I checked it was a trillion dollars but a trillion dollar company getting you know taken on by this small fast moving company that's just a few billion dollars is absolutely incredible and I do wonder if Google are going to take some steps to really change this and you can see here recently this was the person that um you know worked on the VP of engineering for Google and worked at the browser company you know software engineer and now this person is working at openai so this is going to be really interesting to see what kind of things that they do now the article does state that openi isn't remotely close to launching a browser but we do know that openi tends to work with secrecy so I wouldn't be surprised if they do now of

### Perplexity Pro Search [10:40]

course in the browser space there was also this piece of information which is basically where you can now search like a pro with perplexity Pro search so I'm not sure if you guys know what perplexity browser is a lot of people that I personally know do know what this is but essentially what we do have here is a way for you guys to shop more effectively so a lot of times the reason that people actually use perplexity is because this is the kind of model that allows you to find things really quickly and it means that you can save a lot of time now with this what you can do is you can use the AI Vision model to snap a picture of something and then you can search for this model essentially go ahead and buy that specific model so you can say I need this item but I need it under $100 I need it in this way color and you can simply use this model to buy and check out really quickly so I am wondering if this is going to be the future of e-commerce where you literally just say okay I'm looking for a shirt I want it to be this texture that you know uh material I want it to be you know this thick or this thin and then the AI gives you a bunch of different products from websites and then you literally just click and it's like oh I know exactly what shirt you like and it basically shows you one you buy it and it gets straight delivered to your door I mean I wouldn't be surprised if there are some generative apps out there that are being built like that in fact I think I did see one prototype of a company that was building something like that like literally you just you literally just ask for what you want like oh I need a new washing machine oh can it be $400 because a lot of the times and this is something that I've personally experienced myself is that you spend so much time searching for the perfect product online wouldn't it be so much easier to just say I need a product that is XYZ the AI searches across the internet and you get a response of the top 10 things that are specific to you so I will love to see if this product

### Phases of AI [12:28]

does get implemented so then of course we had this chart that basically shows the next phases and the next growth areas of AI so we've got three phases of AI you can see we've got this phase one here of course someone else is going to explain it in a minute but of course you can see here we've got phase one so you can see right which is research attention is all you need all of those papers phase two which is what we're just ending so you can see gpt3 GPT 4 Claude uh 3 you can see that this was you know a really nice area this is another scur and you can see this is actually quite like the chart that I uh put in another video um this is basically where we're on this new S curve and you can see that 01 is basically at the bottom of this S curve so this is basically going to be where we are in phase three so you can see uh yeah this is basically similar to what I showed the other day obviously this is a lot more detailed but you can see that there's going to be a lot more Innovations and scaling coming along the way and basically this was from uh the scale a AI CEO um Alexander Wang and he's basically saying the scaling phase has ented and we're moving towards the innovating phase where other breakthroughs are going to lead to superintelligence in six years or less I

### Modern Era of AI [13:31]

like to sort of break down and to contextualize the modern era of AI into sort of three major phases so the first phase uh was research which really was roughly 2012 until 2018 um so basically lasting you know going from uh the first deep neural network uh Al uh alexnet which was really trained to do basic image recognition you know this was the era when all AI could do was tell you if there were uh cats and YouTube videos and that was you know this incredibly powerful thing all the way up until the first GPT model so the Transformer and the first uh GPT model trained by Alec Radford at open AI that kicked off the next phase scaling so the first six years 2012 to 2018 was sort of This research phase then 2018 to 2024 really up until today is the scaling phase you know the amount of uh resources going to these models grew more than 10,000 fold over this time period um last year or this year I should say something like $200 billion is aggregate going into training these models and this has resulted in incredible advancement in performance um you know we've gone from gpt1 which was nothing special to uh to 01 which is sort of PhD level in math and computer science and then 01 I think kicks off this new phase of you know we sort of think about as an innovation era which is basically now until super intelligence we'll see if that's six years or maybe even a bit less than that um the Hallmark I think here is that you know we're spending $200 billion on the models we probably can't spend a lot more than that you know probably can't spend you know 200 trillion on the models so there's only limited amounts of scaling left in terms of orders of magnitude and so we need corresponding uh Innovations to sort of come alongside uh definitely Advanced reasoning and test time compute uh is one of those uh and we think there's probably a few more handful of others will get us you know to Super

### Scaling Models [15:14]

intelligence now interestingly enough the C- here CEO Aiden Gomez actually says that scaling models has entered the flat parts of the curve but these models are already so smart that it's actually taking them a little bit of time to assess their actual output puts and this is now going to be applied to different research domains so it's interesting to see what other CEOs are saying about the same Paradigm that we're currently entering where are we in scaling laws like how much capability Improvement do you expect over the next few years we're pretty far along I would say like we're starting to enter into a sort of flat part of the curve um and we're certainly past the point where if you just interact with a model you can know how smart it is like the vibe checks they're losing utility and so instead what you need to do is you need to get experts to measure within very specific domains like physics math uh chemistry biology um you need to get experts to actually assess the quality of these models because the average person can't tell the difference at this stage between Generations yes like there's still much more to go do uh but those gains are going to be felt in very specialized areas and have impacts on more researchy um more researchy domains I think for Enterprises and the general sorts of tasks that they want to automate or tools build the technology is already good enough or close enough that a little bit of customization will get them there now

### Runa H [16:44]

there was also this company that actually went under the radar so this is Studio a platform for developers to run AI automations at scale and Runa H which is apparently the most advanced AI agent to date which is going to be navigating the web and interfaces through pixel level interpolation and semantic understanding and you can now turn your instructions into humanlike action with Precision so for those of you guys who are wondering why we haven't managed to get certain agents yet other companies are also working on certain Integrations with certain applications and you can see that this is a company that is working slowly behind the scenes that is able to perform really well at a bunch of different task I've noticed that a lot of these companies don't just come out with their stuff they usually have this in I guess you could say like alpha or beta and of course this is one of the companies that has it right now and I think they're alpha or beta stage so this is going to be something that I think I'm excited for a lot because while yes AI agents are really good and yes they work really well a lot of the times we end up using agentic workflows and there are really specific use cases that are just limited by an AI agent not being able to access a web page your computer to be able ble to do certain things and so it is quite frustrating now of course like I said before reliability is going to be one of the main things for these agents I mean if you have an AI agent that is ordering you food what if it orders the wrong thing what if it spends too much money what if it posts something on social media that is completely false and it's not supposed to these models sometimes do hallucinate so I really can't wait to see what this is I've personally signed up for the beta I ideally want to do a video on as soon as possible but if you're actually wondering about how well this work works

### Success Rate [18:28]

we can see that if we look at the success rate for this kind of agent we can see that Runner H is 66. 9% and clae computer use is at 52% So currently it seems like they are state-of-the-art so remember how opening eye said that they were going to release an AI agent in around I think January of next year so it seems like this company has put them in a bit of a predicament because they have another wall to climb in terms of usability so openi is going to have to improve their AI agent in order for them to want to release theirs because they're going to have to make sure that theirs is also the best so it's going to be super interesting to see exactly how that works another thing

### Creative Writing Update [19:09]

that happened this week was the fact that gp40 managed to get this creative writing update this was posted on Twitter by Kyle Shannon and I wasn't truly aware so basically the reason this was so surprising was the fact that like because I don't really use chat gbt for creative writing all the time I'm usually using something like Claude you kind of get caught off guard when it has these kind of updates that are I guess you could say qualitative it's quite hard because it's not like a math benchmark where it got the question wrong before and now right this is the kind of Benchmark where it just is a little bit more creative and there's kind of that Benchmark that you know you could only sort of talk about but you couldn't really demonstrate but this example right here is one that is really good because it shows you all exactly how creative the model is now and I think the reason that open I did this is because one of the things that a lot of people have spoken about is the fact that Claude is just so much more creative than chat gbt and it was one of the things that like even I was speaking to you know some developers and some people and they were just basically saying that look this model is just not creative at all this model cannot just talk the way that Claude does so I don't know how they managed to do this or train this but you can see that you know when you're using writing styles now I would definitely try and Benchmark Claude and chat GPT to see if you're able to get your models back into open AI or if Claude is still Superior so it will be interesting to see what the writing Community is saying but so far it seems pretty interesting so this is definitely something that you might want to test out so now there was also interestingly another robot called proxy it's a collaborate robot built for the environment and it's built for the real world so they made this little ad I'm going to do probably a full video on this but I wanted to Showcase this because I wanted to segue into something directly after this every day we wake up and we move you see them every day although you may not notice them their early morning crws the weekend shifts and the late night teams technology has transformed the world since I was first inspired by robots as a kid but surprisingly simple manual tasks still takes so much of our time and energy we're designing these robots to help us not replace Us in every sector proxy moves and swerves naturally in sync with the people working around it and as Ai and Hardware evolve together so will the ways it can handle items in the world I'm here to help it allows us to focus on those things that make us uniquely human communicate collaborate and solve problems a future that's even more alive with the help of robots now the reason I wanted to show you guys this robot quickly was because this robot isn't like your traditional humanoid robot this is the kind of robot that I guess is taking on a different form factor now one of the biggest debates that's been happening in the AI robotics Community was the fact that there are humanoid robots but they apparently useless okay now let me just let me just get to exactly what I'm trying to say here like a lot of people basically say that like okay we're creating humanoid robots they look cool they look amazing but they're just really inefficient in terms of their form factor like creating a robot that is a human form just doesn't make sense like why not give the robot actual wheels and this is where the second clip comes in which is what Nvidia CEO was basically stating about humanoid robots today and he was basically stating that

### Humanoid Robots [23:15]

look humanoid robots are the only type of robots that can scale to extremely high volumes because they can be deployed to the world as is and it's basically same thing that look if we make humanoid robots then we don't actually have to change the world to fit their environment what is missing of course for robotics is an AI that understands the physical world chubbt or large language models today understands the cognitive knowledge and intelligence but it doesn't understand physical int it doesn't understand necessarily that when I set the cup down um that it's not going to go through that table and so we need to teach an AI how to understand physical intelligence well let me demonstrate in fact that we're making good progress well one of the the demonstration that's that all of you have probably seen is you could generate using generative AI text to video and I can surely generate a video that starts out with a picture of myself and you prompt uh Jensen pick up the coffee cup and take a SI well if I can prompt the AI to go and pick up the coffee cup why can't I then generate the tokens right to cause a robotic arm to go pick it up and so the the the Gap the leap from where we are which gener AI in robot General robotics is very close and so I I'm very excited about this area there are three types of robots that can be manufactured in high volume and pretty much only three all of the other types of robots that we've seen historically you know robotics has been around for a long time U but it's very hard to scale into high volume scaling into high volume is important because you need the technology flywheel the high volume allows you to generate High R& D which allows you to make great technology breakthroughs which makes better products which causes the volume to be even higher and that flywheel that R& D flywheel is vital to any industry there are only three robots you could really do this for but two of them are going to be the highest volume and the reason for that is these three robots can all be deployed into the world as it is today we call it Brownfield deployment well the three are cars because we created the world for cars over the course of the last 20050 years and the second uh drones because you know the sky is fairly unlimited but the highest volume one of course is human or robots and it's because we created the world for ourselves and with these three type of robots we can pretty much scale robotics into extremely high volume and that's one of the advantages that a manufacturing ecosystem like this one you know really has and the

### Brett Adcock [25:48]

actual founder CEO of figure Brett Adcock says here yes and many people are still missing the point here humanoid robots are brown fueld solution you need no change to our world for a humanoid to do useful work and many will ask is human the ideal form this is asking the wrong question humans are an odd form but that's not the point is that we built an entire world to conform to our form factor and humanoids can scale to tens of billions of units because of this and it's just hard engineering from here and of course I actually added the CEO there because they actually recently had an update so figure 2 is now an autonomous Fleet with 400 % faster and seven times higher successfully and they're committed to delivering millions of useful robots to companies and homes worldwide and their autonomous Fleet operations is really reliable now you can see that they're running a th placements a day in the BMW Factory and they are running all day and all night and this is only improving because of course as they are doing this they are currently able to collect more data and they're going to be able to improve their models overall so it's really interesting to see how robotics is developing especially since when we look at this company like figure I remember literally looking at the press release for the time that they were just raising funding for this company and now they have a robot not only like 18 months later that is in the factory they're on their second iteration they've got the third one under works and this one is able to do a lot of reliable stuff in BMW Factory so if this is the kind of progress we're seeing in 18 months it kind of makes me think okay what kind of progress are we going to see with these humanoid robots in the next 10 years that's going to really change the way the world is because over the last 6 to 12 months I think humanoid robots has taken a really huge step and definitely in the right direction and of course demos sais in a recent interview

### Demosais [27:42]

actually spoke about how we are on the brink of a new golden era of Discovery pushed by science and exponential advancements in AI I think we are on the cusp of that I really do feel like we're on the brink of a new golden era of Discovery like the whole uh of today's you know Symposium is called um and I think what we need is a lot more interdisciplinary science so using AI bringing in the right way asking the right questions with domain experts um and I think this it's almost Limitless what its applications could be and of course AI itself is a scientific discipline is improving all the time so there's applying today's Technologies um directly to the other fields um and then there's also continuing to improve um AI itself and that as well is a sort of exponential Improvement so you know there's a lot of um I think uh progress to be made in just the next few years

### Nvidia Edify [28:38]

then of course we had Nvidia edify which is a multimodel architecture for developing visual generative AI models for image 3D 360 hdri and physically based rendering which is PBR materials and this uses the Nvidia a foundery and service providers can train and customize edified models to build commercially viable services built on nvidia's Nim so this is something that allows you to create 3D models in a really effective way simply by using natural language so for those of you that are thinking about getting into certain Fields but you think ah maybe it's a little bit too difficult for me this is the kind of thing that is definitely bringing down the skill Gap in terms of what your creativity allows you to do didn't really want to leave

### AI Wages [29:19]

the video on a sort of a you know bad note here but um this is the last story of the video but I want to talk to you guys about this because this is the chart that is recently going viral again and this was something that I've already um covered in a video I had an entire series on my channel dedicated to post AI economics I'm no longer focused on that but since this graph has once again become a you know a talking point in the our community um I just want to explain what this actually means so this is basically a graph from the IMF uh they did a report in 2023 and they basically spoke about how AGI is going to change the future of wages and basically what they're stating is that we have you know two scenarios here we have Baseline AGI which is you know 20 years and then we've got aggressive AGI which is in 5 years from now which is basically 2029 which is the latest date that most people working at these AI labs are predicting so I think what we can see here is that okay in an aggressive AGI scenario and in the base uh AGI scenario the productivity of the average human goes up which makes sense because if you have AGI it's able to do more the productivity goes up and then of course we have this second graph which is of course the wages um and this is where things start to get a little bit concerning because you can see that aggressive AGI right here is a situation where yes your wages go up first but then at the peak just before AGI uh you know is really you know realized you can see that your wages start to plummet down to where they were originally and this thing also happens with Baseline AGI so basically what this actually means is that when your wages go up because of course you know um your wages traditionally go up you can see the blue line is just what happens with no AI your wages are traditionally going to go out with inflation I mean you know let's not get into politics here but that's what's supposed to happen um and then what we do have is with aggressive AGI your wages go up normally and then they plummet down because you know these companies are going to basically be using AGI to do basically everything and then what we have is a price War for goods and services that basically brings the wages down for the average person to uh you know a really interesting rates because I don't know how the world is going to work at this point like I really if AGI is going to be doing a lot of the task asks so it will be interesting I mean one Saving Grace for those of you who don't really like AI is the fact that hallucinations is still going to be a problem so it seems like if you're a specialist in your field you're going to still be largely valued but um it's going to be really interesting to see how wages and how this entire economy comes because I've seen tons and tons of statements and it's not hype it's not fake it is a real thing that these policy makers are going to have to think about so it's a graph there that I wanted to share with you all and just explain to you exactly what it means
