# Big AI News : OpenAI Teases o2, New Robots Advance, New AI Image Generator Crushes, Claudes New

## Метаданные

- **Канал:** TheAIGRID
- **YouTube:** https://www.youtube.com/watch?v=8eo4NIwVYW8
- **Дата:** 03.11.2024
- **Длительность:** 26:18
- **Просмотры:** 28,864

## Описание

Prepare for AGI with me - https://www.skool.com/postagiprepardness 
🐤 Follow Me on Twitter https://twitter.com/TheAiGrid
🌐 Checkout My website - https://theaigrid.com/


Links From Todays Video:
News done
https://x.com/tsarnick/status/1852093013604077965 
https://x.com/runwayml/status/1852363195035406442 
https://x.com/tsarnick/status/1852809421950529720/video/1 
https://x.com/ArtificialAnlys/status/1851707166744584335 
https://www.reddit.com/r/ChatGPT/comments/1ggixzy/ama_with_openais_sam_altman_kevin_weil_srinivas/ 
https://x.com/fchollet/status/1852057682695655634 
https://x.com/alexalbert__/status/1852393994892042561 
https://x.com/fofrAI/status/1852041044386550010 
https://x.com/TheHumanoidHub/status/1851868369483632740 
https://x.com/UnitreeRobotics/status/1852246146049282531 
https://x.com/multimodalart/status/1852793140605854008/photo/1 


Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos.

Was there anything i missed?

(For Business Enquiries)  contact@theaigrid.com

#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience

## Содержание

### [0:00](https://www.youtube.com/watch?v=8eo4NIwVYW8) Segment 1 (00:00 - 05:00)

so did you know this week that unry published this video learning in two days humanlike natural walking AI driven robots need to learn new actions as fast as possible this is very important so essentially what they've done here in this short video is they've trained their unitary G1 humanoid agent to be able to walk with a humanik gate now you can see right here that this is impressive considering the kinds of abilities that the unitary G1 was able to do before this is a stunning presentation of the acceleration that is possible when you do have a team of dedicated Engineers that are able to constantly update the robot and see what is possible now with this I think it's pretty incredible because I do remember that with this first area right here this humanik walking this was something that was really tough for many companies to do and of course some people are saying that this isn't energy efficient it's not the most flexible yada yada but I do think what this does show us is just how quickly things are moving before if we would have seen this human like gate I'm sure that we most definitely would have been surprised because this is you know right now only the second robot that I have personally seen across the internet that is able to walk in such a manner that looks remarkably human I remember the video I did recently on engine AI it was so viral that many people were speculating whether or not this thing was even CGI or not but from this other company here you can see that unitri um you know humanoid robot platform is able to really manage to get this down and I think this also shows us what is going to happen in the future because when we see other robots do other things we're going to see if those other robots from other companies are going to be able to apply those new capabilities and of course later on in the video we do get the surprising nature of these robot dogs and with the amount of reinforcement learning that these dogs do have in simulated virtual environments when they do come back to the physical environment they're able to do a variety of different tricks that seem wonderful and amazing the robot is able to stand you're able to kick the robot while it's on two small legs and it's able to completely balance itself and immediately get back up I think this is something that shows us that once these robots have you know a billion a million trillion hours in the virtual world I wonder how they will act in the physical world now the unitary robot wasn't the only robot showing us some new updates we had the figure robot right here that was actually showing us some very interesting stability you can also see that this robot was able to show us some Dynamic stability basically this is a really important feature for robots because essentially what robots are is they are currently delicate robots and I mean that by like this just huge delicate machines that thing is probably worth $250,000 it's an engineering feat but if that thing can't take a shove and if it can't take a push then it becomes you know I don't want to say completely useless but a lot less useful considering the fact that this kind of humanoid robot it's supposed to operate in human areas and human environments if you're you know a human likely you've been in a crowded environment before maybe someone shoved past you unintentionally maybe you've been on a busy Subway or a busy tube if you're in the UK either way being able to handle such things where you're able to handle a push or a shove and do it you know really well in this form to be able to stabilize yourself almost immediately that is something that ideally we want to see from humanoid robots as quickly as possible because this means these platforms are stable they are rugged and they're built for the real world and the crazy thing about this as well is that like remember how I said that this thing probably costs quar of a million dollars the ring that you guys need to understand is that when you see those demos where they have like a hook hooked up to the back of the robot that is usually because number one they don't want to drop the robot and you know lose a bunch of money in terms of you know the funding that they have because robots are really expensive but of course number two robots usually just aren't very stable sometimes in fact most times they have problems with even getting these things to actually work when it comes to even walking so they kind of need to hold it up so this goes to show just how dynamically stable this is the fact that you can walk past this robot brush it and is able to stand up with no way of falling down did you also know that the Google deep learning researcher one of the most respected ones actually spoke about AGI now this is Francis cholle he actually has his own AGI test which is the arc AGI reasoning Benchmark and this is basically you know some would regard this as the gold standard for AGI because it doesn't involve you know simple LM reasoning it's like a real test that like if you get above 80% they're basically saying that look you have an AGI level system so that test

### [5:00](https://www.youtube.com/watch?v=8eo4NIwVYW8&t=300s) Segment 2 (05:00 - 10:00)

you know if that Benchmark it's broken that's going to be like an AGI day now he said something really fascinating because he also said that the first AGI system will be much worse than the previous a systems at pretty much everything except for tests that focus on measuring general intelligence for example in his Arc test and he says like a baby it will be generally intelligent be able to learn efficiently and adapt but will lack any specialized competence and he says however it will be able to use learn using much less data and the capabilities it picks up will generalize much more strongly to the variability and unpredictability of the real world and as a result after a few years it will beat previous AI systems at absolutely everything including many problems that we thought were completely out of reach before so this is where he's giving his insight into what he expects for the future on AGI and this kind of information is important because a lot of people have this idea that we're going to train these models we're going to use that data and you know everything is going to work out but when we do get AGI system you have to think about how humans are humans I mean some people would argue that humans have a lot of training data like Yan Lun but when we actually look at how many hours of training that it actually takes for humans to learn driving humans can learn driving on 10 hours 15 hours but you know a neonet you know like whatever Tesla is using currently it takes them you know petabytes terabytes huge amounts of data for them to be able to get to where they are and of course apparently they're still not completely everywhere full self-driving just doesn't completely work so the point is that you know when we actually have human level AI it's going to be able to do I'm not saying 10 to 20 hours of training but imagine an AI system that you know only needs 100 hours of data 100 hours of real world data and it's able to go off and actually work really effectively like imagine what happens when you give that thing a million hours of data it's going to be crazy and what he also said was that the arrival of the first AGI will not be the you know most shattering Earth event that people expect it will go entirely unnoticed by the general public now this is a question to you guys what do you think about this because I personally agree to an extent because when we look at how AI is advancing most people don't know about Claude most people do not know about perplexity if you're watching this Channel and you do know what those tools are trust me I can assure you that you are in a select view because you are essentially in a bubble of you know people that watch this kind of content and people that are actually focused on the air development if you go into the real world I've asked my friends tons of times you know have you used perplexity have you used Claude most of them have never heard of those tools and they aren't even bothered to try them so it's something that probably will go unnoticed by the general public despite the significance of the event but it is something that you know Fran solay this person says that it's going to permeate throughout the in maybe 5 to 10 years so he says it took 150 years between the initial steam engine prototype and widespread economic disruption but AGI will take 5 to 10 years for the same so it's a much shorter timeline but I do think that the future is going to be super interesting because every month this Benchmark gets you know 2% 3% more so it will be to see how new models do cope with that now one of the most insane pieces of information that I just really wanted to talk about was the fact that Sam mman actually says that O2 gets 105% on the GP QA now of course this is you know it's not a leak because this is from samman himself of course this is you know speculation he says you know I do appreciate that I can make myself laugh so hard a nice way to go through life but essentially you know Sam mman has spoken about o 2 pretty frequently recently and that kind of indicates to us that 01 is nearing its completion and O2 is probably nearing the end of its training date or training cycle or however you know they're working with the model which means okay surprisingly again with AI that things are moving much faster than we did think because Sam mman is here stating that you know I heard O2 gets 105% on a pretty challenging Benchmark and that means that things are moving really incredible so I wouldn't be surprised if we do get you know O2 in 2025 and that was you know something that I'm not going to say is impossible but this is something that I would not have expected to happen this early in the day and for those of you who are thinking okay 105% on GP QA I'm going to give you guys a bit more context with actually you know what that means so PhD level science questions are basically the GP QA and you can see on the right hand side exactly how much 01 improves over GPT 40 we can see that this improves significantly on you know all things so look at GPT 40 versus if we look at 01 preview then of course you can see that there's you know a stockk Improvement across the board and of course you can see in the PHD level questions which is GP QA we can see GPT 40 then of course 01 preview and of course o01 so it looks like you know AI is pretty much surpassing the expert humans at this level which is pretty ins when you actually think about it and if

### [10:00](https://www.youtube.com/watch?v=8eo4NIwVYW8&t=600s) Segment 3 (10:00 - 15:00)

Sam Alman is saying that look the gpq is just going to be you know all the way up here which basically complete the benchmarks that's going to be basically meaning that we're going to need new benchmarks after these inference time reasoning models are here which means that the kind of intelligence that we're going to get next year along with agents is going to be I think probably I don't want to say the first mind-blowing moment but it might be the moment where people start to realize okay we're no longer the smartest things on the planet anymore and these things have a ridiculous amount of intelligence now of course it could just be Sam mman hyping up his own company as of course open AI always know a way to continue creating hype around their products but I do believe that since we've entered a new paradigm and since that we've got such a major jump in the first iteration of the model and the fact that they're saying look there's many other things that we can try I still think that we have a long way to go in terms of what to happen now what's crazy about all of this is that we actually did get a ma with opening eyes Sam mman Kevin will s vas and Mark Chen so this is something where they actually revealed a lot of information now one of the things they said okay and I've done a video on this I should have released by now but one of the things they actually said was they said okay well they didn't say okay Sam Alman said okay and if I can find this tweet not a tweet is actually a comment but um basically he said something about saturating all the benchmarks now I'm going to show you guys this because it's insane so you can see here he's basically saying that look a bold prediction for 2025 is that they're going to saturate all the benchmarks which basically means that all of those benchmarks the people see you know like this one's got 80% this one's got 70% all of those benchmarks are going to be completely crushed by opening eyes so this is something that I think is you know incredible when we start to think about how far we've come and the marginal improvements that we do get to see on these benchmarks and I think if we're seeing the jump for what it is from 01 to you know GPT 40 I mean I think it's going to be something that's incredible because most people haven't realized just how insane these models are we don't have use for them on a day-to-day basis which is why even when AGI gets here most of us probably won't even realize of course those of us in the AI space will pay attention but the prediction for 2025 and remember a lot happens in a year like if you think about you know where we are now compared to where we were a year ago in terms of reasoning in terms of like you know inference in terms of the kind of models that we had in terms of the coding and like you know the entire you know company space that we had things have moved very quickly now if you want to see just how quickly opening ey have been moving you can see this graph here it basically shows you how things have been accelerating we can see that for the first time it was gpt1 gpt2 gpt3 which was you know pretty crazy like you know we got one a year then of course we just went boom we had chat gbt then you know like couple months after that we go gbt 4 and now guys you can see that like every two 3 months we're getting new products we're getting new things we're getting new updates and you know open AI are racing to stay ahead of the competition across all fronts and areas so this is pretty crazy so this is from a private opening ey webinar but I want you to take a look because you can see okay how much is going on like it's just it's crazy it's genuinely crazy just how quickly things are evolving and that the pace is actually accelerating you know if you think about open AI going back to you know 2018 when we launched GT1 it was almost every year we would come out with a new frontier model and this gave people and companies a lot of time at this point you know was just kind of early developers a chance to react and think how these models would perform in the products they building you know fast forward we launched chb2 in uh chbt in November 2022 uh so almost two years ago but when you actually kind of fast forward really into that 2023 into 2024 you know things started accelerating faster uh GB Enterprises hit one uh recently at the end of 2023 but then we started launching the gpts and we launched a GTP store at 2024 we came out with teams another business product 40 which is our multi modal uh gbt 40 came out in May 01 just came out our first model capable of reasoning a couple months ago we just launched canvas a c a couple weeks ago searches came out today I included a lot of major products and in Frontier models but there's you know hundreds of other product updates that have happened so yeah you guys can see here that the pace isn't slowing down in fact I would argue that it's speeding up and I really do want to see what next year has in store for us because we have ai agents they're going to be absolutely incredible of course like I said before samman has been talking a lot about O2 we can see right here that he says um you know when someone said okay and this is the craziest part this is the next story okay um they actually released accidentally 01 okay and I got to use it I'm going to show you guys how crazy it is so um someone said okay unleash the full 01 currently that's referring to the fact that we don't yet have the uh 01 that we need currently we do have 01 preview which is like a you know a distilled version of the strawberry model and of course okay he's

### [15:00](https://www.youtube.com/watch?v=8eo4NIwVYW8&t=900s) Segment 4 (15:00 - 20:00)

stating that look okay it's not going to be that much longer until we get 01 okay but how long until the cuse are going to be unleash O2 so we've seen two references on O2 just today like I said before could be AI hype but I don't think it is completely given the fact that you know a large majority of this Paradigm is inference you know time where you know you're not you know training the model for as long but you're just you know put it spending that inference on actually when it asks the question it spends that time reasoning throughout the solution so it seems that maybe O2 is already here and if he's already saying that look our goal in 2025 is to saturate the benchmarks it's quite clear that they have a path that they you know are feeling really confident in and it seems like you know 01 is just the beginning of that path now like I now another company managed to release something that I personally believe is really effective so basically Claude released the PDF capabilities in you know their app and this is game changer because I was always a little bit skeptical of Claude using PDFs because if you've ever used PDFs with Claude one of the major problems that it actually did was it didn't actually look at the images what it did was it actually so what it actually does okay and what it actually would do was it would extract all the text from the PDF and it wouldn't go systematically and actually read that PDF and as someone that would want to work with long context documents this was something that's just remarkably frustrating so it wasn't able to actually analyze the visual elements in you know association with the actual text which meant that the reasoning was quite unfortunate and this was a small thing because when you could you know input your PDF on the right hand side you would see that it would say extracted text yada y yada and then you would be able to see the text it extracted and often times I'd be like this text is formatted weirdly this doesn't look really right I can't believe this is doing this but now if you want upgraded PDF features if you've been using PDFs and you've been getting some subpar responses Claude now has this feature where it examines everything so I'm thinking that you know in fact I don't want to say I know how to do it because I don't okay maybe they just putting it through the image model but either way this is a good feature um and I'm definitely going to now for those of you that watched my previous video okay remember that how we spoke about there was this recraft 33 model that was crushing benchmarks and we had no idea with as to you know what on Earth this model is now I'm surprised that uh a company from a London based AI graphic design startup managed to just create a model that completely demolishes everything okay and it demolishes flux 1. 1 Pro IDR V2 mid journey and even stability AI so I don't know how they managed to get their hands on this model I don't know how much training they've done but whatever technique they've used whatever they have done it is you know working because they've managed to do something that is crazy they've taken on these big companies and they have one the arena weight is incredible and when I show you the images that this is able to do you're going to see that this is absolutely insane so shout out to F AI for these images you can see on the left we have 1. 1 Pro which is flux which is a really good model not to say the flux is bad but just take a look at what happens when you use recraft now the prompt here is a photo of a wet cat but if we go to the right you can see that looks a lot more like a realistic photo of a wet cat so this one here we've got a cat it looks amazing looks HD it's in the rain but it's not exactly a wet cat that right there is exactly what a wet cat would look like especially after you know cats are mainly running away from you and not trying to have their Buffs and you can see here that the quality for these models is absolutely incredible so you can see that this kind of quality looks incredible it looks insane and this is something that looks really nice now there's also some more examples simple language model tasks this page contains the number of and the model is able to say that the word strawberry has three Rs which is absolutely insane then of course someone says that the next you know image prompt is that this page contains the result of 2 + 5 and you can see that it says 2 + 5 equal 7 that just looks really super realistic there like if someone sent me this image I don't know how would tell that that's completely AI only in fact the pen I don't even know I've maybe I've never seen that pen before but I'm not like a pen expert so I can't really you know pinpoint that then it says write to objective in English and of course you can see it has two right there and then what is the name of the US president and you can see Joe Biden so this is something that other models just simply cannot do and I think their success boils down to the fact that like maybe considering and this is just something that's just came to my head now but considering the fact that the company is uh from a graphic design based startup considering the fact that they understand the design elements I think that gives them the edge when looking at how to select certain images and of course you know the kinds of design choices and realism

### [20:00](https://www.youtube.com/watch?v=8eo4NIwVYW8&t=1200s) Segment 5 (20:00 - 25:00)

that they do make I think a lot of companies you know they have a lot of R from people that are in certain you know Industries like they're just focusing on getting people that can really train these models really well but it's good to have a designer here because you know when you've got a graphic design firm these are going to be the people that are you know the end users of some of these products so it makes them really integral to you know how the thing is developed and we can see here like this is pretty crazy I'm going to show you guys one more example and this one should really show you guys where the model is excelling so the prompt is squid in kitchen we can see that there's a squid it's on the kitchen it looks super realistic as it does and then of course on the right hand side you can see we've got a squid in the kitchen by flux 1. 1 Pro so this is a model I can't wait to test on new prompts to see how realistic it is where it excels where it fails if it passes my own personal benchmarks and if I can use this for certain tasks now for some more creative news we did also have Runway get Advanced Camera control being now available for Gen 3 Alpha turbo so this is something that I think is going to change the game because this allows an incredible level of user creativity now you can put the camera to the right the camera you know up down up you know left right up down flip it all around and you can get multiple different angles of something that is yours and why this is incredible is because a lot of the times with Runway we can't really control well we used to not be able to control where that camera went so having this feature is going to you know open up uh next level you know for creators that really want to work with this kind of technology so this is something that I think you know changes the game you can see that it results in a lot more Dynamic scenery uh and these shots that you're able to do are just a lot better in terms of what we're able to get because the model is able to be like okay I want to move left right it's not just completely random for example this you know this image right here of the sushi you can see exactly how this works and exactly you know how crazy it is in terms of the kinds of realism that we will get so I mean the kinds of pipelines that we're going to get when it comes to AI video I mean you know a lot of people think it's just text of video but Runway I think what they're doing and I think this is the smartest thing they are doing is that they're developing the kinds of tools that allow people to really direct their own short films their own movies and I think that is you know really important because it's not just hey this is you know AI text a video it's hey you can move the camera left right you can have full creative control and freedom over the kinds of shots that you do want to have and this is where someone is zooming out completely to reveal a you know an abandoned city or you know a city that's just been through ruins and Chaos so I mean this kind of thing I mean it's really gamechanging like we're about to I wouldn't be surprised if by next year we get our like full you know fledged first AI movie or we get like a full you know YouTube video like YouTube channel that is just like full of AI you know shorts or whatever of people just using AI to create really cool and interesting content so I will really love to see exactly what that looks like in the future now my star for suan actually said that memory and vision are coming to AI where it can remember everything about what you learn and see and I'm going to include this clip because after I show you guys another clip you're going to see that whoa okay and from the comments that I've seen a lot of people are requesting this feature but I'm going to show you guys uh that thing in a moment the moment because it's frustrating that it forgets that you told it that you were going out to dinner last night and that you know you had an undergrad at XYZ I mean so the bad assistant the bad yeah it's kind of that's boring right and so you don't bother really giving it your all just think about the moment when like it really does remember everything that you've said and it's able to retrieve from that history of your conversations accurately and the timeline for that I think that we're 18 months away from really getting that to a workable level if you look out far enough yes it's quite likely that your AI is going to learn to click on buttons in your browser um enter text in text entry boxes it's certainly going to be able to read the page and understand the images and all of the text on the page that's going to be a very different you know format I think one of the things that's going to come out quite soon I would say is this idea of co-pilot's vision it's going to really see what you see like and so you'll be able to have a companion that really browses with looking at the IM see what you see physically or see what you see as how you use I think at first digitally right but then very shortly after that there will be a visual awareness whether it's through earbuds or glasses yeah airbuds with video in them that's my yeah I think that's going to be a new format is these visual now if you're familiar with who Mustafa slan is he's currently the head of inflection AI well in fact that's not true he used to be the head of inflection AI he's currently the head of Microsoft AI okay Microsoft are doing some pretty cool things with their product called co-pilot so co-pilot is basically going to be what advanced voice mode should have been so what we're going to have here is an AI that can see your browser in real time so as I'm making this video or as I'm making other videos I can just message co-pilot it's going to be able to see exactly what I'm doing and it's I'm going to be like hey what do you think about this video

### [25:00](https://www.youtube.com/watch?v=8eo4NIwVYW8&t=1500s) Segment 6 (25:00 - 26:00)

what do you think about you know the script do you think this is good do you think that's well it's going to be like an entire AI assistant that lives in your computer that you can talk to while you're browsing I mean I think it's going to change like quite of a lot of things cuz I mean you know it's actually pretty useful when you're working just to have like an ear um that can also see what you're doing because one of the things I realized is that now that advanced voice mode is on desktop um it's really cool to like you know talk back and forth between ideas you know navigate you know some of the ideas that you're doing but you know once you have vision and you're able to show it certain things it's going to be able to be like okay this is really cool this is really nice I like this I like that and I think that just completely changes the game because I won't have to completely copy and paste in everything in it's going to see what's on my screen by it's going to be by my side so it's pretty crazy so um I mean I can't wait for this genuinely and if you guys enjoyed this uh video as well uh don't forget to subscribe to my newsletter um I'm going to be trying to cuz since there's so many stories I've been uploading so much I think it's best if every single week I summarize them it's going to be completely free um no subscription required cancel at any time zero payment it's just AI news that you most likely did Miss and there's still stories that I didn't include in today's video simply because I just don't have like this video would be an hour long if I covered every story within the last 3 days so if you guys want that let me know in the comment section below and I'll see you guys in the next video

---
*Источник: https://ekstraktznaniy.ru/video/13832*