# AI NEWS : OpenAI Drops "Blueberry Model?" Metas Stunnign New AI Voice, Sora 2  and more

## Метаданные

- **Канал:** TheAIGRID
- **YouTube:** https://www.youtube.com/watch?v=zZ_lZVVOMaw
- **Дата:** 28.09.2024
- **Длительность:** 20:11
- **Просмотры:** 36,258

## Описание

Prepare for AGI with me - https://www.skool.com/postagiprepardness 
🐤 Follow Me on Twitter https://twitter.com/TheAiGrid
🌐 Checkout My website - https://theaigrid.com/


Links From Todays Video:


Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos.

Was there anything i missed?

(For Business Enquiries)  contact@theaigrid.com

#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience

## Содержание

### [0:00](https://www.youtube.com/watch?v=zZ_lZVVOMaw) Segment 1 (00:00 - 05:00)

so there have been many stories this week in Ai and I'm sure because there has been so many different stories you may have missed a few things so let's dive into today's stories which are going to give you an update to what you all missed so one of the stories that I actually nearly missed was this story about open a revamping Sora AI video this is an article basically talking about the future development of the app and I think it's rather fascinating because we as individuals have haven't actually gotten access to Sora yet whilst the competition from other models like clling 1. 5 Runway lumabs have intensified so you can see here that this is the information a reliable source for AI news leaks especially and it says between the long awaited strawberry launch and its monster ongoing fundraising it might be easy to forget about Sora a video generating model it teased in February to Global Fanfare a number investors told me at the time that the impressive Sora demos cast a worrying Shadow over the gaggle of competing AI video startups that had already raised hundreds of millions in Venture Capital these concerns may have eased 6 months later open AI still hasn't launch Sora but they've said that could soon change open AI is training a new version of Sora so I'm guessing that this is going to be Sora too that in hopes it will quickly generate clips that are higher in quality longer than the ones it demonstrated earlier in the year according to a person who spoken to open AI Executives about it so what's crazy about this is that we didn't get access to Sora and I think there were a variety of different reasons why this happened number one Sora wasn't particularly received well by the global Community many most people I would say like on that were you know vocal about Sora actually rejected the release I spoke to many creatives I spoke to some individuals um and they all pretty much just didn't want this tool so if those individuals would I guess you could say reject the tool and considering the fact that open was going through a period where they are pretty much struggling for compute the compute was so bad that they didn't even give any to their super alignment team which they pledged that they were going to and the team ended up disbanding and by compute I basically mean gpus for inference for many different things and anytime you use services like chat GPT to generate audio to generate images you know it uses their inference it uses their compute which is uh pretty taxing considering the fact that they are trying to grow their company and also train new models at the same time so it seems that what open ey I think they're trying to do here is like once again leap fog the competition because when Sora was released it was unheard of for AI video to be that good but I think they're trying to get to the next stage now where they just completely demolish the competition in terms of what is possible so you can see here that they are speaking about training Sora which is where the company collects millions of hours of video for training data and they said that these videos need to be high resolution and contain a diverse area of styles and subjects and it says but why has open AI taken so long to make any progress on Sora and what were the issues that prevented it being released as a product in the first place of course one of the things I did forget was that you know the Sora announcement one of the things that was an issue was that it did actually take really long to generate Clips it took around you know more than 10 minutes to generate a short clip of a minute or so and the crazy thing about this is that if you've actually seen Runway gen 3 it is so quick in fact I could probably generate an image right now in fact I'm going to show you guys how quick it is because it's genuinely outstanding so if you literally go to Runway okay so let me just say uh nighttime City cyberp Punk okay and I just want to show you guys how quick this is in real time and the reason I'm showing you guys this is because this is going to show you guys what opening eyes is competing with so if I put a nighttime City cyberpunk camera pans around cinematic drone shot um and then I'm just going to put uh do this so let's do generate okay why is do I need to add an image okay for example I'm going to go ahead and use this image and I just want to show you guys how quick this is so this is a 10-second one so what I'm going to do is I'm going to set this at the first frame and I'm going to click generate so you're literally going to see in real time how quickly this generates so this is of course their quicker version but I'm not going to skip the video at all it's now in Q I think it's only been around 10 seconds but guys look at this look at how quick this video is being generated this is why I'm saying that like of course you know you could argue that Sora has higher quality yada yada um and you know of course opening ey is is more known but you have to understand that this was you know 10 20

### [5:00](https://www.youtube.com/watch?v=zZ_lZVVOMaw&t=300s) Segment 2 (05:00 - 10:00)

seconds okay from start to finish okay maybe 30 seconds now that it hasn't been done yet okay so this was 30 seconds and I'm not sure about the quality but let's just take a look so this was 20 to 30 seconds but the crazy thing about this is that this was incredibly fast and of course with better prompting this video could have been made better but the thing is Sora took 10 minutes okay and that's not I would don't want to say it's not feasible because you know 10 minutes is a lot quicker than going out and filming all that stuff but if you're trying to you know offer a project you can't really just be offering it at you know 10 minutes to generate a short clip of about a minute or so I don't think that that's going to be something that's remarkably effective um so you can see right here for instance one of the filmmakers Patrick cedarberg said he had to ask the model to generate hundreds of Clips before landing a usable ones in other words Sor was uous to use basically saying that look whilst it's fun I had to generate so many uh different times and this is actually quite true you know when you're experimenting with certain AI softwares it sounds good in theory but a lot of times like even when I'm messing around with AI voices you know you have to generate them like a few times in order to get like a good output that you can genuinely use which is uh kind of frustrating so I don't know if they're going to like have something that generates 100 clips and then the verifier like goes and checks for the best one that would even take up more inference and compute but um yeah you this kind of thing has to be useful and so I think this part of the video was rather important because of course you know future the improved Sora model is going to be fixing many of these issues Sora is one of the biggest things that I thought I think you know brought a AI into the you know immediate mainstream because it was just a tool that everybody immediately knew what it was it wasn't just like text it was like okay text the video this is insane so I think that this is pretty crazy but I think Sora is like a backend product I think it's going to be you know for Movie Studio and stuff but like I said before other companies have caught up Runway um maybe not P just yet but clling AI are striking deal with these you know major companies to get stuff that they can actually use on a day-to-day basis and with that kind of speed that you just saw from Runway it's going to be really difficult for open AI to compete but um in some other AI news there was also the meta connect the meta connect I got to be honest this was like really incredible I just was blown away by what meta was has done because llama 3. 2 was overshadowed by a few other things going on at open AI such as the advanced voice mode and a few other you know notable AI developments such as the drama going on an opening eye but llama 3. 2 was a great release because this 70b model actually now includes Vision in it and it's actually good vision and it's an open- Source Vision which means that we're going to have developers that also do some crazy things so of course we did have you know 1B and 3B models that were like on device that were really effective and I think this just goes to show like what is being done in the future in terms of what we're going to be able to get with future AI systems that are going to be really effective I can't wait till certain phones have these models you know just natively on device when you're able to use them for certain things just making your phone a lot more effective like you know doing certain agentic capabilities as we're seeing in this demo here I think this is certainly where the future of AI is provided that these developers manage to take advantage of them and you know we create an ecosystem where people can build things natively on these phones unfortunately what meta do have and I say unfortunately because it's only unfortunate because I live in a location which is highly regulated this kind of thing isn't available to me yet so there is this AI voice and it was actually was demoed recently and I'm going to show you that in a moment but this is kind of like their version of chat gpts of course it's not as great but it's still remarkably effective considering the user base meta have with their distribution you have to understand that these guys have like I think threads has 100 million customers already in terms of you know users that have signed up to the platform Instagram and WhatsApp already have over a billion people so with that kind of distribution that's literally like 17th of the world's population or 1/ 18 which is insane when is the next total solar eclipse is on August 12 2026 and will be visible from Greenland Iceland Spain Russia and a small area of Portugal is we are experiment experimenting with automatic video dubbing on reals uh starting with English in Spanish um so this is going to be a way that you can see more of the content that's out there no matter what language you speak and as a Creator this is going to let you reach a lot more people around the world no matter what language you speak so now you're going to be able to take your content and it's going to be in your authentic voice speaking other languages if you want and we're also doing automatic lip syncing um so that way it actually looks like you were speaking the other language so let's check this one out

### [10:00](https://www.youtube.com/watch?v=zZ_lZVVOMaw&t=600s) Segment 3 (10:00 - 15:00)

with coconut banana and this recipe was one of the first ones I uploaded to social media a little while ago I rediscovered it and made a couple of modific now I think this right here is largely going to be one of the most underrated things in AI because what we have is like I wouldn't say it's a unification of many different countries although some people would see it as that but I think it's going to be really interesting to see how different cultures connect now that we have ai because what's going to happen like over the next 5 to 10 years is that AI will be able to translate content easily there was even a speech I think it was at some kind of event it was I genuinely don't remember the event but I remember that the translated version got so much more support than it than the original spoken one because it got translated to English and a lot more people could understand exactly what was going on and it didn't just sound like you know voiceover it sounded like that person speaking and I think we're going to be seeing like for those of you guys who like me and you speak English you're going to be able to see a lot more different concentrators from smaller countri is popping up you know going completely viral because of this area so if you're someone who you know maybe English isn't your first language and you speak a different language I would say taking advantage of translating your content to other languages um and that broader reach is going to really benefit you and definitely for businesses definitely for people who are trying to you know explore how they can increase their revenue Beyond just their current uh Nation this is going to be something that I think is really effective and really underrated because having that translation I mean you know there's so much great content out there that you know you could consume and you could enjoy that this kind of you know technology enables us to do so I think it's going to be you know incredible literally incredible but it's very underrated at this current moment in time and YouTube is going to be doing a similar thing as well so maybe you'll hear me speak Spanish or French um if that is your native language now take a look at some of more methods announcements I was even supposed to put these in the original video but it I was just ill at the time so I wasn't able to do as much as I could so do apologize for that but take a look at some more of these announcements because meta like I think they're probably going to be one of the key industry players that is taking a completely different approach to many of these other companies like while these other companies are focusing on the foundation models I think meta is just focusing on you know user distribution which I think is probably going to be more important in the future um so now meta AI can natively understand images as well as text now we're using this new capability to build some features that I haven't seen anyone else build I mean this is pretty interesting and stuff creative tools that we're adding to imagine edit it let you upload any photo and edit it precisely um with natural language in meta AI across the apps my family loves this we spend a lot of time taking photos and making them more ridiculous you know one of my daughters actually you just wrote a book and we used meta Ai and she used it to illustrate the book it's just a lot of fun and awesome um all right but moving on to another thing this one I think is actually probably the biggest AI news that we have today we are adding natural voice interactions to meta AI um and you don't have to wait a while we are rolling this out today and over the next few days um I think that voice is going to be a way more natural way of interacting with AI than text and probably I think it has the potential to be one of if not the most frequent ways that we all interact with AI um it is just a lot better so now you're going to be able to have natural voice conversations with Med across all the major apps you got you know Instagram WhatsApp messenger Facebook um we want to make this fun all right so in addition to the different system voices that we built in that I think are are pretty good um we've also worked with some of the most iconic voices out there to bring them to meta AI like John Cena um Judy Dench Kristen Bell uh Keegan Michael Key and Aquafina um should we try it out all right hey are live demos risky live demos can be risky yes they can be unpredictable prone to technical issues and potentially embarrassing however they can also be engaging memorable but I'm going to do it anyway go for it just be prepared for anything to happen fingers I mean um live demos can be risky but you can see they're Maric pulled that one off and honestly you just have to feel bad for Google because like every time they've done a live demo it's just failed spectacularly I don't know if the universe is conspiring against them but yeah I think that just goes to show like you know how people are saying that opening I have this massive lead um they are really under so much pressure now to really get stuff through the door and in another video I actually spoke about how

### [15:00](https://www.youtube.com/watch?v=zZ_lZVVOMaw&t=900s) Segment 4 (15:00 - 20:00)

open AI staffers the safety researchers are having to work 20 hours per day during tight deadlines in order to get stuff out the door because of course it's pretty much the last hurdle to pushing these products out there so I mean when we look at meta AI now having their realtime voiceover we had Google have theirs and now literally just last week we finally had opening eye rolls theirs out of course we are in this arms race in terms of AI productivity now with open AI there was really something that I wanted to show you guys because this is some thing that I think is changing in terms of how people view AI systems and the reason I say this is because if you remember the early days of AI and I say early days when really I'm just talking about like two to three years ago but like there was this researcher from Google Blake CL Moy I'm pretty sure most of you guys heard about him but I'm going to give you guys the to Long didn't read version basically this guy Blake CL Moy um he basically was a Google engineer and he thought that the AI had come to life now I'm not going to be someone who says it has or say it hasn't I think these AI are black boxes um there's a lot about humans we don't understand anyways so it's anybody's best guess you know some people are like these things are just chat Bots a number people like okay but like how do you know that I mean we're just simulating certain things in our head um and basically the long story short is that this guy went viral there were so many people saying that this guy's you know a for thinking that these chat Bots are sent but recently opening ey have usur that kind of statement that it may be that these neural networks are slightly conscious and the reason I said this is so fascinating is because um this article was from 2022 when it really went viral this was even before chat gbt which was really you know fascinating but let's take a look at what they're saying about openi 01 now uh 01 models when we trained it and we actually started talking to we started asking it these questions and it was scoring higher on these math tests we were giving it we could look at how it was reasoning um and you could just see that it started to question itself and have really interesting reflection um and that was a moment for me where I was like wow like we we've uncovered something different this is going to be something new and it was just like one of these coming together moments that that was really powerful so when you read the the thoughts do they does it feel like you're watching a human or robot it's like a spiritual experience it's a spiritual experience but then you can empathize with the all like oh that's a mistake that a lot of people would make or you can see it sort of questioning common conventions and yeah it's spiritual but like oddly human in its Behavior so that brings the question do you guys think we're going to come full circle on these models and we're going to have people starting to say that yes these models are completely conscious they have feelings I mean there's this entire philosophical debate and I think I don't really know you know it's one of those things that's kind of confusing because when we start to question what these models are capable of doing we have to always then kind of question ourselves because one of the things I can see time and time again is like oh these models aren't reasoning they're not doing this but it's like you know like so many times these models make mistakes of human make and so many times these models literally improve the same way a human would improve like for example you know if you say show you're working this is exactly how a human will improve like if you give an llm time to plan this is exactly how a human would improve their ability to you know perform any kind of task I mean the whole thing is you know just for like an entire another video I did a video on this a while back I'll leave a link to that but let me know what you guys think if you think these models are conscious or not cuz I think the debate either way both of them makes sense genuinely because if we're saying that they're not kind of makesense and if we're saying that they are kind of makes sense too so with that being said let me know what you think about that now there's also this artificial analysis text to image leaderboard and basically this leaderboard is where we can see the text to image models in terms of their capabilities and recently there has been this Secret model that has come out of absolutely nowhere and this model is called blueberry guys this is rather fascinating now at first glance most people are speculating that this blueberry model could be referenced to open I strawberry but potentially an image generation model that just simply surpasses everything before I personally wouldn't be surprised if this was just a play on opening eye strawberry like a blueberry model and they released something that's really cool because Sora does have incredible image generation capabilities but I think it's surprising that I remember only recently that we got flux 1. 1 Pro that was just so further ahead of chat GPT and pretty much anything else that it was quite surprising ing especially considering the fact that it immediately went open source and then now only you know one or two months later we have something called blueberry that is quite Anonymous I mean it could be Claude it could be Google I mean there are just a variety of different things that we haven't seen yet from these major companies so I'm truly wondering what company this is some people think it's Sora I think it probably might be open a eye considering the fact that they haven't really explored their image generation capabilities and we know they are really

### [20:00](https://www.youtube.com/watch?v=zZ_lZVVOMaw&t=1200s) Segment 5 (20:00 - 20:00)

good so I do think this is most likely going to be open AI but let me know what you think about this super Secret model because I think we're probably in for something big if this one already is a high jump amongst the others

---
*Источник: https://ekstraktznaniy.ru/video/14052*