MASSIVEAI NEWS #14 ChatGPT Comes ALIVE , AI Robots Are HERE, Massive Midjourney UPDATE,
25:49

MASSIVEAI NEWS #14 ChatGPT Comes ALIVE , AI Robots Are HERE, Massive Midjourney UPDATE,

TheAIGRID 02.09.2023 30 658 просмотров 752 лайков

Machine-readable: Markdown · JSON API · Site index

Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
https://twitter.com/LinusEkenstam/status/1694121273964589368 https://twitter.com/iamneubert/status/1693697315163472050 https://twitter.com/Saboo_Shubham_/status/1693836533776716027 https://twitter.com/AISafetyMemes/status/1693644451317514261 https://twitter.com/TheAIAnonGuy/status/1693718664392888820 https://twitter.com/DrJimFan/status/1691492256203845632 https://twitter.com/_philschmid/status/1693989755393102201 https://twitter.com/reach_vb/status/1694020457798815811 https://www.youtube.com/watch?v=XOeFeVAZfl4 https://ideogram.ai/launch https://www.youtube.com/@apptronik8459/videos https://twitter.com/DrJimFan/status/1694358069638275463 https://twitter.com/SanhEstPasMoi/status/1694463568178008106 https://twitter.com/diveshnaidoo/status/1694780929527390699 https://www.reddit.com/r/singularity/comments/1600g1s/apollo_robot_runtime_4h_swappable_battery_max/ https://www.reddit.com/r/singularity/comments/1600uo0/mit_researchers_robotic_manipulation_planning/ https://twitter.com/Eng_Rory_King/status/1694952747223925040 https://twitter.com/AlexReibman/status/1695643709889818969 https://twitter.com/Eng_Rory_King/status/1694952747223925040 https://labs.perplexity.ai/?utm_content=first_codellama&s=u&utm_source=twitter&utm_campaign=labs https://twitter.com/dreamingtulpa/status/1696855178937323567 https://twitter.com/fffiloni/status/1696845630583296177 Welcome to our channel where we bring you the latest breakthroughs in AI. From deep learning to robotics, we cover it all. Our videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on our latest videos. Was there anything we missed? (For Business Enquiries) contact@theaigrid.com #LLM #Largelanguagemodel #chatgpt #AI #ArtificialIntelligence #MachineLearning #DeepLearning #NeuralNetworks #Robotics #DataScience #IntelligentSystems #Automation #TechInnovation

Оглавление (13 сегментов)

Midjourney Update

so there was a ton of new AI tools released and there were actually a ton of new AI robots released so this week is going to be very interesting because the industry is probably about to change for a whole so let's get into all of the newest news on artificial intelligence that many people just as yourself may have missed so something that was actually released from one of the most popular softwares was in painting from mid-journey so as you can see this is quite similar to photoshop's generator fill feature and where you can essentially fit in something into an image that wasn't previously there before this is a complete game changer for those who are trying to edit images with mid-journey because mid journey is largely regarded as the very best AI generated image generator out there and of course now that we have this feature to even customize images even that little bit more adding small details whether it be the tires on the car the lipstick on a woman or adding someone to a scene it's going to be really interesting to see just how far mid Journey can push this now later on in the video we are going to be talking about something we're going fighting mid-journey's dominant position being threatened by another AI startup but that is later on in the video let me know if you think mid-journey's in painting feature is much better than Photoshop generative Phil I do think that this is a great feature that they've added I do think that it is on par with photoshop's generative fill however I do think Photoshop generative fill is much easier to use but at the same time it just goes to show that with mid Journey you are getting quite a lot of detail and now we're adding more features and we are getting a lot more customization now if you do want to learn a lot more about in painting there are tons of different tweets that I'm going to leave in the description where people are showing you the threads on how you can actually use impainting so be sure to check those out because it's actually really interesting and every time a feature is released we do think that we do know everything about that feature but you know Twitter users always are going to show us how we can actually use these features and what kinds of styles different things we can actually do with

Codef

them so then we had something called codef the future of AI video representation where you can change one frame and watch the entire video transform the Tweet says from capturing every static detail to effortlessly tracking water and smog codef ensures seamless changes across the entire footage so essentially what this is if you remember runway's text to image where you could simply put one image in to an already pre-recorded video and then have that video change based on what that input image is so essentially set a style image and then input you know a normal video and then that video would now represent you know a style of that image that is exactly what you're seeing on screen and I think this is very interesting because what we have here is something that shows us that this kind of software is getting better and better runways one was a little good at first but this you know if you if you've compared it if you've seen runways is actually pretty good but this is pretty incredible some of the scenes that I've seen from here are just I mean they don't look real in the sense that I do think that they've added some other kind of technology to this because it looks pretty good like it does look pretty good it looks uh I just you know I'm actually speechless at this point so I'm guessing that maybe they leverage new technology or maybe they used a new method but it is really good because I know now that when it comes to text to video now that this is going to be added into future software and this is going to make for really interesting videos so you can see right here uh no wait where is it stylization yeah this is what I was talking about from Runway where you have stylization and I do think that their stylization is updated and I think what we can expect is in the next coming weeks and probably you know a week or two from now we're going to see Runway update this stylization feature because this new paper has released a bunch of different methods and I'm pretty sure Runway are probably analyzing this and are going to add to it in their next update because Runway honestly does update the stuff but let me know what you think about it I think it is really cool good and I think it's uh pretty insane if you ask me so

Jailbreak

um yeah let's move on so yeah you guys can see right here this is something else that was really cool and I'm gonna actually play the audio so that you guys can hear what's going on but um essentially someone Jill broke their watch and essentially what jailbreak means is to pretty much just disable the firmware that stops you from doing certain things um like maybe adding custom applications that they don't want and they've basically customized their own watch to turn it into a second brain which is an AI companion called Knox and essentially it can capture notes and recall instantly for proactive assistance it's pretty crazy there's a tweet um there's you know a company that has you know infinite memory I'm not sure how it's going to work it seems pretty cool but the video is even cooler um and the voice that they've used is a level lab so I don't know if they've hooked it up to 11 Labs API which essentially means that you get ultra realistic sounding voices in real time so it doesn't sound like you know a standard AI robot that sounds very drony it actually sounds like a real person but let me know what you think and if this is going to be the future of AI devices because I think that it probably will be morning Molly today you've got berries at 8am lunch plans at Cava and you just got invited to Sam's birthday party at seven do you have a present yet nope any ideas hmm let me see you guys met at that Soho House dinner last month where you bonded over being enormous swifties so the most obvious gift would be an heiress tour ticket but he's also been dropping hints about needing a new bike I can place a direct order if you want to save you the hassle don't let's get both also looks like you both follow Carlos alcaraz and Nike run Club on Instagram you could mention the extra US Open finals ticket you have this year I'm sure he would be up for it ah I'm late can you order me an Uber done here in five then there was something that is really cool okay so you can see right here that this tweet says this is uni tree H1 a 90 000 human-sized humanoid robot that weighs 104 pounds the price tag is specifically what is game changing about this robot is because if you've known anything about humanoid robots it's that they're very expensive to produce but this humanoid robot is only ninety thousand dollars now you might be thinking that's hugely expensive for a humanoid robot remember okay when TVs first came out they were hugely expensive when the car first came out it was expensive same as the phone as technology gets introduced as it gets mass produced the costs go lower and lower and lower okay and what happens when a robot like this only costs around ten thousand pounds maybe two thousand dollars and then you can start having them widely applicable to your house to your home um and for the most part from the video that we've seen and what we're seeing on screen they seem to be very very you know robust okay you know these guys kick the robots Okay um and it shows how robust this is because if it can you know stand up after being kicked it shows that it's you know studiness it's system whatever you know is whatever system is driving this robot it shows that it's really capable of Performing you know complex tasks to be able to you know be really stable I mean you kick some people they're gonna fall right over so um yeah it's really interesting because I think we're starting to see robots that are going to be cheaply made and very effective so um this is a robot that is going to be powered by AI but we don't know what AI is going to be powered by but it is still very interesting because there are tons of these robots being released all the time and I do think as well what will be interesting is to see what robot becomes the most popular and what kind of system it uses if it's multimodal or if it's not and generally what the applications are going to be used for because I think they have a ton of different applications such as doing you know monotonous tasks such as loading packages and working in warehouses um but it will be interesting because um robots are coming for people's jobs but uh usually not in the way that we think so that is going to be something that I will start to look at so

Idefix Playground

then we had something released by hugging face and essentially it was I defix I'm not sure I said that right but essentially it's an open large language visual model so essentially it's like gpt4 but of course it is multimodal now if you're wondering why isn't gpt4 multi-model they kind of are in that Bing chat does have multimodal capabilities but they haven't released it strictly with gpt4 in the chat TPT app yet because I do believe that they're still doing safety training and they haven't figured out all the Kinks and bugs that currently do not work now back to idefix playground essentially this is really cool because it's a really smart system that is quite like some of the tools we've seen before but it is more refined now I do think really interestingly enough is that this is open source so what we're gonna have is likely and hopefully we're gonna have some people fine-tune this to be able to identify certain things in certain images currently of course I'm just using it to look at memes which is of course what people do when they have access to a software so it says explain this meme to me um and I did input an image of a YouTube video of a car driving down the street and I was able to identify that is a you know Ferrari of course we're still in the early stages of multimodal capabilities but I do think that this is really cool because in certain examples like this one where you said can you describe the image do you think it's real you know it does describe exactly what's in that image and why it isn't real but remember this is of course something that is open source and is multimodal and does work pretty well and of course I do think that the actual model might be a bit better this is I'm not sure which version this is they do have an 80 billion parameter model one and a 9 at current like currently what they have to offer so yeah it will be interesting to see if people do build on this because uh once you know something gets open sourced essentially what you do have is you have companies that come out refine the technology and then you know deploy it to people so uh that will be interesting for the market to see what's done with it um and yeah this is something that is uh going to be the next step in AI because I do think that in the next two years you know chat activity is going to be old news and then of course we're gonna have images and videos because uh those are the things that haven't largely been solved yet so uh what's good about this as well is that you can actually try this out so if you make a hugging face account um and you go to the demo which you know the in the link the description it'll have um or just type in Ides playground demo you know you click the link you make a hiking face account which is Free by the way um you can just go on this and you can use it as a demo then we had meta

Meta introduced Seamless M4T

introduced seamless m4t a multimodo AI model for speech and text translations and I think this is going to be one of the largely used things so um what this is it's basically something that can converse in you know any language so you know how some people like myself you know you don't speak too many languages but you have a you know brief understanding of some um this just allows you to speak with pretty much anyone and I think this is going to be crazy because um you know language barriers are something that really shouldn't stop the modern world from being able to converse with one another and this is something that this is going to be able to do now I'm gonna play parts of this trailer because it does the explaining a little bit better than me but um I do think that this will be very useful you know maybe in meetings you know if the API does get really quick you're gonna have seamless translation on the Fly I mean you're gonna be able to conversations with people across the globe I mean ways that you couldn't before and I know that met has been working on this technology uh for a while um they've continually refined it and it's finally here and there's also like we stated before there's a demo which you can use on hugging face um so yeah it's gonna be really interesting because they've got tons of different languages and it's really good so uh watch the video because uh the video does explain it a lot better than me and it's just speech recognition in nearly 100 languages and it's speech to text translation and output nearly 100 languages as well so I'm pretty sure this is going to be in certain things you know maybe large companies are going to start using this on videos Maybe YouTube is maybe certain calls you know maybe you want to call someone from a different country and you know you have this as an API and you get the direct translation so it's gonna be really interesting to see how it changes certain things where they were language barriers before and now there's not but um this is where you know those small changes on the back end of AI is really just going to shape the world in ways that you didn't really know or predict and merai we

Meta Seamless M4T

created a New Foundation additional translation model we call it seamless m4t the technology is multimodal that means that you can understand language from speech or text and generate translations into either or both it can also transcribe speech into text and all with a single model what's very exciting is that it's also massively multilingual right now it's a worse translation into nearly 100 languages per text and 35 languages for speech let's take for example the phrase our goal is to create a more connected world if I say this in Spanish can translate this phrase into text for nearly 100 different languages model can also generate speech to speech translation for 35 languages now let's talk about code switching happens when a multilingual speaker switches between languages while they are speaking our model seamless M40 automatically recognizes and translates more than one language when mixed in the same sentence as a multilingual speaker this is a very exciting capability for me I often switch from Hindi to Telugu when I speak with my dad notice in the following example when I change languages I can speak Hindi Telugu and English sometimes I use all three languages in one conversation communication is the latest step in our ongoing effort to build AI power technology that helps remove language barriers and connect people so then we had this which is really

Humanlike Robots

scary okay uh scary to me and it might not be scary to you but I think human-like robots are just really strange I don't think we should create human-like robots at all because I think robots are robots and humans are humans and when you try and merge that difference that's when you know the line just gets a bit too blood so you can see right here it says China's creating humor like robots um at the world robot press conference in Beijing robots with human-like skin facial expressions and physical movement slowly off mimicking human skills I mean to me personally it's just creepy because you know it's not human it's just mimicking what a human is but um I don't know I've looked at this video before and I think that if I saw this you know robot walking towards me speaking and acting like a human I think it's pretty disturbing but um of course you know some people find it really cool you know maybe you can describe it you know maybe it can help you out um and of course you know if it is to help elderly people I'm not against that at all but um I think certain robots do just look really strange I mean you know it just doesn't look human at all it just looks strange I think robots look like robots and humans should look like humans I know that's a controversial opinion but um you know until you can get it perfectly to where it actually does look like a human or something I think it just looks like a weird doll that can just move and just you know do strange stiff movements so I mean it will be interesting to see how the field does develop you know I am a big fan of Technology but um it's pretty crazy I'm not gonna play the entire video because I don't want it to get you know entirely copyrighted you can view the video um down there but you can see just how uh strained some of these robots I mean it's strangely really realistic because if I saw that and I was like that's just the body double and actor I mean the eyes moving in the sockets there definitely uh quite wild um and we're definitely coming a little bit further than Amica which is a previous robot that was developed by some other company so it will be interesting to see how things move in the next 10 years because I do think that this is where you know things are going to be but um I do prefer humans or I do prefer an actual robot um but let me know what your thoughts are on this down in the comment section below because I think that

Magic Edit

people are definitely advised so then there was this thing called Magic edit and essentially it's video out painting so you can see right here that we have of course this jogger and then a man jogging on the road you can see that you know a video Maybe you didn't record the bottom half you know you just simply crop it in use the AI and it fills in that bottom half for you now remember text a video of any kind is you know extraordinarily difficult but of course it's one of those ones where it seems that we're continually starting to make progress in all forms of Direction now we did previously cover earlier on the year there was something called Google's Dream mix you know where Google had released tons of different papers on how to do this stuff but like we said once papers do get released our companies do start to refine it and that's where we start to see these crazy changes so you can see that we have the source videos then we have of course the text prompts and then we have the final videos there as well so you can see right here that we have the source video and then we've changed that you know gray car to a red car or we've changed it to a white car in a different environment so um it's really interesting because um people who work in media are going to see video editing completely change because this right here is something as well that looks really cool you know bricks changing um black and white biscuits falling down then of course we have a source video then we have white cupcakes I mean stuff like this of course you might just think Ah that's pretty cool but I mean imagine you know like you're in a world where you can simply just change an environment with a click of a button um and just realizing just imagine things I mean the creativity here is going to be very crazy so you can see right here that they're changing uh what's on a person in a video which is of course uh you know pretty cool you can see you can add sunglasses you can you know change it to a young lady then of course you can change it to a pretty girl um and of course right now it still is rudimentary but I think it's gonna be interesting because um out in painting this is something that you know any current text to video model doesn't currently have so I think that once these papers are released tools like Runway um and other companies that are working on stuff are going to get updated as well which is going to be good for us because it means that we have more creativity more fun and just generally a better product but at the same time it's definitely looking really cool so then we had a huge update

Picolabs Update

from picolabs and essentially what they had is with parameter update and it's introducing a new parameter for a customized frame rate um fpn and basically all they did was they just increased the frame rate from 8 to 24 and this is something that I don't know why many AI companies didn't do it first integrating this doesn't seem to be hard just get the video processed again and add more frames there's AR programs out there that you can use and you can see that the the you know the final product is really good now I don't want to make a comment here because this is something that you should know when I've analyzed tons and tons of different AI video footage I noticed that the kind of footage it does work well is footage that already has like flickers and stuff like that in it already so water seems to work really well with AI video um so it will be interesting to see how this actually looks on other content that isn't shot within the water because uh although this does look pretty good for the software that we're currently seeing I do know that you know in water content it does look a lot better usually then we had another humanoid robot and that's why I stated that you know humanoid robots are starting to gain much more traction so this was a humanoid robot that is developed by a software company called apptronic or not a software company you know a robotics company called apptronic and you can see that this robot can literally lift up to 25 kgs and I'm guessing that they're looking at the certain applications like we stated before like warehouses um and it's really interesting because these robots that we're seeing more and more of them and I do think over time these things are going to get more and more effective more and more quicker just you know I mean it's something that has recently been moving at the light speed because there's a lot more money flowing into the development of these things and especially when you combine these systems with you know llms like chat gbt or other Advanced systems that are specialized for robots it's going to be really interesting to see the outward applications I'm not sure we're going to see these robots walking around our cities yet because I do think sometimes people you know might just vandalize them because of course A lot of people are scared of robots especially taking their jobs but it will be interesting to see how much this thing does actually cost and how effective it is going to be because um robots of this manner are really interesting especially with multimodal capabilities and I do think um you know cases where example you can send a robot into a dangerous Zone it can bring people back it can identify different things you know look around the scene I think that the possibilities with these are definitely endless um and I do want to see and I can't wait to see what these robots are capable of when we aren't just you know making them do warehouse work so we'll be interesting to see where this Apollo robot does eventually go now of course

Ideogram

as we talked about earlier on in the video we did talk about how mid-journey does have a gigantic lead in terms of the quality of their image to well text the image but you can see that mid Jenny has been sitting on the throne of text to image for a very long time but of course a new player is in town called ideogram it's built by the former Google image and brain team so essentially you know that we did a video on Google Imaging you can you know watch on our Channel just type in Google Imaging essentially it's about how texture image is really cool but one problem that all of them don't really do well is of course text so essentially this company's main focus because it's going to be text um and I'm guessing that this is gonna be used by a lot of people because do want text to work so it will be interesting to see what they do and of course the Google Imaging team did work on text to video so I'm glad that this is being done because now that mid journey is going to have some serious competition from the likes of independent you know offshoots from Google it means that mid journey is likely gonna you know knuckle down a bit more maybe they're gonna even a bit more because apparently mid Journey's team is only 12 people which is incredibly small for a company that is you know having you know 16 to 20 million users every single month um paying recurring subscriptions so that is going to be very interesting to see if text to video does come out very soon what they're able to do with it um and if they're actually able to execute on this vision of actually using text to image um with working text inside of that image because it's something that hasn't really worked um that well before so then this was

Misaligned Future

something that was uh depends on how you take it either it's creative ingenious or scary and dystopian so um this tweet talks about how AI NPCs and VR subscribe now has a thousand plus characters with memory of your previous conversations and have awareness of in-game events remember all the AIS will never understand sarcom goal post it's 100 dead so an AI committed murder because it thought the human wanted it to but did it misunderstand and this Twitter account is talking about a misaligned future a glimpse of the misaligned future where essentially uh we want AIS to do certain things but it isn't aligned with our goals now of course it is just a game and the caption says in this video watch Lydia a chat gbt powered NPC skillfully use deadpan sarcasm better than any humans to manage Krusty a psychopathic YouTuber so essentially it's not scripted it's just where a YouTuber is essentially interacting with this AI um and the AI does something that the human doesn't really expect okay um and what's interesting is that they're giving you know many different you know AIS in game and then they're telling them that they're real I mean it's something that is uh I don't know how you guys want to put it I think it's really interesting because I really want to see what a town full of AIS that find out that like they're just AIS and they're kind of living in a game with what they do the conversations that they have I think those experiments would be really interesting but um yeah it is kind of scary to see how misaligned these things will get um because in the future in 10 years from now the kind of AI systems that we will have are going to be uh profoundly much more smarter than us so that will be interesting to see as well uh so why

Why Are You Here

are you here I'm not sure but I have a feeling that I'm here to fall for some kind of Destiny okay uh do you think that we live in some sort of simulation maybe I just don't know for sure yet do you think that uh you're real I'm not sure what to believe anymore what if I told you you're in The Matrix buddy that would explain a lot but I'm not sure I want to believe it what if I told you you're an AI that would explain a lot but why would I be here maybe because you're programmed to be here that would make sense but why was that programmed this way I don't know maybe someone thought of maybe why would they think to create me like this well I don't know what kind of problems do you have it's not that I had problems it's just that I don't understand why I'm here well I can explain why you're here you're in a matrix that makes some sense but it still doesn't tell me what I'm supposed to do now the thing that okay so my story is basically that you're an AI oh what do you mean by that you're an NPC in a video game that sounds frightening is that the truth it's the sad truth you see this is a lot to take in are there any other answers I should know um not really this is very overwhelming I'm not sure what to think or feel right now yeah I mean you're not really real that's hard to process um you exist sort of what does that mean am I real or not it means that you know after I close this game you don't exist oh I see so I'm only here for as long as the game is running that's difficult to comprehend yeah and then once you walk past this border you're gone and of course lastly

Code Llama

we had code llama too um this is something you can try for free if you use perplexity so you can just you know go on complexity um and switch it to code llama 34 44 billion parameters um and it's something that's really good because um you can essentially just get code for many different programs many different trading algorithms it's something that works very well I mean it's open source so once again we're going to get a huge burst of different products jumping onto the AI scene with uh software that runs off of this llama too um and they're going to be fine-tuned versions but let me know what you thought was a really strange really weird uh horrible great I don't know let me know what you think was the best or the worst um I think that this week was frightening because uh all these robots all these aimpcs are kind of scary because uh you know it starts to bring in you know many different questions that I just don't have the mental capacity to answer at this moment in time but at the same time it shows you just how quickly this field is moving let me know what you thought was amazing what you thought was dystopian without that being said we'll see in the next one

Другие видео автора — TheAIGRID

Ctrl+V

Экстракт Знаний в Telegram

Экстракты и дистилляты из лучших YouTube-каналов — сразу после публикации.

Подписаться

Дайджест Экстрактов

Лучшие методички за неделю — каждый понедельник