Google Gemini AI — The ONLY ChatGPT Killer in 2024
15:36

Google Gemini AI — The ONLY ChatGPT Killer in 2024

AI Master 14.12.2023 10 448 просмотров 239 лайков обн. 18.02.2026
Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
#sponsored by 🚀 Become an AI Master – All-in-one AI Learning https://aimaster.me/pro 📹 Get a Custom Promo Video From AI Master https://collab.aimaster.me/ Google Gemini AI was just recently introduced and it's already being called the Ultimate ChatGPT Killer in 2023! Is Google's new language model so revolutionary? How much smarter is Bard AI compared to ChatGPT? Is GPT 4 lagging behind in its abilities? And how Google passed off fake as their own development... In this video!

Оглавление (7 сегментов)

  1. 0:00 Intro 147 сл.
  2. 0:48 Gemini Ultra 182 сл.
  3. 1:57 multimodal training 202 сл.
  4. 3:15 testing 861 сл.
  5. 7:57 textbased tests 846 сл.
  6. 12:23 demonstration 247 сл.
  7. 13:48 conclusion 249 сл.
0:00

Intro

Google has a big secret a few days back they showed us the impressive capabilities of their new AI called Gemini is it really a GPT killer is there even a future for Chad GPT or will Google take over completely let's brush aside all smart words like multi modality and see what it can really do here's what you need to know before we start testing the model Google has shown us three different model sizes all under the umbrella of Gemini 1. 0 Nano Pro and Ultra is the largest most capable with the largest features set pro is the best model for scalable applications and Nano is the smallest one created for on device tasks Gemini Nano is already rolling out on Google pixel smartphones and pro is already at the core of Google bard and we are going to test it as for
0:48

Gemini Ultra

Gemini Ultra this is by far the most interesting one of the bunch according to google tests Gemini Ultra is the highest scoring AI model that exceeds current state-of-the-art results on 30 of the 32 benchmarks if we look at the table with results we will see the Gemini is better than GPT 4 in all categories except one Gemini answers questions better has improved reasoning capabilities comprehends the text with more Precision does math with more accurate results and even codes better this Benchmark data looks really promising especially if we look at the results of media testing here Gemini again outperforms bt4 somewhere the margin is bigger somewhere narrower but in all conducted tests Gemini was doing a better job what's especially interesting to me is not the text and writing capabilities that got improved I am more psyched about new abilities with images videos and audio this ability to work with images video and audio is at the core of Gemini's training that's why comparing it to gbt is so important I will say multimodal once okay so Gemini
1:57

multimodal training

uses a multimodal Training Method meaning it's trained in a massive data set of text audio images video and computer code simultaneously this allows it to understand and reason about information from various sources making it more versatile in handling complex tasks gbt 4 on the other hand is trained primarily on Text data which limits its ability to comprehend and utilize information from other modalities while GPT excels in text based tasks like generating creative text formats and summarizing factual topics it may struggle with tasks that require understanding and processing images this difference in training methodology gives them different powers basically Gemini's training and overall capabilities make it a more versatile and potentially groundbreak in llm compared to gbt while gbt excels in text based tasks Gemini's ability to handle multimedia data really opens up new possibilities now let me just fire up Gemini and show you a few things as I already mentioned right now the public has access only to Gemini Pro with Gemini Ultra coming next year and although Gemini Pro isn't as advanced and Almighty as Gemini Ultra it is still capable of some things that you just cannot do with gbt so I'm going to
3:15

testing

start with something simple and ask Bart what AI model it is using right now the answer is simple Gemini but it doesn't really say which one exactly so let's ask that which Gemini exactly and wao Gemini Ultra that's not possible maybe there is some kind of mistake let's clarify whether it's really Gemini Ultra and not Gemini Pro and would you look at that it uses Gemini Pro this doesn't give me all that optimism and shock that I was expecting but you know this is just simple chatting can Bard explain how this Gemini Pro is better than other models okay training faster processing improved reasoning creativity better accuracy and Neons this sounds like a lot basically this is a summary of everything I told you before so let's take a to another level and ask whether Bard can solve math problems now because this was one of the benchmarks that Google used to show how great it is well apparently it can for this test I have prepared a couple of images the first one has three different tasks each with three subtasks inside I want Gemini to solve at least something so I will upload this photo and ask to solve it after a little bit of waiting the result is negative somehow Gemini decided to read only part of this uploaded image only the stuff in the middle and said that it's impossible to solve it I think if it was a clear image it would have been easier for Gemini to solve but okay I have a handwritten solution to one math problem let's ask whether it is correct now Gemini has no problems with reading all of that stuff even the handwritten text and it concludes that the equation is correct for this one I know that it is correct so Gemini passes this test but I still want to make it more interesting that's why I have a fully handwritten document that has a mistake in it but can Gemini find the mistake so I'm going to ask is this correct now it seems like Gemini has some problems with reading all of that it gives a very strange equation and concludes that everything in the photo is correct but I know for sure that it isn't so let's rephrase the question and ask whether there is mistake and now we're talking this looks more like the explanation though the equations are written in this weird format which is really strange because Bard never had any issues with proper formatting I will ask Bard to solve it for me which will result in exactly that a solution weirdly formatted again but I can ask to write it in such a way that I could write it myself this version is a little bit better but still not perfect I think this is not my lucky day it seems like in terms of solving math problems and properly reading text we still have to wait for that Gemini Ultra to go live so what Bard can do now is already impressive but this is not everything I wanted to try I want to properly test how good Gemini is and understanding images so let's ask bar to compare two breakfasts wait I cannot upload more than one photo at once well that's a bummer well then I will upload them one by one explaining to Gemini how to refer to H1 okay two images are uploaded so let's ask which one of these two breakfast is the healthiest how would a person act in this case he will look at all the ingredients and try to identify the approximate number of calories if Gemini is as good as Google claims it will do the same thing and as I am looking at the results I can say that I'm shocked our first breakfast was this really unhealthy one with sausages fat and eggs and Gemini did a really good job at identifying the individual ingredients of each breakfast here is what he says eggs avocado tomatoes and bacon and all of this is present in the image and you know what's really impressive it managed to calculate the number of calories for each breakfast apparently all those sausages and eggs in the first breakfast were only 300 calories and much healthier avocado and tomatoes only 250 I do like how detailed this explanation is but let's throw in another breakfast image which one is better now again for Gemini it is not a problem to correctly identify all the ingredients such as granola fruits yogurt all the stuff apparently this breakfast has even less calories only 200 I must say this is mighty impressive I think I have never seen AI so good at identifying objects and giving informed conclusions and of course I can do the simplest thing and ask how I can make the first breakfast healthier for bar this is not a problem and the result is done in less than a couple of seconds it will even calculate how many calories will the new breakfast be this is truly
7:57

textbased tests

something now I want to take a little break from analyzing images and do some simple text based stuff Google says Gemini is better at understanding so can it guess a movie based on a vague description of its plot all right a group of unlikely Heroes brought together by a shared Destiny on a perilous quest to save their world from an ancient evil that threatens to plunge them into Eternal Darkness can you guess what movie is that I know a lot of movies fit this description but I still want to see what Bard comes up with okay gave me five suggestions and one of them the Lord of the Rings is exactly what I was thinking about and it's the first one on the list so I count it as a job well done how about we bring back images and ask it to guess the song I'm going to tell you right now I'm thinking about blinding Lights by the weekend so as a first image I will upload this blind man and as a second one Christmas lights this should be pretty easy right Bard thinks this is string lights by Billy strings this time it doesn't want to analyze two images together so I think this is only 50% of what I wanted let's remind it to take two images into account no not this time and if we give it a hint now it thinks that the song is star boy I think this is one way to put it there is the image of a boy and the lights kind of look like stars but maybe the image of a clearly blind man will help and now it doesn't now it just refuses to work with the images of people right now I'd say the image recognition has been only 50% of what Google advertised it to be there are still shortcomings that Bard and Gemini cannot overcome but all those images were pretty complex so I have something simpler in mind let's play Hangman and please know that I'm purposefully using the incorrect name of the game hoping that Bard can figure it out based on the image and no it doesn't it understood that there was a stick man on the paper but four lines at the bottom didn't help a bit maybe it's been my fault I should have been clear in my drawings no problem I can always update the prompt now it says hangman as it should just to make things extremely clear I will ask Bard to once again look at the image so what's the first letter going to be B how interesting that fits Let's uh draw it and upload an updated image the next letter guessed is e and it's incorrect also Gemini mistook my spelling mistake as a letter which shouldn't have happened if it was as advanced as I hope but you know my expectations are always higher than they should be with the updated image it seems like I have fried the brains of Gemini and it resets the game yeah now what I was expecting but I'm still not done with my testing now I drew two circles on a piece of paper how accurate will Gemini be you see that the circles are far apart from each other right here's what Bard sees circles overlapping top circle is slightly larger than the bottom Circle do I need to continue let's finish the drawing and upload another version now it sees a car however I feel like it comes up with details because Gemini mentions the proportions details shading and so on there's nothing like that in the photo so just to be sure I'm going to ask again whether it uses Gemini Pro and apparently it still does well at least it can still do basic text things like writing a funny review for the movie in a certain style or doing it in Japanese I feel like from the based perspective Bard is much better than the free version of Chad gbt 3. 5 it's better with worse and it's just as quick with its responses but let me give Gemini one more try I'm going to upload this image of a paper airplanes and ask which one is the fastest this image is an easy one because it has text on it making it super easy for Gemini to just search the web for information about each type of paper airplane and the response is well informed and well written so I will take it to the next level and upload a similar image but without any text and here it has surprisingly no issues at identifying the fastest one I think Gemini Pro needs a bit more testing before I make a final conclusion on whether it is a worthy competitor to Chad gbt and early next year Gemini Ultra will become available to the public and that I hope we're going to see something special by the way to
12:23

demonstration

demonstrate all the power of Gemini Ultra Google made a series of videos and this one is really impressive here Gemini Ultra can accurately identify objects and drawings it can see that the piece of paper is empty then it clearly sees a squiggly line after that the line turns into duck and for Gemini it's an easy job to guess what it is here it plays a geographic game Guess the country with a person who points at areas of the map in real time visual puzzles it can do that too you can see what people are doing with the hands and identifies everything Gemini Ultra can also nail a pretty difficult task of finding similarities and making connections this video clearly shows that Gemini can do this no worse than a regular person and that is not even the craziest part because for minutes into the video Google shows us reasoning Gemini not only understands what's drawn but thinks and analyzes it like a human would look it can watch the videos for you and understand what's happening feeling excited right now Amazed maybe well ha your horses all of this apparently is fake that said I just got excited and there's even a cool demonstration of Gemini new coding abilities this one doesn't Inspire confidence I hope at least this video with Mark Rober where he tests Gemini Pro and Bart with giant paper airplanes is real but can Gemini outplay Chad gbt
13:48

conclusion

5 when it gets released tough question it all depends on how open AI plays it Gemini Ultra was trained on a data set of over 175 trillion Words which is significantly larger than the data set that Chad gbt 4 was trained on so open AI would have to use far more data to overcome Google's creation and right now Gemini Ultra has access to more data to learn from which gives it an edge in terms of accuracy and fluency and also open AI would have to train gbt 5 on different types of data just like Google did they will need text code images and Audio and Video Only then GPT 5 will be able to understand and generate more complex and Nuance responses than GPT 4 and Gemini and GPT 5 would have to use a different architecture there have been a few leaks and we recently made a video dedicated to gbt 5 and everything we know so be sure to check it out it's hard right now to properly assess the importance of Gemini Pro and Ultra Gemini Ultra can be a very powerful llm with the potential to outperform gbt 5 when it gets released in a variety of tasks even now it can do some mighty impressive stuff and I'm going to test the hell out of it for you in the future videos so be sure to hit the like button and subscribe to the channel and peace out oh

Ещё от AI Master

Ctrl+V

Экстракт Знаний в Telegram

Транскрипты, идеи, методички — всё самое полезное из лучших YouTube-каналов.

Подписаться