Mindblowing o3 Prompts, OpenAI Models & More AI Use Cases
27:05

Mindblowing o3 Prompts, OpenAI Models & More AI Use Cases

The AI Advantage 18.04.2025 44 803 просмотров 1 360 лайков обн. 18.02.2026
Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
Go check out Notion mail to start using the AI Auto Label feature powered by GPT-4.1 and more, all for free! 👉 https://ntn.so/AIAdvantageMail What a crazy week in AI! OpenAI released o3, o4, the GPT-4.1 series, Codex CLI and more. Claude and Grok both got major updates to bring them up to speed with ChatGPT. (Or at least, where ChatGPT was one week ago...) We'll break all that down and more in today's video! Links: https://x.com/OpenAI/status/1912560057100955661 https://www.youtube.com/live/sq8GBPUb3rk?si=9dcPmO4UOW7IxCJi https://openai.com/index/gpt-4-1/ https://youtu.be/FUq9qRwrDrI?si=xnLvY3Hkd3slINdr https://github.com/openai/codex https://x.com/AnthropicAI/status/1912192384588271771 https://blog.google/products/gemini/video-generation/ https://x.com/GeminiApp/status/1909721519724339226 https://app.klingai.com/global/release-notes https://x.com/canva/status/1911826850323615767 Chapters: 00:00 What’s New 00:54 OpenAI Releases 01:14 OpenAI o3 and o4-mini 02:47 GPT-4.1 04:05 o3 Use Cases 15:43 OpenAI Image Library and OpenAI Codex 16:48 Notion Mail 19:05 Canva Updates 21:45 AI Video Updates 23:31 Claude Research 24:34 Grok Studio 25:15 Gemini Deep Research #ai #NotionPartner #NotionMail This video is sponsored by Notion. Free AI Resources: 🔑 Get My Free ChatGPT Templates: https://myaiadvantage.com/newsletter 🌟 Receive Tailored AI Prompts + Workflows: https://v82nacfupwr.typeform.com/to/cINgYlm0 👑 Explore Curated AI Tool Rankings: https://community.myaiadvantage.com/c/ai-app-ranking/ 💼 LinkedIn: https://www.linkedin.com/company/the-ai-advantage 🐦 Twitter: https://x.com/IgorPogany 📸 Instagram: https://www.instagram.com/ai.advantage/ Premium Options: 🎓 Join the AI Advantage Courses + Community: https://myaiadvantage.com/community 🛒 Discover Work Focused Presets in the Shop: https://shop.myaiadvantage.com/

Оглавление (12 сегментов)

  1. 0:00 What’s New 237 сл.
  2. 0:54 OpenAI Releases 62 сл.
  3. 1:14 OpenAI o3 and o4-mini 354 сл.
  4. 2:47 GPT-4.1 293 сл.
  5. 4:05 o3 Use Cases 2871 сл.
  6. 15:43 OpenAI Image Library and OpenAI Codex 279 сл.
  7. 16:48 Notion Mail 589 сл.
  8. 19:05 Canva Updates 644 сл.
  9. 21:45 AI Video Updates 432 сл.
  10. 23:31 Claude Research 225 сл.
  11. 24:34 Grok Studio 158 сл.
  12. 25:15 Gemini Deep Research 426 сл.
0:00

What’s New

This is the first week in AI where many people claim that we reached AGI. And look, that's a bold claim and it's definitely debatable. But the fact of the matter is with the release of Ofrey from OpenAI, which we'll talk about today. You'll see in this very video chat GPT doing things that were unthinkable before. And the crazy thing is even beyond that, this was an amazing week in AI, state-of-the-art video generation, many more Chat GPT updates, multiple other models coming out of OpenAI, and so much more. In this week's episode of AI News You Can Use, the show that rounds up all the AI releases from this week filters by the one that you can put to work today and then we test them or have a first look at all the things that came out in Generative AI this week. As you might see, I'm still on the road, my last week here in Japan. What a country. I don't want to waste too much of your time here. We're here to speak about Generative AI releases, but I will be using some pictures from my trip as a part of our testing here. All I'll say is if you're tech enthusiast and you haven't been to Japan before, oh boy, does this country have a lot in store for you. And it's so peaceful. All right, let's get into
0:54

OpenAI Releases

this. Starting with the OpenAI releases. So there's multiple things here and I'm sort of going to bundle this because we had two big announcements. One of them released multiple developer focused models that are not available in chat GPT. This is the GPT 4. 1 series. These are non-thinking models. Again not in chat GPT only available for the API for
1:14

OpenAI o3 and o4-mini

builders. And the second announcement and that is the big one this week is the release of free the full version just free. That's the brain that was powering deep research. And the thing that made it so damn good, you can now access it through chat GPT on all Pro Plus and Teams plans coming to the education and enterprise plans in about a week from now. Just to make it clear, this is not accessible on the free plans. With this 03 release, they released also all four mini models. As you can see here on my pro account, there's one called O4 Mini and High. But in this video, I really want to focus on the O3 model because I think this is the one that most people care about. On this show, I always try to focus on the tools that really push the bar on what's possible with AI. And while O4 and 04 Mini High are impressive and perform better than O3 Mini and O3 Mini High before that, naming is a mess. The joke has been made a thousand times. They'll fix it soon. Point is, the one model for most users that really makes a difference this week is 03 because it's head and shoulders above anything that we had access to before except maybe deep research. But this thing is way faster and just different in certain behaviors. So before we get into some examples and use cases here, which I find absolutely fascinating. I can't wait to show you some of these, let's just briefly outline some of the specs and which models you would care for based on what you're doing. So as I said, 03 is sort of the king here. As they show in the blog post where you can see all the details of this, it leads on all the benchmarks and beats everything including this 04 mini model on both settings. So that's a really important takeaway. If you want the highest intelligence in chat GPT as of now, you pick all free on a paid plan. The other
2:47

GPT-4.1

release this week, which was GPT 4. 1, GPD4. 1 mini, and GPD4. 1 Nano, those are developer focused models that are particularly good at generating code. These are only accessible through the API where you pay per request, and you can't find these inside of Chat GPT. And again, these are developer focused models both in terms of usage and use cases. They're really good at generating code. And already a lot of tools are adopting GPT 4. 1 under the hood. Like for example, the sponsor of today's video, which we'll talk about later, which is Notion Mail. They instantly switched the model that helps you manage your emails to 4. 1 because it just performs better and works better. So you can expect the increase in performance across all AI powered applications that used other GPT models before, but these are non-thinking models. So let's not even compare the benchmarks to these thinking models that are 03 and O4 mini because the thinking models are always going to crush non-thinking models. So one more time, GPT series non-thinking models. There's a new 4. 1 and smaller versions of it. O3 and 04 thinking models. O3 is the best one that you want to put to work today there. And if you wonder how they compared to some of the competition on the benchmarks, well, we started putting together a table to show you all of that. But in the middle of creating this table, we found a website that already does this really well, not affiliated, just think this is a great resource. And as you can see, you can sort by the various benchmarks that have been published. For example, in GPQA 03 leads above all the competition here.
4:05

o3 Use Cases

Anyway, let's get to the more interesting part here because this is a model that you'll be using and I want to focus on 03 here. I think we talked about the different alternatives and releases and benchmarks enough here. The thing that really matters to me and what I really try to make the point of this channel is how to democratize the access to all of these tools, techniques, workflows. And I haven't seen a tool since deep research that does it as well as free now. Because this is a big point of free isn't just a thinking model. It's a thinking model that has access to tools. What kind of tools? Well, all tools that Chachibd has before we saw limited releases, okay, when 01 preview came out, you couldn't upload images, you couldn't upload files, you couldn't access memories, it could not use the data analysis tool and so on. All of these things work automatically inside of 03. Now, this is a massive deal and we'll see why in the examples that I have lined up here next for you. But before we do that, I want to share one more thing. Look, I made it my life to stay on top of all of this, test all of it, build organization around it, and we teach it at scale to individuals and organizations. At this point, not often does a release stop me in my tracks and maybe rethink the way I use these tools, technology, the way it integrates into my life. As I said in the opener, many people are stating that this is the first release that can be considered AGI. And I think that statement is debatable because if you ask 100 people, you're going to get 100 definitions for AGI. And I'm making these videos to give you my personal take on this. And the conclusion sort of is this model makes me want to rerun every prompt I have ever tried because it has the ability to surprise me with its intelligence and level of insight. And while that might not be the case on every single thing you run, it did happen regularly to me in my first initial look here. So with that being said, let's have a look at the some of the things that I tried here right away. One of them just being this quick prompt from Matt Schumer on Twitter here which says, "Do intensive research on Egor Pogani and give me a massive report on everything you find. " Obviously, you would replace my name with your name here and you could try this for yourself. And what happens here is just so different from anything we've seen except maybe deep research. It goes ahead, it starts thinking, it starts making a plan. Then it uses the web search tool to pull up multiple sites. It pulled up my Instagram, LinkedIn, my ex. It found the company website, the company LinkedIn. We're now going to be posting twice a day on our company LinkedIn. Go check that out if you're on that platform. And it even found my SoundCloud, which I haven't used in years, but fair enough. Then it goes ahead and figures out that, hey, I even made a typo writing my own name. Embarrassing. And it proceeds to think about this, pull up more links, and it just finds like a dozen different LinkedIn posts and other websites, including some of my university, etc. And it comes up with something that Deep Research would always refuse, a profile on me with all the details, and it's quite comprehensive while not being overwhelming, which sometimes is the case with deep research. So, the guardrails on this are lower than what we saw with deep research, and it's quite impressive. It pulls together many relevant things. But let me tell you, there's like six or seven mistakes here that are just flat out wrong. And those are not mistakes that Ofrey made. Those are mistakes that websites that this reference made. For example, it shows that my nationality is Austrian. This is actually not the fact. I lived in Austria for most of my life, but I still have a Slovakian passport. But the great thing is it gives it source right here. And this website here, which I don't know much about. It's called business ABC. It just somehow assumes that my nationality is Austrian. That's why has it in its report. So, it's not perfect, but clearly the big point here is that the guard rails are lower. With deep research, you could have never researched a person. And some of these other insights here actually made me lean back in my term and kind of surprise me myself. It made a comprehensive list of all the different paid products we offer. And it kind of extrapolated the signature content themes that made me think like, yeah, we do we did used to do a lot of AI workflow stacks and workflow videos and do less of these days. We should do more of that. They get less views, but they seem to communicate more value. And then the overarching theme of the channel is productivity hacks. I mean, we're doing this in the form of generative AI tools and workflows, but that's what a lot of people here care about. So, I thought that was great. And then it ran things like the SWAT analysis here in the end. And that's really one of the big points that I see here. It's not that none of this was possible before, but you had to bring in the context yourself. You had to know which prompts to use and then you had to take the time to actually use them to wait for the results. And you think about it, reprompt, things like that. This does all of it for you. You don't need to engage a search function. give it specific websites that you wanted to consider. You don't need to prompt for it to come up with a structure like this that really makes sense here. It just figures all that out for you with a combination of thinking models and tool usage. So obviously you could take this little prompt, replace your name, run this on yourself or on anybody else. Maybe they'll restrict this down the line, but as of now this works and in deep research this never used to work. This was one of the most interesting things that I wanted to try. Now it works in here. But in the second use case that I want to share here, it really shows the depth of the capability. And this one was inspired by Matthew Burman on Twitter here doing something that many people on the internet are trying right now, which is giving it a picture and saying figure out exactly where this person is. I did it on this wholesome picture of myself from the recent Japan trip and Ofrey went off and started thinking about it. What it did is it cropped into the different parts and started analyzing them in detail, translating the Japanese language and interpreting some of these signs. Like for example, here it says the crest has 16 pedals. Although it looks a bit stylized, it zooms in onto different parts and based on the caption and the text, it comes up with the conclusion that this is clearly the Maji Jingu Shinto shrine. I hope I'm pronouncing that correctly. That is located in Shibuyaku, Tokyo, a park in the middle of Tokyo. And here's the clues it followed. And it even states that this is the inner precinct exhibition area, which is exactly right. Then I was like, "Wow. " Okay, so it didn't just get the location right, so it even got the exact area within the quite small shrine right. Impressive. But it did have several clues here like the text and the logo. So I wanted to give it something a bit harder. I took a second picture for my script which had no text, no logos. Matter of fact, it's night and I figured there's no real clues on this one. And I asked the same thing, figure out exactly where this person is. And after about 2 and 1/2 minutes of thinking, but it started by cropping into the various parts, coming up with some assumptions and various locations. This could be maybe Kamakura's Hokuji Temple or the Kaji Temple in Kyoto. It started running some internet searches to find more images that it could then doublech checkck with. And in typical 03 fashion, as you might know from deep research, it didn't just find a website or two. It found 11 ranging from Trip Advisor to specific travel blogs to Reddit to YouTube. Like, yeah, it pulled up a YouTube video in the process to cross reference these snippets of this image. And it noticed things like, hey, it only shows one horizontal rail while the other shrine that I was considering typically has two, but that could be a trick of the vantage point. So, I want to verify it. And then it went ahead and looked for that specific trail that it assumed and again pulled in a bunch of different websites and analyzed images on there to cross reference this and eventually chose me multiple images that match mine and says that no this is not my first assumption but it's a small hillside bamboo walk inside Kodai temple in Kyoto's Higashyama district which is 100% correct. This is not some obvious spot. The main temple is at the bottom. You kind of have to walk up into the gardens and then this is a small passage tucked away in the woods behind the smaller shrine that is behind the big temple. And it even says the person is standing roughly halfway along the illuminated bamboo trail that leads visitors back towards the main hall of Kodai. This is exactly correct and it even gives you the exact coordinates. At this point, I was just blown away by the accuracy of this thing and by the fact that it looked up dozens of websites, looked at the images there and cross referenced them and could identify this. And I told myself, "Okay, okay, okay. I got to give it something even harder. How about something without a person and without a trail that people walk? So, I looked at my camera roll and found this little snap of this beautiful little bird in a river here. And I just figured, how could it know this? I mean, it's just a river with some rocks and a generic looking Japanese building in the back. Well, long story short, it got this right, too. What? But I thought it was really interesting that initially it pulled up Python and wanted to extract the metadata from the image, which I thought of already. And all of these images didn't include that. I screenshotted the original image to make sure that there's no GPS coordinates embedded in this to make this actually challenging. So that didn't end up working. So then again, it did what it did before. It made some assumptions, ran some searches, looked at all the websites, cross referenced this, and after only 43 seconds, it concluded that this elegant gray heron is standing in the Shurikava Canal. And then again, it gave the exact coordinates and a description of where you can find this exact spots with other reference images that look similar to this. And yep, that's the exact location where I took this picture. God damn, how is it this good? At this point, I felt a combination of excitement, but also I was kind of pissed at the tool. I just figured like, how can it be this good? How can I break this thing? So, I just took out my phone at this coffee shop where I was preparing this video earlier today. And I looked out and I was in a beautiful coffee shop overlooking the river of Osaka. And when I looked closely on the other side of the river, there were these rocks that were tied up in a net with some turtles sitting on it. I shot a little video of you so you can visualize what I'm talking about right now. And funnily enough, when I shot the video, a little raven landed there and all the turtles that were sunbathing kind of hopped off the rock. That was kind of a random moment. Anyway, with the telephoto lens on my iPhone, I snapped a picture of this and I just wanted to break it at this point. give it something that it cannot figure out. And yeah, when I ran the prompt after 2 and 1/2 minutes, it actually got it wrong. Boy, it figured that this is in central Tokyo, which is not right. This is in Osaka. So then when I followed up and said it wrong, not in Tokyo, it did several weird things like running Python code on top of the images. And look, it found these signs and it tried to run some filters on top of it to extract the contrast to read what's on them, but the image was just too low resolution for it to read what's on this sign. Uh, side note, if they plug in some of the top AI upscalers right now into this, it might have gotten this right. But after a total of 10 minutes of thinking, it actually figured out that these are the sandbags that line the inner mode of Osaka Joe, the Osaka Castle in Osaka, Japan, which is correct. I mean, it's not exactly next to the castle. It's like a 25minute walk down the canal. But just based on this low resolution image with one full prompt, I got it. I mean, this is just crazy. So, I hope this example shows you the capabilities of this thing. It's not just the fact that the thinking model is just better than anything we've seen before or free, but it's also the fact that it has access to every tool in chat GPT. And while before even some power users might not have took full advantage of the tooling in chat GPT just because you don't think of it or often because one is too lazy. I mean, if you run a search, you kind of have to wait for it. Then you need to take that context prompt on top of it. use the data analysis tool again that took some time. It does all of that for you now and just presents you with the result. So this works really well if you do something that I refer to as goal-based prompting rather than instruction based prompting. So you just tell it where you want to go and it figures out how to get it. That was always a strength of these thinking models. Yet any AI skill that you might have required over the past few years is still relevant here because if you know how to use this individual modalities, you can open up this thought process and reprompt it to do specific things for you if the results are not what you want. And there's still value to prompt engineering because you can do more intricate things like I started trying afterwards here, but I'm still playing and I don't want to spend an hour showing you every single use case that is interesting here. If I find enough, I'll do a separate video as per usual. But let me just say that many of these business use cases, which I now teach in workshops, like this meeting analyzer, just work better than anything before. In some cases, they're just slightly better. In other cases, they're stunning. And just the same as you would have got with a deep research or proper prompting. But one thing remains constant and that's that the bar to get some of these advanced results that even in combination with AI would have took quite a bit of effort to achieve you can now get from a simple oneline prompt and that the guardrails on things like searching for specific people were lowered on this release just unlocks a whole new world of possibilities. If that's agi or not is up for you to decide, but it's certainly impressive. So that's my first look at 03 and then O4 mini and 04 mini high are just versions of it that are faster and then cheaper when you use them for the API. But for most people, 03 will be the one that you want to be really playing with right now. And then for developers, the three different variations of the 4. 1 models are excellent at generating code. And if you need a non-thinking model, that's the new go-to. They will be depreciating the 4. 5 in the API. By the way, that's also kind of interesting.
15:43

OpenAI Image Library and OpenAI Codex

Okay, so there's two more things out of OpenAI here. One of them is a mobile release, which I'll just quickly cover. There's a new tab called library where it just collects all of your images. So if you're generating a lot of images, they created this interface that is sort of more focused on that. That's really nice. But more importantly, there was one more OpenAI release this week, which is OpenAI Codex. And I thought about spending more time on this, but I decided to focus this video on all free cuz I think most people will get more value out of that. Plus, this new release of Codeex CLI that they, by the way, fully open sourced is an Agentic product that works in the terminal and does pretty much exactly what Claude Code does, which was released by Enthropic about 2 months ago, and I've been using that tool regularly since its release. It's really great. It basically runs in your terminal. So I guess you do need the very basics of development but these are the things that with chat the assistance you can figure out in like an hour and then you can basically do things like drag and drop images into your terminal and let it build little applications for you on demand on full autopilot. It's obviously the same thing that Entropic did with claude code and I think it's amazing that OpenAI decided to open source this entire program and they said there will be more coming but as of now if there's one thing you should really check out this week it's all free. Okay let's see what's next.
16:48

Notion Mail

All right, so this next one is exciting for me because if you've been following the channel for a while, you'll know that I love using Notion. We use it for our prompt databases and the entire operating system of the AI advantage. Whether that is the content production, the entire organization or the community. Come to think of it, we actually kind of use it for absolutely everything. I try to avoid other apps at all costs so we have one source of truth and also one source of context for our various AI workflows. And as they reached out and wanted to partner on a video, I was like, "Yes, please. " because the new feature they're releasing is actually something we're interested in ourselves and that is notion mail. It's a brand new email platform that launched earlier this week and there's a couple of features that I want to highlight to you because you might just want to consider this for your workflow. The basic concept is simple. You just connect your Gmail to notion mail and it upgrades your regular inbox with a ton of AI powered organization and writing tools. When it comes to organization, they have the auto label system which is my personal favorite feature in the notion mail release. And it's one of the things when you start using it, you wonder how you were doing it before. Basically, you just tell the AI that you want a new label, and the AI does all the work of the labeling, or you could also call it categorizing. And then once the AI labels all your emails, you can review the decisions and tell it if it did well. Basically, fine-tuning its decision-making, so you can rely on it moving forward. And sure, you have this feature in other email clients, but I found this AI integration to work particularly well as they're a new product and they don't have a million existing features that the new one has to comply with, and in practice, it just works better. Next up is a feature that you've probably seen in other platforms, too, but it's an important one, and that's when AI drafts your email responses. But Notion Mail has one twist that sets it apart from the competition here. And that's a fact that usually the AI writing your email drafts or the agentic app that you kind of plug into it lacks the context of well, anything that matters. And guess what? As this is notion, you can pull the context from your notion pages into the email client really simply. So for example, if you want to draft something and say, "Hey, you can find all the relevant information about the sponsorship in this page. " You just type at, pick the page, that's it. The AI has all the context from that page. Now, obviously, there's way more ways to add context here, but if you already work in Notion, this is the smoothest thing ever, obviously. So, those are two, but there's many more ways Notion Mail will make your email management smoother, and it's completely free to use. So, why not give it a try today? All you need to do is click the link at the top of the video description, link your Gmail account, and get started. Thanks again to Notion for sponsoring this video. And now, let's get back to the next piece of AI news that you can use. Next up, we
19:05

Canva Updates

have some major updates from Canva with all of their AI integrations. And I always like covering Canva because I feel like they take a lot of the trends and things that have proven to work out there and integrate them into their suite of products, which just I understand it as taking complex tools for visual communications and packaging them and then making them easy to use. And they already were doing a ton with AI. But now during their Canva create conference they launched several new features which I would like to show you and I believe a few of these deserve our attention here. So we tried everything that is available now and I think most significantly this new updated Canva AI is really close to something that one might conceptualize as a visual AI tool that creates social media content for you. As you can see we just took the example of a online dog product subscription brand trying to create social media content and kind of assisted us in the process here. And what I like here is that gives you various options, right? They have a lot of these templates already and this AI integration just helps you navigate between those and it really creates an experience that is very smooth for somebody who might not be involved in this at all. I mean, can you get higher quality results by dialing in every single detail? Sure. But is this one of the quickest workflows out there? Yeah, probably. I mean, look at that. You pick one of these presets and then you can edit all the text in there, hit export, and you're done. And honestly, it's kind of funny. It's getting to a point where for some of these smaller brands that just need to get their message out there, it's really hard to argue for some of these fully manual workflows where you pick your templates, fonts, maybe do photo shoots for custom imagery, put it all together in Photoshop, and then iterate on top of that. You just say what you want, you get a bunch of presets, you click one, change the text, and you're done. It's kind of great, and I think this is an easy recommendation for a lot of small businesses that just need content, right? But they went even further. They integrated some of these AI trends that we've seen across all different applications. I mean the one that most people were surprised by is Canva code. Like Canva served the opposite of an technical audience, but now it can write code and create little landing pages and so on. We tested it again on our imaginary dog subscription box business here and hey it built the interface with a onboarding flow with surveys and everything and canvas. So I found that interesting but me personally when I looked through all of this one thing really caught my attention and that is the fact that Canva has now their own type of sheets integration where you can upload data and it works with it. Now, unfortunately, this is one of the features that isn't available as of now, but I thought this looked really interesting because you can add a sheet with data into it and then quickly process everything in there, making bulk workflows really simple. So, from what they showed is you could upload the list of all the products or campaigns or pretty much anything. And then you can use Canva at scale to generate multiple assets and just get it done instantly. And there's also data analysis features and things like that. But I just thought that looked quite promising, especially for social media agency owners. So, if you do that, have a look at these Canva features. But now let's move on to the
21:45

AI Video Updates

next story. Okay, next up we have some major updates to AI video generators. Namely, the one where most people agree right now is probably the best one. Google's V2 is for a short amount of time freely accessible for Google's AI studio. That's their developer interface. And with new releases, they put their products in there for a short amount of time for developers to try out. And as of the week that I'm recording this video in, V2 is available in there. So you can just go ahead and generate a few videos for yourself if you haven't tried this before. This model is head and shoulders above many of the competitors as you can see by some of the examples that are on screen right now. But then allow you could only access it directly through the API where you paid per usage and in their experimental whiscap that we talked about recently but now you can just try it. So that's kind of nice. And then beyond that we have cling 2. 0. And cling was one of the models that was competing for this top rank with V2. And let me tell you this 2. 0 model is incredible. It's probably the best textto image model that we have right now. Even better than V2. It's debatable. Both have their strengths and weaknesses, sure. But the point here is AI video has definitely not plateaued. And I wanted to show you a few comparisons of this Cling 2. 0 model and some of the prompts that we usually like to run on top of them for the show to give you a feeling of how this stuff is evolving. Before we do this though, I do want to say it's ridiculously expensive. Whereas right now a 5-second video costs 100 credits. And here's the pricing for the credits only. So getting one clip that needs a few iterations might just cost you $5 to $10, which is just crazy for most people. But as you can see from some of the comparisons on screen right now, the quality is incredible. Particularly this little text to video example of a sloth hanging out in an inflatable donut really impressed me. The water physics are near perfect here. Even the way his foot is submerged in the water kind of works. I mean AI video has become indistinguishable in many cases, but with releases like this and V2 now it's just pushing that high bar further and further. Interesting. Now let's move on to the next one. So here
23:31

Claude Research

we have a few announcements and I'm going to bob these together. I would call these competitors catching up to OpenAI. And if you're using either Enthropics Claude or Grock, then great. There's a bunch of new features that you should know about. Namely, I'll start with Enthropics research feature, which is essentially deep research now integrated into Enthropics Claude. Many people love Claude for good reasons. The tone is really unique. The coding abilities are state-of-the-art. Although now Gemini 2. 5 Pro and GPT4. 1 are competing for that slot. Which one is the best there? Really, that depends on who you ask and what exactly you're doing. But the point is they added this research feature which has one integration that none of the competitors have as far as I know. And when I say competitors I mean Google's Gemini Advance, ChatGpt and Groc and that's our ability to connect to external documents within the deep research. So you could do something like connect it to your Gmail account, Google calendar, Google docs and it will consider those files while conducting its deep research here. As of now it's available in beta for max team and enterprise users in select countries. So yeah, this is a paid feature and no, the standard plus plan does not have this as of now. And then
24:34

Grok Studio

we have updates from Grock releasing Grock Studio and there's things like data analysis as we know it where it can generate documents, code and then visualize them and in something like a canvas where you can collaborate on the content with the AI here. They're just clearly playing catchup with chat GBD features that have been here for well almost 2 years in the case of data analysis code interpreter as it was called back then. And they also ship the memories feature similar to the one in chat, but it doesn't reference your past chats. We talked about that extensively last week. If you haven't heard that discussion, I think it's an important one. Consider checking that segment from last week's video out. Point here being Grock now has basic memories and adding some of these essential features that I still love using all the time in chat GPT. Okay, onto the next one, which is
25:15

Gemini Deep Research

update to Gemini's deep research feature. Now, it uses Gemini 2. 5 Pro under the hood, which has enjoyed massive popularity since its release. Now, if I may, I'll give you a quick recap of the history of the deep research feature. It first was released within Gemini Studio. They called it deep research, but it was very basic. Then, OpenAI came along and implemented their O3 model under the hood of their deep research product, which blew everybody's mind. I think still today that is the most useful AI thing out there and they just adopted that descriptive name. And now, as we just covered, everybody's following suit with pretty much everybody having a deep research feature now, including a croc and even more specialized players like Perplexity. So, what happened actually last week, and I'm featuring it just now, is that Google updated the model under the hood of their deep research with their best model, Gemini 2. 5 Pro. And I was really excited to see this because I love me a good deep research and I was hoping that this could be some competition for OpenAI, but long story short, I genuinely do think that OpenAI is still better. I ran a few test prompts here and just consistently I found the OpenAI results still be better. We'll put up a result here on screen right now from our testing. Basically the gist of it is that the OpenAI reports are more pointed towards what you exactly prompt. So one might say that the prompt adherence is better or in more human terms it just listens to you and understands what you actually mean gets you exactly those results. Now don't get me wrong still a great product just at this point in time open version is still superior. Oh one more note the deep research results they're more structured like a research paper. If that's what you need then I would definitely recommend Gemini's implementation of deep research. And if you use Gemini Advanced it's perfectly fine. I'm just saying if you start nitpicking OpenAI's results just more relevant and I find some of the insights from those reports to be truly novel whereas I don't have that experience with pretty much all the competing products. All right, so that's really everything I have for this week. Next week I'll be back in my studio. If you enjoyed this, make sure to subscribe. I do the show every single Friday. My name is Eigor and I will see you very soon.

Ещё от The AI Advantage

Ctrl+V

Экстракт Знаний в Telegram

Транскрипты, идеи, методички — всё самое полезное из лучших YouTube-каналов.

Подписаться