# Open AI's New 'MAJOR BREAKTHROUGHS Stuns The ENTIRE AI Industry! (GPTS +GPT-4  Turbo)

## Метаданные

- **Канал:** TheAIGRID
- **YouTube:** https://www.youtube.com/watch?v=89nOj7igU-c
- **Дата:** 06.11.2023
- **Длительность:** 31:27
- **Просмотры:** 22,061
- **Источник:** https://ekstraktznaniy.ru/video/14701

## Описание

https://openai.com/blog/new-models-and-developer-products-announced-at-devday

Welcome to our channel where we bring you the latest breakthroughs in AI. From deep learning to robotics, we cover it all. Our videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on our latest videos.

Was there anything we missed?

(For Business Enquiries)  contact@theaigrid.com

#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience
#IntelligentSystems
#Automation
#TechInnovation

## Транскрипт

### Segment 1 (00:00 - 05:00) []

overshadowed by the very big announcement that they recently did made of course GPT for Turbo we're going to get into absolutely everything and break down some of the key points that many people did Miss so pay attention to this video because there's a lot to break down but it's really worth it because this next stage in AI is going to be completely different gp4 turbo gp4 Turbo will address many of the things that you all have asked for so let's go through what's new we've got six major things to talk about for this part number one context length a lot of people have tasks that require a much longer context length gp4 supported up to 8K and in some cases up to 32k context length but we know that isn't enough for many of you and what you want to do gp4 turbo supports up to 128,000 tokens of context that's 300 pages of a standard book 16 times longer than our 8K context and in addition to longer context length you'll notice that the model is much more accurate over a long context number two more control we've heard loud and clear that developers need more control over the model's responses and outputs so we've addressed that in a number of ways we have a new feature called Json mode which ensures that the model will respond with valid Json this has been a huge developer request it'll make calling apis much easier the model is also much better at function calling you can now call many functions at once and it'll do better at following instructions in general we're also introducing a new feature called reproducible outputs you can pass a seed parameter and it'll make the model return consistent outputs this of course gives you a higher degree of control over Model Behavior this rolls out in beta today and in the coming weeks we'll roll out a feature to let you view log probs in the API all right number three Better World Knowledge you want these models to be able to access better knowledge about the world so do we so we're launching retrieval in the platform you can bring Knowledge from outside documents or databases into whatever you're building we also updating the knowledge cut off we are just as annoyed as all of you probably more that GPD for's knowledge about the world ended in 2021 we will try to never let it get that out of date again gp4 turbo has knowledge about the world up to April of 2023 and we will continue to improve that over time number four new modalities surprising no one Dolly 3 gp4 Turbo with vision and the new tech to speech model are all going into the API today we have a handful of customers that have just started using Dolly 3 to program programmatically generate images and Designs today K is launching a campaign that lets his customers generate devali cards using Dolly 3 and of course our safety systems help developers protect their applications against misuse those tools are available in the API gp4 turbo can now accept images as inputs via the API can generate captions classifications and Analysis for example be my eyes uses this technology to help people who are blind or have low vision with their daily tasks like identifying products in front now something I didn't see many people talking about because I don't think the feature is that good is that they spoke about how they have a new Texas speech software and I think 11 Labs has really set the standard for what we expect effect when we now look at texture speech software because if you have actually listened to some of the stuff that they literally just released on the website I'm talking just a couple of minutes ago it doesn't sound that great I mean it sounds good but if you do want to compare it to stuff like 11 Labs is definitely far superior but what's oddly enough is that the one that I speak to when you're using chat gbt and you speak to it like it's a person sounds much better than their new text speech model but if there's anything we know about open AI is that the models are going to get better and better but take a listen to it because it does sound kind of interesting but it's something that many people didn't talk about but it's a very small part that you probably should just pay attention to and with our new text of speech model you'll be able to generate incredibly naturally natural sounding audio from text in the API with six preset voices to choose from I'll play an example did you know that Alexander Graham Bell the eminent inventor was enchanted by the world of sounds his ingenious mind led to the creation of the graphophone which etches sounds onto wax making voices whisper through

### Segment 2 (05:00 - 10:00) [5:00]

time this is much more natural than anything else we've heard out there voice can make apps more natural to interact with and more accessible it also unlocks a lot of use cases like language learning and voice assistance speaking of new modalities we're also releasing the next version of our open-source speech recognition model whisper V3 today and it'll be coming soon to the API it features improved performance across many languages and we think you're really going to like it then we have is what I think is the most insane section of the video so um yeah just check this out because uh everything is about to change in this part right here to everyone so we're going to show you what gpts are how to use them how to build them and then we're going to talk about how they'll be distributed and discovered and then after that for developers we're going to show you how to build these agent like experiences into your own apps so first let's look at a few examples our partners at code. org are working hard to expand computer science in schools they've got a curriculum that is used by tens of millions of students worldwide code. org crafted lesson planner GPT to help teachers provide a more engaging experience for middle schoolers if a teacher asks it to explain for Loops in a creative way it does just that in this case it'll do it in terms of a video game character repeatedly picking up coins super easy to understand for an eighth grader as you can see this GPT brings together code. org extensive curriculum and expertise and lets teachers adapt it to their needs quickly and easily next canva has built a GPT that lets you start designing by describing what you want in natural language if you say make a poster for Dev a Dev dat reception this afternoon this evening and you give it some details it'll generate a few options to start with by hitting canvas apis now this concept may be familiar to some of you we've evolved our plugins to be custom actions for gpts you can keep chatting with this to see different iterations and when you see one you like you can click through to canva for the full design experience so now so in this section of the open AI Dev day what they had was super interesting because this section they talk about custom gpts for companies and although we don't get to partake in making these gpts and we can make our own gpts I think this is going to revolutionize pretty much every online business and pretty much how we interact with most modern applications because with canva GPT I don't think there's any reason a person would use this over the traditional method of course if you want some more creative control and of course if you do want to do more specific things I guess you could but being able to just simply talk into a large language model and then get the exact graphic that you want is going to revolutionize absolutely everything and I can't imagine what other companies are going to be launching their own gpts and how they're going to be integrated and maybe if those companies have to pay an upfront cost for that GPT and if open AI is making money from this but this is definitely something that I think a lot of people aren't talking about and I think we're going to see a lot of applications the way most people use them change rapidly we'd like to show you a GPT live zapier has built a GPT that lets you perform actions across 6,000 applications to unlock all kinds of integration possibilities I'd like to introduce Jessica one of our Solutions architects who is going to drive this demo welcome Jessica thank you s everyone thank you all for being here my name is Jessica Shay I work with partners and customers to bring their product alive and today I can't wait to show you how hard we've been working on this so let's get started so to start where your gbt will live is on this upper left corner I'm going to start with clicking on the zapier AI actions and on the right hand side you can see that's my calendar for today so it's quite a day everybody use this before so it's actually already connected to my calendar to start I can ask what's on my schedule for today we built gpts with security in mind so before it performs any action or share data it will ask for your permission so right here I'm going to say allowed so gbt is designed to take your instructions make the decision on which capability to call to perform that action and then execute that for you so you can see right here it's already connected to my calendar it pulls into my information and then I've also prompted it to identify conflicts on my calendar so you can see right here it actually was able to identify that so it looks like I have something coming up so what if I want to let Sam know that I have to leave early so right here I say Let Sam um no I gotta go um chasing

### Segment 3 (10:00 - 15:00) [10:00]

gpus so with that I'm GNA swap to my conversation with Sam and then I'm GNA say yes please run that Sam did you get that I did awesome so this is only a glimpse of what is possible and I cannot wait to see what you all will build thank you and back to you Sam thank you Jessica so those are three great examples in addition to these there are many more kinds of gpts that people are creating and many more that will be created soon we know that many people who want to build the GPT don't know how to code we''ve made it so that you can program the GPT just by having a conversation we believe that natural language is going to be a big part of how people use computers in the future and we think this is an interesting early example so I'd like to show you how to build one all right so I want to create a GPT uh that helps give Founders and developers advice when starting new projects um I'm going to go to create a GPT here and this drops me into the GPT Builder uh I worked with Founders for years at YC and still whenever I meet developers the questions I get are always about how do I you know think about a business idea can you give me some advice uh I'm going to see if I can build a GPT to help with that so to start GPT Builder asks me what I want to make and I'm going to say I want to help startup Founders think through their business ideas and get advice after the founder has gotten some advice uh Grill them on why they are not growing faster all right so to start off I just tell the GPT a little bit about what I want here and it's going to go off and start thinking about that and it's going to write some detailed instructions for the GPT um it's also goingon to let's see ask me about a name how do I feel about startup Mentor that's fine uh that's good so if I didn't like the name of course I could call it something else but it's you know going to try to have this conversation with me and start there and you can see here on uh on the right in the preview mode that it's already starting to fill out the GPT um where it says what it does it has some like ideas of additional questions that I could ask um and you know what I actually so it just generated a candidate of course I could regenerate that or change it but I sort of like that so I will say that's great and you see now that the GPT is being built out a little bit more as we go now what I want this to do um how it can interact with users I could talk about style here but what I'm going to say uh is I am going to upload transcripts of some lectures about startups I have given please give advice based off of those all right so now uh it's going to go figure out how to do that and I would like to show you the configure tab so you can see some of the things that were built out here as we were going um by the Builder itself and you can see that there's capabilities here that I can enable um I could add custom actions these are all fine to leave um I'm going to upload a file uh so here is a lecture that I picked that I used to that I gave with some startup advice um and I'm going to add that here in terms of these questions uh this is a dumbb one the rest of those are reasonable uh and like very much things Founders often ask um I'm going to add one more thing to the instructions here which is be concise and constructive with feedback all right so again if we had more time I'd show you a bunch of other things but this is uh this is like a decent start and now uh we can try it out over on this preview tab so I will say um what's a common question what are three things to look oops for when hiring employees at an early stage startup now it's going to look at that document I uploaded um it'll also have of course all of the background knowledge of gp4 that's pretty good those are three things that I definitely have said many times um now we could go on and it would start following the other instructions

### Segment 4 (15:00 - 20:00) [15:00]

and you know Grill me on why I'm not growing faster but in the interest of time I'm going to skip that uh I'm going to publish this only to me for now uh I can work on it later I can add more content I can add a few actions that I think would be useful um and then I can share it publicly so that's what it looks like to create a GPT with thank you by the way I always wanted to do that after like all of the YC office hours I always thought man someday I will be able to make a bot that will do this and that'll be awesome so with gpts we're letting people easily share and discover all the fun ways that they use chat GPT with the world you can make private gpts like I just did or you can share your Creations publicly with a link for anyone to use or if you're on chat GPT Enterprise you can make gpts just for your company and later this month we're going to launch the GPT store you can list a GP thank you I appreciate that you can list a GPT there and we'll be able to feature the best and the most popular gpts of course we'll make sure the gpts in the store follow our policies before they're accessible Revenue sharing is important to us we're going to pay people who build the most useful and the most used gpts a portion of our revenue of course we have the section where you can make your own gpts and I think this is going to be where a lot of money is made in terms of normal people who don't have particularly a codebase So currently what you have is a bunch of companies that wrap chat GPT and essentially they change the software into something that is usable for external applications so there's some software out there that you know you can use to talk with your PDF there's some where you can talk to it in a specific style but chat gbt and openai are essentially giving you that capability and they're saying look you can make your own bot you can upload it to our website um and if loads of people use it we're going to pay you a percentage of that and I think that is going to be very similar to the App Store and with this I think it's going to be really interesting because now people have a huge financial incentive to create these Bots share these Bots and we're going to be getting a whole lot of creativity and customization so I would say if you are someone who is looking for that next business model online this is definitely going to be something that is really interesting and something that if you're particularly good at AI you can definitely take advantage of today we're making that a lot easier with our new assistance API the assistance API includes persistent threads so they don't have to figure out how to deal with long conversation history built-in retrieval code interpreter a working python interpreter in a sandbox environment and of course the improved function calling that we talked about earlier so we'd like to show you a demo of how this works and here is Raman our head of developer experience welcome Roman thank you s good morning wow it's fantastic to see you all here it's been so inspiring to see so many of you infusing AI into your apps today we're launching new modalities in the API but we are also very excited to improve the developer experience for you all to build assistive agents so let's Dive Right In imagine I'm building wust a travel app for Global explorers and this is the landing page I've actually used gp4 to come up with these destination ideas and for those of you with the Keen ey these illustrations are generated programmatically using the new darly 3 API available to all of you today so it's pretty remarkable but let's this app by adding a very simple assistant to it this is the screen we're going to come back to it in a second first I'm going to switch over to the new assistant playground creating an assistant is easy you just give it a name some initial instructions a model in this case I'll P gp4 turbo and here I'll also go ahead and select some tools I'll turn on code interpreter and retrieval and save and that's it our assistant is ready to go next I can integrate with two new PR of this assistant API threads and messages let's take a quick look at the code the process here is very simple for each new user I will create a new thread and as these users engage with their assistant I will add their messages to the threads very simple and then I can simply run the assistant at any time to stream the responses back to the app so we can return to the app and try that in action if I say hey let's go to

### Segment 5 (20:00 - 25:00) [20:00]

Paris all right that's it with just a few lines of code users can now have a very specialized assistant right inside the app and I'd like to highlight one of my favorite features here function cooling if you have not used it yet function calling is really powerful and as Sam mentioned we're taking it a step further today it now guarantees the Json output with no eded latency and you can invoke multiple functions at once for the first time so here if I carry on and say hey what are the top 10 things to do will I have the assistant respond to that again and here what's interesting is that the assistant knows about functions including those to annotate the map that you see on the right and so now all of these pins are dropping in real time here yeah it's pretty cool and that integration allows our natural language interface to interact fluidly with components and features of our app and it truly showcases now the harmony you can build between Ai and UI where the assistant is actually taking action but next let's talk about retrieval and retrieval is about giving our assistant more knowledge Beyond these immediate user messages in fact I got inspired and I already booked my tickets to uh to Paris so I'm just going to drag and drop here this PDF while it's uploading I can just sneak peek at it very typical United flight ticket and behind the scene here what's happening is that retrieval is reading these files and boom the information about this PDF appeared on the screen and this is of course a very tiny PDF but assistants can par long form documents from extensive text to intricate product specs depending on what you're building in fact I also booked an Airbnb so I'm just going to drag that over to the conversation as well and by the way we've heard from so many of you developers how hard that is to build yourself you typically need to compute your um beddings you need to set up chunking algorithm now all of that is taken care of and there's more than retrieval with every API call you usually need to resend the entire conversation history which means you know setting up a key value store that means like handling the context window serializing messages and so forth that complexity now completely goes away with this new stateful API but just because OPI is managing this API does not mean it's a blackbox in fact you can see the steps that the tools are taking right inside your developer dashboard so here if I go ahead and click on threads this is the thread I believe we're currently working on and see like these are all the steps including the functions being called with the right parameters and the PDFs I've just uploaded but let's move on to a new capability that many of you have been requesting for a while code interpreter is now available today in the API as well that gives the AI the ability to write and execute code on the fly but even generate files so let's see that in action if I say here hey will be four friends staying at DC rbnb what's my share of it plus my flights all right now here what's happening is that code interpreter noticed that it should write some code to answer this query so now it's Computing you know the number of days in Paris number of friends it's also doing some exchange rate calculation behind the scene to get this answer for us not the most complex math but you get the picture imagine you're building a very complex like finance app that's crunching countless numbers plotting charts so really any task that you'd normally tackle with code then code interpreter will work great for you all right I think my trip to Paris is solid so to recap here we've just seen how you can quickly create an assistant that manages state for your user conversations leverages external tools like knowled and retrieval and code interpreter and finally invokes your own functions to make things happen but there's one more thing I wanted to show you to kind of really open up the possibilities using function calling combined with our new modalities that we're launching today while working on dev day I built a small custom assistant that knows everything about this event but instead of having a chat interface while running around all day today I thought why not use voice instead so let's bring my phone up on screen here so you can see it on the right awesome so on the right you can see a very simple Swift app that takes microphone input and on the left I'm actually going to bring up my terminal log so you can see what's happening behind the scenes so let's give it a shot hey there I'm on the keynote stage right

### Segment 6 (25:00 - 30:00) [25:00]

now can you greet our attendees here at Dev day hey everyone welcome to Dev day it's awesome to have you all here let's make it an incredible day isn't that impressive you have six unique and Rich voices to choose from in the API each speaking multiple languages so you can really find the perfect fit for your app and on my laptop here on the left you can see the logs of what's Happening behind the scenes too so I'm using whisper to convert The Voice inputs into text an assistant with gp4 turbo and finally the new TTS API to make it speak but thanks to function calling things get even more interesting when the assistant can connect to the internet and take real actions for users so let's do something even more exciting here uh together how about this hey assistant can you randomly select five death day attendees here and give them $500 in open eye credits yes checking the list of attendees done I picked five devday attendees and added $500 of AP credits to their account congrats to Christine M Jonathan C Steven G Lis K and surj S all right if you recognize yourself awesome congrats um and that's it a quick overview today of the new assistant CPI combined with some of the new tools and modalities that we launched all starting with the Simplicity of a rich text or voice conversation for your and users we really can't wait to see what you build and congrats to our lucky winners actually you know what you're all part of this amazing OPP Community here so I'm just going to talk to my assistant one last time before I step off the stage hey assistant can you actually give everyone here uh in the audience $500 in open credits sounds great let me go through everyone all right that function will keep running but I've run out of time so thank you so much everyone have a great day back to you Sam pretty cool huh all right so that assistance API goes into beta today and we are super excited to see what you all do with it anybody can enable it so next one you're about to see is Sam alman's closing words and I do want you to pay attention to all of the words that he says because in one sentence he does mention that these AI agents or these AI systems that he's building are going to be able to plan and of course do more advanced things in the future and they're working on increasing these capabilities so like I said before I don't think AGR is far out of the scope but it's really interesting to see just how far we are coming in such a sort space and the progression and what they're hinting SL leaning towards so pay attention because these closing statements show us exactly where the company is headed and of course if you had been watching the channel you understand that their core values have been updated so this is of no surprise time gpts and assistance are precursors to agents are going to be able to do much more they'll gradually be able to plan and to perform more complex actions on your behalf as I mentioned before we really believe in the importance of gradual iterative deployment we believe it's important for people to start building with and using these agents now to get a feel for what the world is going to be like as they become more capable and as we've always done we'll continue to update our systems based off of your feedback so we're super excited that we got to share all of this with you today we introduced gpts custom versions of chat GPT that combine instructions extended knowledge and actions we launched the assistance API to make it easier to build assisted experiences with your own apps these are our first steps towards AI agents and we'll be increasing their capabilities over time we introduced a new gp4 turbo model that delivers improved function calling knowledge lowered pricing new modalities and more and we're deepening our partnership with Microsoft in closing I wanted to take a minute to thank the team that creates all of this open a has got remarkable Talent density but still it takes a huge amount of hard work and coordination to make all of this happen I truly believe that I've got the best colleagues in the world I feel incredibly grateful to get to work with them we do all of this because we believe that AI is going to be a technological and societal re Revolution it will change the world in many ways

### Segment 7 (30:00 - 31:00) [30:00]

and we're happy to get to work on something that will Empower all of you to build so much for all of us we talked about earlier how if you give people better tools they can change the world we believe that AI will be about individual empowerment and agency at a scale that we've never seen before and that will Elevate Humanity to a scale that we've never seen before either we'll be able to do more to create more and to have more as intelligence gets integrated everywhere we will all have superpowers on demand we're excited to see what you all will do with this technology and to discover the new future that we're all going to architect together we hope that you'll come back next year what we launch today is going to look very quaint relative to what we're busy creating for you now thank you for all that you do thank you for coming here today so with Sam alman's closing statement I think it's clear that we're moving more away from just a standard chatbot to specific applications and to AI agent AGI and I think that's going to be really evident over the next two years because they're saying they're going to be increasing the capabilities of these models so I wouldn't be surprised if we start getting a bunch of different AI agents that we can literally just Deploy on our computers and of course fundamentally you have to remember that this is going to change the internet so if you have any predictions leave them down in the comment section below and um yeah it'll be interesting to see what happens tomorrow
