Sam Altmans Statement On AGI Is Bigger Than You Think!
17:00

Sam Altmans Statement On AGI Is Bigger Than You Think!

TheAIGRID 09.11.2024 36 583 просмотров 850 лайков

Machine-readable: Markdown · JSON API · Site index

Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
Prepare for AGI with me - https://www.skool.com/postagiprepardness 🐤 Follow Me on Twitter https://twitter.com/TheAiGrid 🌐 Checkout My website - https://theaigrid.com/ Links From Todays Video: https://www.youtube.com/watch?v=xXCBz_8hM9w&t=6s&pp=ygULeWNvbWJpbmF0b3I%3D Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos. Was there anything i missed? (For Business Enquiries) contact@theaigrid.com #LLM #Largelanguagemodel #chatgpt #AI #ArtificialIntelligence #MachineLearning #DeepLearning #NeuralNetworks #Robotics #DataScience

Оглавление (4 сегментов)

Segment 1 (00:00 - 05:00)

so Sam mman recently made a statement about 2025 that has a lot of people surprised and are debating currently whether or not this statement is true or completely false this is the statement in question what are you excited about in 2025 what's to come AGI excited for that so of course you can see there Sam mman clearly stated that in 2025 it's quite likely that we will have AGI and I think this is one of probably the most profound statements that we've ever had in any interview from any open AI employee now of course this statement does warrant some skepticism because AGI has been a thing that is heralded as this incredible piece of technology that most of us can't even fathom due to the immense possibilities that happen if AGI becomes a reality now one of the things that I found interesting was the fact that a lot of people who are actually working at openai commented on this statement and their comments are actually supporting samman's statement basically stating that look this statement isn't ridiculous it's actually on par with what we're seeing inside the company so someone that worked on the old one model this is noan brown he worked on reasoning at open Ai and has done a remarkable number of things to advance the state of reasoning in AI he actually stated that I've heard people claim that Sam Alman is just drumming up hype but from what I've seen everything he's saying matches the median view of opening eye researchers on the ground so this is where we actually get a look to what individuals are saying that is going on inside the company if Sam mman is stating this of course he does have an incentive to State the best about his company after all he is CEO and after all he does have to maintain a good Public Image and drum up hype for his industry but of course if individual researchers at the company are saying look this is not hype and this is basically on par with what researchers are saying then I think it gives us more confidence in that Sam mman statement is not something that is completely ridiculous but more in line with what these AI researchers are starting to say now someone else from open AI also shares the same view Adam GPT the GTM open AI actually stated this he said that it's not AI hype it says I've been intently listening to Sam for my 3 years at open aai and he is precise with his words and comments I personally think the disconnect where he is viewed as hyping things up is that he speaks very transparently about where and when things are going SL trending with the pace and progress of AI if things that are talked about down appear immediately people often feel disappointed and it's a bit of a catch 22 and I personally love that he communicates outwardly and publicly about our progress it's important we continue to accelerate on our path towards AGI so you can see here that Adam GPT actually supports samman's statement and basically says that look this is just a disconnect because most people can't fathom how fast the company is moving and this is truly where they are now one of the things that also intrigued me was the clear date 2025 for next year as you know it's going to be 2025 but I couldn't help but keep thinking that this date is a pretty significant one in question now one of the Twitter accounts that I follow and engage with quite a lot is the Twitter account the legendary Jimmy apples now for most of you probably don't know who this account is but it is account that somehow manages to get early a information months and years early before anyone else in the AI industry knows what's going on of course other than those at the company now in his bio you can see he says 2025 which means that if his predictions hold true it's quite likely that in 2025 something significant in the AI space is going to happen now he did put this 20125 date in his bio a couple of months ago so it's clear that he knew something SL know something that a lot of people don't and I do remember that it was I think around a year before we got any news about Orion that he was actually tweeting about Orion and most people didn't even realize what he was tweeting about until we knew what was happening now let's say this statement where stman clearly says AGI in 2025 and of course he's excited for that let's say you think it is a decent amount of hype whilst that might be your initial response to that Monumental statement take a look at this and want I want to show you guys is something from Dario amade who is the CEO of anthropic which produces the models clawed and is competing with open a for market share now I think it's clear to state that there are three companies at the frontier of the AI race currently open aai anthropic and Google and in this statement he actually gives his date for when AGI is going to come now it's important to note that in this article he doesn't state that AGI is Agi but he refers to it as powerful machine intelligence you can see here he says I dislike the term AGI but obviously many people are skeptical that powerful AI will be built soon and some are skeptical that it will be built ever at all and I think it could come as early as 2026 so if samman's statement that AGI is next year seems a little bit hypy I don't think it is when in the grand scheme of things the CEOs of Frontier labs are talking about the fact that advanced machine intelligence or powerful AI is going to come at 2026 which is only a year later than Sam alman's prediction now something interesting that Sam Alman also spoke about was of course the very fascinating scaling laws when we started yeah the core beliefs were deep loaring works and

Segment 2 (05:00 - 10:00)

it gets better with scale and I think those were both somewhat heretical beliefs at the time we didn't know how predictably better it got with scale that didn't come for a few years later it was a hunch first and then you got the data to show how predictable it was but people already knew that if you made these neural networks bigger they got better yeah um like that was we were sure of that um before we started and what took the like word that keeps coming to mind is like religious level of belief was that wasn't going to stop now of course the scaling laws are Infamous in AI for predicting the future Trends and what's interesting is that if you've been paying attention to the AI space you'll know that there are of course these new Incredible scaling laws train time compute and test time compute basically the fact that these models get predictably smarter at inference time if you allocate more computes to the actual model's thinking it's able to think for longer and harder about problems and come up with reasoning steps to be able to solve those same problems but with a greater degree of accuracy on this graph you can literally see that as train time compute goes up the accuracy on the benchmarks continue to increase and of course for the test time compute you can see that as more compute is applied these models get predictably smarter every single time and I do Wonder considering the fact that they've only applied a certain amount of computes once we get access to rediculous levels of comput where we have giant data centers that are spanning the size of entire cities I do wonder what this graph will look like so mman also said that the future of 01 scaling is quite interesting the CPO responded to this comment on Reddit that basically says how will you influence scaling llms will you continue scaling llms as per the initial scaling laws or will influence compute time scaling means smaller models with faster and longer inference will be the main focus they stated that it's not either or both it's better base models plus more strawberry scaling and inference time compute basically stating that look they're going to make the base model better and on top of making that base model better they're also going to scale up the amount of compute that they use on inference time meaning that overall you get a better base model to begin with even if you don't use any compute on it but with the added compute it just becomes even better now of course some people have been skeptical of ai's reasoning everybody had some reason of oh it's not really learning it's not really reasoning I can't really do this it's you know it's like a parlor trick and these were like the eminent leaders of the field and more than just saying you're wrong they were like you're wrong and this is like a bad thing to believe or say it was that there's got to you know this is like you're going to perpetuate an AI winter you're going to do this that and we were just like looking at these results and saying they keep getting better then we got the scaling results now whether or not llms can actually reason has been a widely debated topic for quite some time now ever since the Inception of llms some people just think that these things are statistical sance they're just simply predicting the next token and of course some research basically suggests that llms don't actually reason but rather mimic reasoning patterns from their training data I'm not sure if you guys watch the video I made around 3 weeks ago that basically talks about how Apple showed that on certain benchmarks when you change the names of certain individuals it actually drops the respective results by around 10 to 15% and that's even for Frontier models like 01 and GPT 4 which is quite surprising considering that these models are supposed to be reasoning and of course an unrelated Clause can decrease performance by up to 65% basically stating that these llms are pretty fragile when it comes to their reasoning abilities now this is not the only time that we've heard this narrative being spun in the AI Community one of the individuals that is famous for his stance on AI well I guess you could say against current llm architectures is of course Yan Lan I do want to say is of course one of the most notable figures in the AI industry as he did win the Nobel Prize of computing which is basically the touring award and one of the things that Yan Lan clearly states is that large language models are not a sufficient path to AGI because they miss essential capabilities for intelligent beings such as understanding and reasoning about the physical world basically stating that look llms don't really understand where things go and it's quite hard to infer this kind of physical data just from text one of the benchmarks that was essentially trying to prove this was of course simple bench and this was a reasoning Benchmark that was created with the soul goal of seeing if these models are actually able to reason about the physical world or standard problems quite like how humans do I don't have any examples of the questions here but some of them include things like if you put five ice cubes on the fire how many ice cubes will be there in the next minute and then having a multiple choice that has different numbers that the AI can pick from and then seeing what reasoning steps they took to get to their answer it's pretty fascinating considering the fact that the human Baseline is 83. 7% where the llm Baseline is around 40% and a lot of these questions don't really require Advanced knowledge you could ask them to any normal person and they would find out the answer pretty easily another thing that was actually really incredible from this interview was the fact that samman basically said that the path to AGI was solved this is the first time ever where I felt like we actually know what to do like I think from here to building an AGI will still take a huge amount of work there are some known unknowns but I think we basically know

Segment 3 (10:00 - 15:00)

what to go do and it'll take a while it'll be hard but that's tremendously exciting so samman stating look I think we basically know what to go ahead and do is something that is really profound stating that look we know exactly the kind of steps we need to get to artificial general intelligence the technology that's going to revolutionize the world as we know it which is quite akin to basically inventing the steam engine we know exactly where to go and I think that is really important considering the fact that a lot of times what companies I wouldn't say waste time doing but it's part of the natural AI ecosystem is that they have to conduct research now research is very time intensive it's very computer incentive and of course it costs a lot of money and the problem with that is that you might spend months on a problem only for you to realize that this isn't something that will work so then basically stating that look the path to AGI is clear now I think this is actually a really profound statement because it means that things are speeding up even more and one of the craziest things that was said recently in the open AI AMA where you had individuals asking open AI questions about the future of AI was that they said basically they can look at the current hardware and understand that with the current AI systems they're going to be developing the AGI is possible so there won't be any new architecture breakthroughs in terms of the hardware so for those of you guys thinking that we're going to need Quantum Computing neuromorphic chips or any of that kind of stuff it seems that it's quite likely that we will be able to get to artificial general intelligence just by using the current Hardware the next thing that I also found super interesting that most people didn't really catch was of course the thing about level three to level four and I think this is probably once again something that is really profound because this actually means Society will change faster than we initially thought well I had been telling people for a while I thought that the level two to level three jump was going to happen but then the level three to level four jump was level two to level three was going to happen quickly and then the level three to level four jump was somehow going to be much harder and require some mediumsized or larger new ideas so basically this is where Sam is referring to the fact that we are currently at level two but the level two to three jump was basically where you have level two being the reasoners which are able to solve human level problems which is kind of what 01 is and of course level three being the systems that are agents and basically what is saying that look we know exactly where we're going to be going with agent systems that can take actions so essentially stating that look we know that the jump from level two to level three was pretty nice but of course the jump from level three to level four was going to probably be a monumental jump considering the fact that we're looking for AI that can innovate level four is innovators and AI that can Aid in invention the reason this initially seems difficult because it means that AI need to be able to generate novel ideas and be able to also evaluate those ideas and additionally do research so somean basically stating that look we are going to get to level four a lot faster because there's no crazy breakthroughs that are needed to get to innovators it actually means that the 01 series of models are a lot smarter than we initially thought if these systems are already able to show us that future versions of them are going to be able to Aid in Innovation that means that getting to innovators is truly going to change things quicker than we initially thought now of course I was looking around for some research and I found that there was also this research paper that basically looked at automating a research and this is coming out of Stanford University and they basically said that look automating e research is exciting but can llms actually produce novel expert level research ideas after a year-long study we obtained the first statistically significant conclusion llm generated ideas are more novel than ideas written by human expert researchers and of course there was also AI research that was basically from Sakana AI where you had automated AI research done so it's quite likely that open AI may have built on top of that too now the crazy thing about this is that getting more out of the models is not something that is new news s when in this clip basically says that look if we manage to have some kind of framework around A1 we actually can get even more out of these models and that demo and a few others have convinced me that you can get a huge amount of innovation just by using these current models and really creative ways I think things are going to go a lot faster than people are appreciate so that you have it seems that and what samman is basically referring to right here basically say look you can get a huge amount of innovation just by working with the models and of course doing different things to them is basically where you have different prompting strategies and you have different ways to interact with the model than we initially thought remember how when chat gbt was first released everyone would just input a prompt get a response and that's how you interact with the model it wasn't until that people started interacting with the model in different ways that we got major improvements for example if you look at the GSM 8K benchmarks underneath the filters area you can see that there are many different ways that you can actually evaluate the models one of those is majority voting where you basically take like 10 different responses and then the model votes which one it thinks is the best of course you've got Chain of Thought which is where you ask the model to think step by step there are other things like self-consistency and basically what I'm trying to highlight here guys is that interacting with the model where you just simply input one piece of dialogue in and then you get a response out that is isn't how realistically you're going

Segment 4 (15:00 - 17:00)

to push the frontier forward you're always going to have to change how you interact with the model that way you can figure out what is the most efficient way to get the smartest response and interestingly enough Sam aln has also spoke about how O2 gets 105% on GPA which is just simply unheard of and I do wonder and this is still playing on my mind that is this due to the fact that there was another recent breakthrough at open AI what was one thing that surprised you in the last month it's a research result I can't talk about but it is breathtaking leg apparently the research result is breathtak L good now samman also said something in a different blog post that he actually responded to here one of the craziest things that samman actually said was that it is possible that we will have super intelligence in a few thousand days it may take longer but he's confident that we'll get there now super intelligence in a few thousand days is basically a lot more impactful than AGI is basically humans at scale but super intelligence is basically where you have a god level AI that is able to do things to us that we would basically deem magic in the essay you actually say a really big thing which is ASI super intelligence is actually thousands of days away maybe I mean that's our hope our guess whatever uh but that's a very wild statement yeah um tell us about it I mean that's big that is really big I can see a path where the work we are doing just keeps compounding and the rate of progress we've made over the last 3 years continuous for the next three or six or 9 or whatever um you know 9 years would be like 3500 days or whatever if we can keep this rate of improvement or even increase it like that system will be quite capable of doing a lot of things I think already uh even a system like 01 is capable of doing like quite a lot of things from just like a raw cognitive IQ on a closed end well- defined task in a certain area I'm like one is like a very smart thing and I think we're nowhere near the limit of progress so overall it seems like 2025 is leading up to be one of the biggest years in AI that we may have seen and it seems like open AI are once again very much ahead of the larger pack

Другие видео автора — TheAIGRID

Ctrl+V

Экстракт Знаний в Telegram

Экстракты и дистилляты из лучших YouTube-каналов — сразу после публикации.

Подписаться

Дайджест Экстрактов

Лучшие методички за неделю — каждый понедельник