work on MicroHart, the most serious of all product names. Uh so arguably giving computers to uh humans was a good idea. So we're doing the same thing for AI. It's kind of like inception. We're giving computers to computers. Um so Macro is building a fully capable digital realtime very important human emulator. So it's able to do anything on a computer that a human is able to do including using advanced tools in engineering and medicine. So there should be rocket engines fully designed by AI. And in a sense it's one of the last few remaining areas where AI is significantly worse than humans which is why I think it's one of the most exciting areas to actually innovate in and actually change the field. — Hi everyone. So yeah, my name is John and uh yeah, so we're building these strong reasoning models which are now going to control our CLI. Like we're actively using these every day. They are like tremendous like productivity boost to the whole team. I know the voice team is like killing it on that. And you know this is the reason why we need the compute you know we need the large scale compute to run these models to boost our own productivity. But um you know 80 to 90 95% of the world uh world software has a GUI. Um so that's like you know great representation and you know to truly make people's lives easier uh we need to develop models that are capable of solving day-to-day tasks uh on GUI. So macro hard you know we will emulate a company where the output is digital and so this is the obvious next step for agents. uh macro hard will enable true endto-end orchestration across the desktop and it will lead to immense economic prosperity. Um so yeah, we're entering an era where we need to tackle the hardest of tech problems, but in order to solve this, we need to hire the best people. So, you know, think of the smartest people that you've worked with and put them forward for a position here. And if you can't think of anybody like go through your phone book, go for your LinkedIn. Uh you'll be surprised like how big your actual network is. And they just need three properties obviously that we want to optimize for. Are they clever? Can they solve hard problems? And the second property is are they driven? Do they have the ambition? Do they want to win? And the third is are they a nice person? — Like do you want to actually work with them? Um yeah so thank you. — Yeah the mac the macro hard project is um over time actually will uh probably be our most important uh project because uh what we're talking about is um emulation of entire human companies. So when you look at the most valuable companies in the world they are uh their output is digital. Um so they don't actually uh make hardware. So it should be possible to completely emulate uh any company that where the output is digital. Um and this will usher in an age of prosperity likes which uh we could barely imagine at this point. You need imagine to imagine it. Um so this is a big deal and this is why the words macro hard are painted on the roof of the training cluster. uh because that's what it's going to bolt. So, — it's also pretty funny. — Yeah, meant to be a joke. — It's me again. You might remember me from macro harding computer use from a long time ago, but I also actually work on core product infrastructure and API. In fact, this is what I've done for most time uh at XAI. So, anytime you use any of our products like grock. com API authentication, you go to status. x. ai. This is done by the core product infra team and uh a large portion of them actually sit in London and we work with Haime over there. So we keep the lights on at peak hour 4 p. m. every day. Uh we get paged at night when stuff goes down. Also thank you to anyone in Palto getting paged. Um there's really important work reliability security core product infrastructure. So if you actually like if you're really interested in solving difficult distributed problems with like messy data, this is the team to join. Hey everyone, my name is Diego. Um yeah, so I think one of the main bottlenecks in this next year uh for these models is going to be very high quality evals and training data. And uh one of the ways we solve that is by taking the world's foremost experts in these respective domains u bringing them here and um uh having them evaluate the model. Uh we do this for domains like medicine, finance, uh law, we have voice actors, we have video editors um who contribute daily um to making rock better. Um and uh yeah, we're going to be continuing to work on very high quality evals over the next few months. Um we have some exciting stuff in um you know, the frontier of uh useful tasks in finance and law. um you know we're trying to build evals that are are useful and training data that represents uh useful work um and not necessarily proxies of intelligence I think a lot of the open source eval do today. Um — yeah. I' I'd like to say like um we're shifting from using these sort of uh common uh internet evals which I think are actually not a real indicator of usefulness to uh having uh expert tutors in each domain. Um so every domain of uh engineering, medicine, law, whatever the case may be. Um and the the actual eval is does the expert in that arena or does our group of experts in that arena human experts uh agree that uh Grock is extremely useful um and that the results are correct? That's the that's actually the only eval that really matters. — Yeah, exactly. Um in you you'll see this in Gro 420. Um but we've made some improvements because of that type of data in truth seeeking and kind of minimizing political bias. The responses are much more cogent. Um, so yeah, that that's exciting. Uh, and we are also working on Gracipedia. So the, uh, the goal of Rockedia is to create a distillation of all human knowledge. Um, I kind of like to think of this as like a modern day version of the Library of Alexandria. Um, and in the quest to build Encyclopedia Galactica, um, which it will one day be called. Uh, we've gone from essentially having nothing to around 6 million u, um, articles. Um, for context, Wikipedia is around 7 million English articles. Uh, and, uh, yeah, we're improving on hallucination. Um uh and our our goal is essentially for rock 5 to not have to search out of the data center. Um so yeah so in the ML