Dmitri Dolgov is Co-CEO of Waymo. He joined Google’s self-driving car project in 2009 as one of its first engineers, and was repeatedly promoted until he took it over in 2021. Waymo is Google’s most successful moonshot and now provides over 500,000 fully autonomous rides each week. Cheers, by the way. Yeah, cheers. You grew up in Russia, right? I grew up in Russia. It was actually the Soviet Union. Right, exactly. My dad is a physicist. The Soviet Union started falling apart, and then he had a visiting position in University, in Kyoto University for a year. We moved there as a family. Then he went to Berkeley, and I tagged along. Then I graduated from high school. I was thinking about the next thing I wanted to do. I really like that technical school in Russia. The Russians are serious about their physics. They are. I went back to Russia, and I got my bachelor's and master's there. What year was this that you went back to Russia? 1994. Okay. That was almost peak Russian optimism in the sense where it was opening up. It was. I actually remember talking to my mom about it. Of course, my parents grew up in the Soviet Union. They've seen it. They were born right before the war, and then they saw… They lived through some really tough times. I remember talking to my mom… In fact, I got my green card here in the US before I went back, and she insisted that I do it. Actually, at that time, I wasn't thinking of coming back. But then I was pretty excited about where Russia is and the trajectory it's on. Being young and naive, there's no turning back. Why did you decide to come back? There's more of a—Yeah, I know. It was pretty clear to me. I wanted to continue studying math and computer science. While the undergrad and master's that I got in physics and applied math, I think was still an incredibly strong foundational school of Russian math and science. Graduate school, it was very clear to me that the best way to do it was in the US, so I came back. I'm struck by the founders of the two most valuable UK companies are Russian math nerds who both went to the same school, Nikolay at Revolut and Alex Gerko at XTX. It's a strong diaspora. There's a company, not far from here, where one of the founders also has a similar pedigree. A company that we're closely related to. Exactly. You know the classic engineering interview question of "What happens when I type google. com and hit enter? " As in, talk me through whatever you like, HTTP, DNS, and BG. You can go down to whatever level of the stack you want. Do you want to maybe just describe, when I take a ride in a Waymo today, what's happening at a technical level? What is the architecture? Let me answer your question. It was happening in real-time, but this is going to be only a part of the story because we're going to be talking about the inference, the real-time inference part of it. If we want to have a deeper, richer, technical conversation, I think it would be interesting also to zoom out and talk about the entire ecosystem of what goes into building, evaluating, and deploying the Waymo Driver. But when you're driving around or being driven around, we think about what we're building as a driver. Obviously, it's not a car. It has a number of sensors that are positioned around the vehicle. We use three different sensing modalities. There are cameras, there's LiDARs or lasers, and there are radars. Those are the primary ones. There are also microphones, directional microphone arrays, but those are the primary three for sensing the world. They all have very nicely complementary physical properties. They all have 360-degree coverage around the vehicle, so the Waymo Driver sees 360 all the time. All of the data goes into a computer, you would expect. The software that process, now it's all AI, specialized AI in the physical world. It processes the sensor data. Nowadays, we talk about it using AI terminology as encoders that take this data in. Then there's the decoder, the action—the generative part, if you will—in the car. The generative task there is to figure out how to drive. That is, of course, connected through a specialized interface to the car where we can actuate the vehicle.
Segment 2 (05:00 - 10:00)
That's why you see the steering wheel turn, and it drives you around. Okay, so I get into my car, there are three main families of sensors, LiDAR, radar, and cameras. It is using that to first build a model of what's going on in the world, where are all the other cars and things like that, then you say make decisions, and then actuate that with the car. That is the system that you're living in. Is all that inference done locally or presumably, yes, nothing's in the cloud? Nothing real-time? Nothing real-time in the cloud. There are some things that can happen in the cloud, but they're not required. Got it. What's an example of a nice-to-have that happens in the cloud? You can imagine a situation where we do… Some of it is not directly related to the task of driving, but let's say after you leave the car, we want to check that the car is not dirty. You didn't leave anything there. If you did leave an item, well, if you left it a mess, then we want to send the car to one of our depots, get it cleaned up. If you left an item there, maybe your phone, we want to detect that, and then send it to our listening phone and let you know. That we do by asking a model that actually lives off-board as opposed to having to put it on the car because it's not a real-time task related to the driving. That's one example of something that— There are all these debates that go on Twitter around self-driving. I can think of end-to the end versus the more modular approach. There's cameras only versus array of sensors. I can't tell, are these debates actually interesting to an expert in the field, or do you think these are just settled matters, and they're just grist for the algorithm? I understand where the questions are coming from. I do find that often the way they're posed and the way the debate happens is losing a lot of the nuance and a lot of detail that really matters. To me, the most interesting technical questions are in that level. Because the way we think about building the Waymo Driver, it starts with a large off-board foundation model. I can imagine building a big model that understands how the physical world works and understands the important properties of what it means to drive, the social aspects of driving, and what it means to be a good driver as opposed to a bad one. That's the foundation. Then we specialize it into, let me call it, three main off-board teachers. There are still large, high-capacity off-board models. There's the Waymo Driver, there is the simulator, and then there's the critic. Those then get distilled into smaller models that you can run inference on faster. The Waymo Driver becomes the backbone, the backbone of what's in the car. The simulator, of course, is what powers our synthetic generative environment that can run on the cloud for training and for evaluation enclosed the system. The critic is the value. Does the simulator ever run locally? No, it doesn't. However, what I think is interesting, in a way, the way the decoder works, the way the model works. If you think about the generative task in the simulator of creating those realistic worlds and how other people behave, how cars, pedestrians, cyclists, and the task that you have to solve on the car in real-time, there is this fundamental shared capability of understanding how these objects relate to each other and predicting what they might do in the future if you are running on car and then generating, some sampling, those probabilistic behaviors in the simulator. It's different models, but this is why the shared foundation model is able to power both. Similarly, if you think about the critic, the job of the critic is to find interesting events and then be opinionated about what's good behavior and what's bad behavior. Similar fundamental understanding. If you're running inference on the car, you still have to figure out which of the multiple hypotheses of these future worlds you want to take action to steer towards. Okay, and these are all downstream of the same foundation model? That's right. You start with the foundation model, then you specialize and fine-tune, still off-board model. Those are the teachers, and then you distill. Each one of the teachers trains its own student, the driver, the simulator, the critic. You started working on self-driving 20 years ago.
Segment 3 (10:00 - 15:00)
As you think about the tech evolution, is this just a scaling laws story where we had to be able to throw enough compute at it? Were there architectural approaches we needed to wait to have be invented? Was it just a story of we needed 20 years of going down the wrong cul-de-sacs before we eventually arrived at the right approach? Knowing what you know now, could you have a successful Waymo in market in 2015, or was there some enabling technology? No. Technology breakthroughs that happened over the years were critically important, primarily in AI, but also in other areas. Compute. They have a compute... Now, I wouldn't characterize it as going a thousand different dead ends and then having to retract and then finding the one right path, I would characterize it as iterative learning and evolution, and then transformers came around, but Transformers, for example, are very general architecture, powers of LLMs, powers of our models. But how you apply them to that space, I think this is where— It didn't just fall out of Transformers. Exactly. Then, of course, people like to talk about architectures, but architecture is important, but really a lot of it comes down primarily to your metrics, to your evaluation mechanisms, to all of the training recipes, and of course, data. LLMs are good at text or tokens, specifically, and obviously perform best at domains that have some single corpus of text they can work on, like coding, where it's very helpful that everything was just textual already. Part of the success has been creating textual representations for domains such that we can then put a lens against them. Can you describe how you encode the world that you're seeing? Are you just building a 3D bit map, essentially? This is where I think we get a bit into this question of what is the interface between the encoder and the decoder parts. I think that touches also on the thing you flagged earlier where people like to debate end-to-end or not end-to-end. Let's talk a little bit about end-to-end and then get back to what is the interface between those two. When we say end-to-end, what do we mean? We mean that it is some large ML model. Typically, you don't build them monolithically. You have different parts and different subgroups. But what's important is that you can propagate/back prop the gradient and the loss function all through the different layers. Every layer, you can learn the weights and the representations that matter for the final task. You don't force it through some narrow funnel between, let's say, the encoder and the decoder. I think of a simple view of end-to-end being pixels go in, and car actions come out, which may be a bit of an oversimplification. That's exactly right. This is the basic vanilla version of it. If you think about what will it take to build the driver that's capable of fully autonomous operations. You think about this entire ecosystem of the driver, the simulator, the critic. If that's all you do—pixels in, trajectories out—it becomes very difficult to do all of those three and achieve the high level of safety and performance that we require, and it becomes very difficult to do it at scale. However, it's a very easy way to get started. You collect some data… Kind of like the LLM world. The easiest thing you can do is pick a model. The easiest way to get started nowadays would be just take a VLM. It already has a language-aligned camera encoder, and then it has a decoder that can predict, generate text, and you can fine-tune it and say, "Instead of text, generate trajectories. " Very doable. In fact, a while ago, we published a paper called EMMA, that did exactly that. It will actually, in the nominal case, drive pretty darn well, which is mind-blowingly impressive. That is very funny. There's something to it— You're saying you can take an off-the-shelf model, which has nothing to do with driving to start with, and you'll get these good results. That's right. In the normal case. I just want to be clear. It's orders of magnitude away from what you need to—
Segment 4 (15:00 - 20:00)
Yeah, you should not try it on the streets, but it works. But for example, if you— It's like a talking horse. It's impressive that it's talking. Exactly. You can actually… The product that you wanted to build was maybe a driver system, not a fully autonomous system, then maybe that's all you need to do. Then for that, you don't need all this other machinery of the simulator and the critic because the number of nines is drastically lower. This is interesting because there is some intuition behind why that works. If you think about the hard parts of driving, it's not unlike having a conversation, except if in the LLM world, you're modeling a language, or may be modeling a dialogue in the space of sentences and words. What makes driving hard is also this multi-agent social interactive part of it. If I do something that's going to affect you, it's going to affect somebody else, and the history matters—it's not local and just geometric—context matters, semantics matters. But it's in a different… It's not in the language of words, it's body language, if you will. We see that empirically validated if you do this approach. Let's say we build this thing, just cameras, camera encoder, pixels go in, trajectory go out. The quality is sufficient to drive. In the normal case, it's not sufficient to deal with the long tail of the edge cases and hit the high bar of superhuman safety that we require. Then you start asking the question, what else do you need? If all you did was observing how other people drive when you trained this system, maybe observing just passively how people drive and how they interact, maybe also driving the car yourself and then using imitative learning to train it. You find that that's not enough. You have to do something in closed loop. things like RLFT, which is also parallel to what we see also. RLFT? RLFT. Reinforcement Learning-based Fine-Tuning. Okay. Similar to the Reinforcement Learning with Human Feedback in the LLM world, right? You want to do maybe proper closed loop driving where you explore all kinds of different situations, and then you give it a reward signal to keep it in distribution. For that, then you need a realistic simulator. If you want to have a good RL system, you need to have an opinion for the reward function, this is where the critic comes in. If you have a purely end-to-end system, let's look at the simulator. What do you do? You're then constrained to just go from pixels to trajectory. That's all you can run the system on, right? It's a very high dimensional space. It's a hard problem to generate everything. But even if you solve that, it just becomes incredibly inefficient to run it in the full way of pixels to trajectories and simulation for training or for evaluation. This is when intermediate representations come in. There are some intermediate representations in the world in this task, in the physical world, we know are correct. They are not sufficient, but they're not generality limiting. There's an object here, there's a concept of a road, there're signs, there's speed limits. This is where augmenting that learned representation, those learned embeddings from the encoder decoder with that more structured representation is what we do. We find that this gives us additional knobs to simulate in that space, just pixels to trajectories. It allows us to have additional safety validation layers in real-time. It also gives us additional mechanisms to specify the reward function for evaluation of the critic or for training. This is, again, we've gone full circle of it. Is it end-to-end? Yes, it is. Then, if you want to do it at scale for full autonomy, it's augmented with all of this other stuff. That's very interesting on the simulating point. It's just very hard to simulate for an end-to-end model because it's easier to deal in intermediate representations rather than coming up with a pixel-perfect view of the world. You need both. Having end-to-end architecture that's augmented with that structure allows you to play in both of those worlds. Yeah, yeah. What are you looking to do as a self-driving car? It sounds funny, but I think people maybe don't realize that there are many different things that you're looking to solve for, where you're looking to get the person to their destination, them there
Segment 5 (20:00 - 25:00)
reasonably promptly, but also drive quite smoothly, and also have many lines of safety, and also not annoy other drivers and get honked at, and... What are some of the reward functions or things you're optimizing for that maybe are not obvious to people? Safety is the primary focus. But of course, we also want to be a smooth driver for both people in the car and other actors. We also want it to be a predictable, well-behaved one so that it can nicely fit into the whole social ecosystem of our roadways. It seems like one of the issues that has quickly emerged with self-driving is the fact that people can't have nice things or not everyone is nice to the robots. Whether you're driving through a dodgy area or getting blocked, or maybe I'm not going to drop you off here. Maybe I'm going to go around the block and drop you somewhere better, but all of these, as you say, other human issues. How do you go about solving those? A lot of the ones that you mentioned are just things that we need to work on and understanding. Honestly, as I said, that if we're not dropping you off exactly where you wanted to be dropped off, or we don't give you a good interface to tell us, that's on us. We got to make it better. It feels like the drop-off is actually a pretty nuanced part of the self-driving journey. The highway stuff and the 35-mile-an-hour roads, that is all nailed, but there's just a lot of nuance in the drop-off experience. I'd say they're all hard. You picked freeways, and you picked drop-off for different reasons. For drop-offs, you're absolutely right. There are a few things that are maybe not obvious. You just think about this problem, but it's understanding where you want to go and making it as convenient as possible for you. Pickups from drop, it's not exactly symmetric. But then also understanding the context of the situation, where do you stop? You don't want to block a driveway. double park, although in some cases where if it's a quick one, maybe it's okay. There's a lot of nuance that goes into doing that well so that it's smooth, frictionless experience for the rider as well as other folks. Freeways, for most of the time, not much happens. They're very well-structured because we designed them that way. But there is still that long tail of really complicated stuff that happens where the consequences of a bad event are much more severe. The speed is much higher. Everything is quadratic in speed. But we see a lot of stuff. Imagine grills falling off of freeways. Imagine people getting into accidents and spinning out of control. You see one of those flatbed trucks with just a bunch of stuff piled in it, and you're driving behind it? I don't know. I always find it very nerve-wracking. I know. We've seen them leave a trail. It's a different set of problems. But I feel like the general sentiment with the Waymo is that the driving has mostly now been solved by you guys, and it's a question of scaling up and maybe some super long-tail stuff, really snowy condition. Is that your sense internally, or is there actually much more nuance within that? I would say it's not like we're done with engineering. I would say that we've clearly moved past the stage of scientific research and deep core technology development to this new phase of accelerated global scaling and deployment. We still have work to do, but I don't see today any limitations or any gaps in the core technology. The driving is good enough now. Well, the core technology, I think, is good enough that I can't think of any aspect of driving that is not supported by the fundamental technology. Now, that said, there is a lot of work to do in specialization and in validation before we can deploy responsibly. We're not driving everywhere in the world. We are planning to start operating in London and in Tokyo this year. Do we have a driver that you're using today in San Francisco that we can just plop down in London and go? No. But what we're seeing is incredibly encouraging from the perspective of is the core technology there?
Segment 6 (25:00 - 30:00)
Now it's a matter of collecting the data, doing some specialization and validation. Science are different in both of those places. People drive on the other side of the road, but that's actually not that hard for computers. The core technology generalizes really well, but there's still work that you have to do. What generalizes least well? Increasingly, we're finding, especially now that we're able to hook the Waymo AI to the AI in the digital world, the VLMs, and inherit the general world knowledge from VLMs, we're seeing really strong results from zero-shot or few-shot learning because of that general knowledge that we bring in. But there are a few things like, say, cold weather, cold winter weather, where it affects the entire stack. It's not just the AI, we actually have to— The hardware. You need to have the proper cleaning solution, heating elements in it, and then you think about things that are completely solvable by computers, like motion control and slippery surfaces. That takes a bunch of work. You don't get that for free from just pulling it, some VLM decoder. Was it the case… My impression, not knowing anything is that in the early days, there was maybe a lot of San Francisco-specific work or Phoenix-specific work in the early markets, whether it be mapping or something else and that you guys seem to either have solved that in generalizing it or just scaled up your ability to do the city-specific work. What enabled the rapid-city expansion? We usually think about it, the capability of the Waymo Driver as well as deployment, not primarily and directly in that space of cities or zip codes. I think about the operating domain. Then the freeways, cold weather, snow, rain, fog, density, et cetera. That's what we are building. That's where we're evaluating, and then that maps to a particular city, be it within the operating domain or outside of it. If we rewind history a little bit, our initial deployment in where we started offering a fully autonomous commercial service for the first time was in 2020 in Chandler, Arizona. That was on what we called the fourth generation of the Waymo Driver. This was, if you remember, the Pacifica minivans with different hardware, different software. There, we were super focused on doing the whole thing end-to-end: learn how to build the driver, evaluate it, deploy regularly, operate it end-to-end 24/7 with customers, learn from the customers. Then we were very focused on that operating domain of mostly Chandler, which is a medium, low-complexity one. Then, when we made the jump to the fifth generation of our system, this is what's on the highways today, we really wanted to take a huge bite out of that operating domain. We collected data all over the United States, all different states, different cities. When we chose to deploy in the hardest parts of San Francisco, hardest parts of Phoenix, we made a big jump on the hardware side, and most importantly, on the software, the AI side. I would say that was the big discontinuous jump. That's what you're seeing now after we've scaled up and iterated all of the aspects of building and deploying the driver. This is now why you're seeing us go in parallel and scaling in the US and globally. So driver v5 was just a much more generalizable stack than v4? What was it about it? Was it just that it had been trained on a much wider data set? It was when we made this big bet on AI. There was a lot more little AI models and ML models in the fourth generation. We made a much bigger bet and jump to AI is the backbone for the fifth generation. as the core engine, as in you're saying that Gen 4 had lots of small little AI subsystems? Yes. We made that jump, and we've been iterating and improving the model since then. As we're seeing with Waymo rolling out widespread autonomy, it has second-order changes on the entire system. In this case, traffic patterns or other drivers' behavior, or eventually, how cities are laid out. Autonomous systems are coming in many domains. In commerce, soon agents are going to be transacting without human intervention. We're basically getting driverless commerce.
Segment 7 (30:00 - 35:00)
Stripe is building the economic infrastructure for AI. As part of that, we're letting payments be initiated by humans or by agents. If you want to sell to agents or let your agent spend money all around the web, check out Stripe's Agentic Commerce Suite. Can we talk about hardware a second? Lots of hardware questions, but one is maybe, everyone in this space has a very charismatic demo of a vehicle that is custom-made for self-driving. It's often the van with no steering wheel, seats facing in both directions. You guys have one. Tesla has the steering wheel-less Cybercab. Cruise had the Cruise Origin. Yet, we're still driving in Jaguars that have a steering wheel in the front and are pretty similar to consumer cars. It's interesting to me because if we were talking about this 10 years ago, we might say, "Well, yes, developing a custom car, that's relatively straightforward. We know how to put a bunch of sensors on a new car. " But the software will take a long time. What's interesting is we've made huge progress in the software, but interestingly, the cars are still derivatives of cars that people are driving. I'm curious why you just think the custom hardware has not happened as of 2026. It's obviously a small improvement compared to Waymo is the big improvement, but it's just interesting that it still hasn't happened. Well, let's say our sixth generation of the vehicle and the driver is our version of that. Oh, no, I know it is. It is the Ojai platform. That still has the… We can talk about whether you want to have the seats pointed backwards or not. I actually think it looks nice in a demo, but practically speaking, it's maybe not the way to go. But it is a custom-designed vehicle, and we put a lot of thought into moving away from a car that's designed around the driver to passenger. It's much more spacious. It's happening. It's not open to the public yet, but I took a ride in it the other day, fully autonomously, and that's coming this year. Yes. How much better is it as a passenger experience? You'll tell me once you give it a try. I love it. It's all about the space and the convenience of ingress and egress and the screens and the interface of the passenger. We put a lot of thought into every aspect of it. It has sliding doors. It's very easy to get in. It has a flat floor. If you sit in the back, you can fully stretch out, and there's so much space there. From the outside, it looks fairly big. But the actual footprint of that is barely larger than the I-PACE. It's amazing that you walk in, and it feels like you're in a living room. Yes. I guess my question is just, Waymo does 25 million rides a year, run rate-ish, with the Jaguar I-PACE. It's interesting that so much scaling has happened with self-driving so far on the old retrofit. Maybe that's to be expected. Well, it matches the high… I don't think it's a given. You're right. But if you think about the value proposition, of course, there is the safety of it. You don't have to worry about it. There's also the privacy, being in the car by yourself, maybe with other folks, but not having to share that space with another human. No, Waymo is a great product. But I guess this is why we're seeing such consistency in car. It drives well, very predictable. You can go beyond that. You can specialize even more to make the experience even more magical around the rider. But I guess it would have been disappointing without the specialized car, and I think I would have been surprised if we leveled off at some other much lower level of customer adoption because a car seems like more of an optimization improvement, but the core of the value proposition comes from those other factors. Yes. I guess it just take risk on one thing at a time. We'll start by doing the software layer, and then we'll build a specialized car or something like that. That's right. It's also, as you said, it's a big investment. You have to de-risk the fundamentals. Throughout our history, we were very focused on setting the biggest goal for the company to de-risk the most important questions. We talked about the third generation where we wanted to deploy
Segment 8 (35:00 - 40:00)
something and go end-to-end. We talked about what was the goal with the fourth generation, sorry, the fifth generation, and then there's the sixth generation. That was the sixth generation where it made sense to spend all this effort into the custom vehicle. The sixth generation is both the custom vehicle. Is it also a new generation of the driving stack? It is the new hardware. The sensors, the hardware, the self-reliant hardware they're putting on the Ojai vehicle is the sixth generation. It is very different from the fifth generation. It is simpler. It is more capable. It is much lower cost. Just thinking of a fraction of the cost. It's comparable to what you would get like a fancy ADAS system nowadays, the driver-assist system. The software is pretty much the same. When we talk about generalizability of the Waymo Driver, we talk about weather conditions, we talk about cities, but it also generalizes well to different vehicle platforms and different sensor configurations. Okay, so Gen 6 is a new vehicle and a new sensor stack, but it's almost a TikTok cycle happening here. It's a similar software. That's right. Then we're going to put the sixth generation Waymo Driver on other vehicle platforms like the Hyundai Ionic that's coming later in the year. What is different about the sixth generation hardware stack, and how did you make it cheaper? It still has the same three sensing modalities, but we've made significant optimizations in all three. Unification, simplification, and there's just the… Just writing the— Yes, is it a classic case of manufacturing scale where we're not even— Well, scale hasn't fully come in place. But all of those, if you think about the supply chains, the industries, cameras is pretty mature. Radars, way many years ago, used to be bulky, complex, very expensive when we were putting them on planes. But then we started putting them on cars, now you can get a decent automotive radar for tens of dollars. There is a variant of the automotive radar. It's called the imaging radar. It gives you a richer… That also has come down in cost drastically, but it's a little bit behind your standard automotive radars. LiDARs are following the same very predictable, very well-known trend. We're writing that, and we're also learning from the previous generation to just make improvements and simplifications and optimizations. It's a very silly question. What are LiDARs versus radars better at in a self-driving context? LiDARs—Are they complementary? They're very complementary. It's all blasting. Echolocation. Effectively, like blasting photos out there, and then they bounce off of something, they come back, you measure what comes back. The frequencies are very different. The laser gives you… Its very high resolution. You can think of it as like a laser beam that goes out, spins around. It shoots out millions of these laser pulses per second, and then each one comes back, and you're sampling the 3D structure of the world with very high resolution. LiDAR for very fine-grained mapping. That's right. Radar has much lower resolution, but because of the physics of it, it degrades much better in adverse weather conditions. So fog, snow, heavy rain. It could be included by particles between it and the target. Imagine driving in super dense fog. Yes. We're close to San Francisco, so we probably don't have the think that hard. It can be really hard to see. So cameras degrade. Laser, depending on the size of the particulates, can degrade better or worse than camera. Radar is not well-affected. So you can imagine driving on a freeway, then radar will give you really good returns for cars that are absolutely invisible in the camera space. That's interesting. Does that mean there are some environments where you'll be relying significantly more on radar? Where the performance is good enough? Well, it's a combination of the sensors. We rely on… Each one is noisy. How the noise characteristic show up in different environments is different, but it's not like we switch from one to another. estimate what's happening with the world through cameras and through radars and through LiDAR, and then we compare. No, they're like… There's an encoder for camera, LiDAR. There's an encoder… They all go into the system that gives you, jointly, the best view of what's happening in the world around.
Segment 9 (40:00 - 45:00)
If it's a nice, bright, sunny day, cameras are very valuable. If it's pitch dark, or you have sun in your face, or you're blinded by the headlights from an oncoming car, then the camera will degrade. There's still some noisy signal, but it will degrade, and LiDAR is completely unaffected. Are there technical problems that are your white whale, or you're still chasing, or you are particularly interested in solving? Even if they're niche for the… We really want to have driving when it's actually snowing nailed or steep hills in San Francisco. Are there problems you've been very interested in historically or still are? I'm super excited right now about the accelerating global expansion, more cities in the United States and going internationally. I understand I'm not answering your question about the technology, I'll come back to that. But really, that's the thing that I'm today most excited about. Just getting to a place where any major metropolitan area, you can fly into the airport, and then take a Waymo and go anywhere you want to go. That is insanely exciting to me right now. Then technically, what I'm most excited about is all of the rapid progress in AI, and the world models, the foundational model work. It is just such a massive boost to how much we can simplify the system, bring down the cost, and how we can scale globally. There's just some magic that happens that I don't think I would have anticipated a few years ago. That I find from the technical perspective, just insanely thrilling. When you talk about the progress in AI, what are the most fun parts of it for you these days? I think it's seeing the capability and the scaling laws from this approach of starting with that cornerstone of the foundational model, and then specializing to teachers and then distilling. You get such big wins in performance across the board. You invest something into the architecture or get a better data or training recipe, and then you invested that early stage, and then it just has massive amplification and ripple effects. That, in some ways, is kind of magical. Then you see it on the car. I've had some moments where a car does something, and you look at a log, and I've been surprised. It does things that I didn't think it was capable of doing. It's that… When you see emergent behavior, that's a proud moment? One example, yeah. When you build a system, and then you think you understand how it works, and you understand fully the limits of its capability and performance, and then it does something almost magical, it's exhilarating. One example I can give you, I think I've shared some videos of that publicly in some talks, was this example where the situation that happened in San Francisco. A fairly benign situation where at an intersection, our light is red, there's near cross traffic, a bus goes by, and it stops partially blocking the lane. Our light turns green, so we start to go. We're nudging around the bus, and then you see a pedestrian being detected on the other side of the bus. Then your car responds appropriately, it slows down, goes a little bit wider, and then a pedestrian actually emerges from the bus, and we go on our own. The first time I looked at that log, I was like, "What's going on here? " I know we have pretty darn good sensors, and the software is very capable, but we don't see through stuff. That's not how cameras or LiDARs and radars work. It saw the pedestrian through the bus? on the other side of the bus. It's not like you look at the windows, you're like, "Okay, radars shouldn't… This massive metal box. " Look at the sensor data, and radar shouldn't be able to go through it. You can't see in the camera because there're reflections and there's people on the bus. It's not like you can see through the windows. What is going on? Maybe it's noise or some coincidence. The first time I saw it, I couldn't actually believe it.
Segment 10 (45:00 - 50:00)
I was like, "No, there's something. It doesn't smell right. " What actually turned out was happening is that our peripheral LiDARs bounced under the bus, and there was just a little bit of very noisy reflection of the movement of the person's feet that was enough for the AI models. "Hey, likely there's a pedestrian there, and I'm going to… I detect it as such, and moreover, there's enough data there to predict what they're going to do. " It just blew my mind. Is this the perfect example to explain what we were talking about earlier, the value of one, fusion across a sensor suite, but then secondly, building, I mean, relatedly, building an intermediate representation of what's going on, where if you're just dealing with pixels, I mean, the person behind the bus does not exist in pixel space, and so you need to have some representation of the world that exists to be able to reason about the person behind the bus. I think it's an example where using that intermediate representation to boost the level of performance of all parts of the model is what's happening here. Just imagine solving this problem with a black box, purely open loop imitative system. It could be— Hard to impossible. Is it impossible, no, but in practice, what would it take to achieve that level of performance? It's very, very difficult. What metrics can you share on just where the business is at today in terms of rides, revenues, cars on the roads? We have about 3,000 cars on the roads. We're doing about half a million rides per week. That translates to about over 4 million fully autonomous miles per week. We are operating in a fully autonomous mode in 11 cities in the US, and 10 of those, we have riders, public riders. What's the ghost city? The ghost city is Nashville. We just started there. We just opened it up to riders in four new cities in one day. That was one of those little but super exciting moments where I thought back to the history, like how long did it take us from the first time we started fully autonomous, rider-only operation to the first time we had external riders in four cities. It was about eight years. Then the other week, we just launched four in one day. Yes. It seems now clear that in 15 years, most miles that are driven will be autonomous. There will be some burn in period, and it's lots of old cars in the road. I think it'll actually take a little while. Some of that will be by level four, level five systems expanding in new cities and that expansion continuing. Some of it will be, you referenced the existing driver assist systems and getting up to level two and level three, and existing systems across current car brands getting more and more capable. What do you think that working your way up from the lower levels versus working your way expanding from existing products like Waymo? What will that convergence look like? Because we're going to eat it from both sides. I don't believe we will. I actually think this— That's a great answer. Cars will get smarter. There's going to be advances in driver-assistance systems, and if there is at the same time from level four autonomy, there is simplification, and the sensors of today are not going to be the sensors of tomorrow, so they'll be much more integrated, they'll be simpler, there'll be much lower cost. From that perspective, there is a path of convergence. There's also a path of convergence from the product lines. There's ride hailing, and you can take a ride through the Waymo app today. Eventually, that'll be on your personal car, so that I see. You can talk about the technology, and I see it just as fundamentally two different problems. There's driver-assist systems, and then there is full autonomy. I think it's deceptive to think of them as incremental on one spectrum of complexity. Okay, but you think one cannot work one's way up from driver-assist systems to full self-driving? You think you have to start building a full self-driving system? I think you have to tackle… If I think about the hardest parts
Segment 11 (50:00 - 55:00)
of building a fully autonomous, rider-only system, they are very different from what you do for a driver-assist system. Of course, some work in this space helps you. I don't want to say you can't make the jump, but it is a qualitative jump. When can I buy a Waymo so that I don't need to wait for it when I want to go? When I'm ready, I can walk out the door and it's there. I'm not going to give it a date today, but you're not the first person to bring this up as a product request. Duly noted. I'll add it to the list. Just that waiting for the car, it should be nice just in the garage there and keep your stuff in it and everything. It's not the first time you've heard that request. It seems to me operationally very intensive and very hard. A self-driving car is actually not self-driving, it takes a village. You have all of the human operator ready to step in. There was that thundering herd incident that you guys talked about in San Francisco that highlighted that for people. Then there's just keeping the cars clean and keeping everything running in that regard. Can you describe just what the operational infrastructure that sits behind Waymo looks like? Sure. I will say that we are overall in all of those areas on a path of increasing efficiency and automation. The number of manual steps that one had to do five years ago to launch a Waymo, versus where we are today, is drastically different. But nowadays, if you look at one of our depots as a fully, automatically orchestrated dance of autonomous vehicles. The way it looks like today is cars will automatically go to pick up their riders, serve their trips. If for some reason they need to come back, maybe they're low on energy, maybe somebody left a mess in the car, they will automatically come to the depot. If it is, so cleaning today is a manual process. It'll get flagged, and the car. We have fleet management systems say, "Hey, car number 378 needs cleaning. " Actually, on the sensor dome, we're able to display icons. We'll show you like a little emoji. They'll put their hand up, yeah. There's people whose job it is to clean the car. They'll come and clean it up. If that cleaning is not required, and it's just charging, we'll also automatically pull into a charging stall, and we'll say, "Hey, I need charging. " We don't yet have automated charging. In the future, you can imagine that being fully automated, but a person will come in and plug in a cable and the car will charge, and they'll say, "Hey, now I'm ready to go. " It will get unplugged and the car will pull out of its parking stall and then go on its merry way. One of the new Porsches, I think it is, has inductive charging, just like your iPhone, where you just drive over the charging mat. I was amazed that works at car scale, but yeah, possibly in the future, they'll just be able to drive onto the charging mat, or do you think just robotic plug-in will be easier? We'll see. I don't know. I think there're some questions about efficiency and how that plays into the overall cost and which one will be most cost-beneficial. It remains to be seen, I think. How well-behaved are the Waymo riding population in terms of not leaving a mess in the car? We have wonderful riders. We have the most amazing customers in the world. Generally, I would say they are very good. I think there is something about… I talked about not having a person in the car, it's not somebody else's car. In some ways, you want to preserve the… I think generally people want to preserve the nice aspects of it. It's a broken window thing where it's so clean to begin with. I know. I think that's the general trend that we see. Because it's not somebody else's space, you're in it, it feels like it's your own. You don't want to mess up your own space. I don't want to speculate too much on the psychology thing. However, I will say that it varies. You can imagine a college town on a Saturday night, and that's a different distribution. Yes. Will I be able to get a Waymo at any address that has USPS service in the US, or will there be some head/tail dynamic
Segment 12 (55:00 - 60:00)
where Ketchikan, Alaska is just never worth it? Eventually, it will, absolutely. There's no doubt in my mind. I think it's just a matter of when and what modality would make the most commercial sense. This is for your ride-share versus privately owned. For a ride, it's not a technical problem. I mean, technology is solved. But then if you're in the middle of nowhere and there's just not enough density of the trips, does it make sense for the ride-hailing service that Waymo is running to have cars on standby? Probably not. They can be deployed somewhere else, and you probably don't want a horribly bad ETA. This is where a personally-owned vehicle that is equipped with the Waymo Driver is maybe how you will see it materialized. Relatedly, what will the second-order effects of, say, majority autonomous traffic be? It feels like a lot of things will work better where, as you say, when someone merges into a lane very poorly and everyone all the way back has to slam on the brakes, that's an antisocial behavior. It feels like higher quality and more prosocial driving will just basically reduce traffic a little bit, even for the same number of cars on the road. But presumably, there'll be other second-order effects. We'll want higher throughput traffic lights and, yeah. How else will things change? The first thing I think that you mentioned is that's a huge deal. I just need to think about traffic jams. What's that saying with the Navy SEALs? "Slow is smooth and smooth is fast. " That's what you're like, traffic jams are like, you accelerate abruptly, then you come to a stop, and sometimes you have a traffic jam, like what happened? Well, an old lady crossed the road three hours ago, and we still have the standing wave there. If everybody was a smooth, predictable driver and a consistent driver, and you would still have those traffic jams at the time of, but then the time constant should clean it out, I think, would be very different. But longer term, things like parking lots. Right now, if you look at what is our most interesting pieces of land allocated to, it's parking lots, it's garages. Why is that? Well, because, again, your car is just sitting there 90% of the time. If more cars become fully autonomous, then there's no need of that. Then imagine, just imagine what you can do with your favorite city in the world if you don't have to spend that money, that huge fraction of it on just keeping these chunks of metal sitting around. I don't think people often realize how big a deal parking minimums are for the layout of the urban landscape. The coffee shop near where I am would like to have outdoor seating but can't because it would reclaim parking spots. Yeah, wouldn't it be wonderful? I have a few more questions, but I'm curious to talk about Google's relationship with… Self-driving where, again, it feels like right now, Waymo is, aside from everything else AI-related, the most exciting thing happening at Google, but it was a very long journey to get here. I feel like you could say that Google almost started working on it too early because you were saying there's been a bunch of recent enabling technologies, and so did it require Google starting when it did so early, or could one have spun up this project in 2015, 2020? Then, how did Google keep the faith when it almost felt like it was perennially two years away? Yeah, on the latter part, I just have to give credit, huge kudos and gratitude to Larry and Sergey and Alphabet's leadership in our company. It is part of the culture and the DNA of the company is to have that vision and have the stamina and conviction to go the distance. To the other part of the question, was it too early? I don't know. I think what we've been seeing, clearly, all of the breakthroughs that we've seen over the years have changed how we're building the system. But the complexity of the problem is such that you need to go through these iterative cycles. We've seen many waves of technology. There were breakthroughs in 2013, ImageNet came around. There was this narrative, "Okay, that is the right time to start or be a self-driving company. " Then transformers came around and VLMs.
Segment 13 (60:00 - 62:00)
All of those are super powerful. You have applications in other spaces. In the digital world, they certainly have an impact on our AI in the physical world. But there are no silver bullets. They drastically reshape that early part of the curve. It's always been the nature of this problem. It's very easy to get started. It's deceptively But it is super hard to go the full distance. Edge case... It's the number of nines. There's the standard engineering rule of thumb that every next nine takes 10x more. Yeah, maybe there is a more optimal path, but I don't see that there's some magical moment where the true complexity of the problem goes away, and then you can just take some off-the-shelf components under your business. If that were the case, then I think the industry would look very different today. Last question I have. You've been promoted a lot at Google. It feels like Google really recognized your talents. Just what do you think Google does? Google is famously one of the very best in the world at technical talent and say, the current AI wave more broadly happening, is either stuff happening at Google or generally Google alumni. But just what have you observed firsthand from how Google does this so well? I would say that culture of Google of not accepting the status quo, having a big vision, and investing in technical talent—the people who can go the distance and realize the vision, that is part of the culture. I think this is what you're seeing with the breakthroughs in AI in the digital world, and like all of the early investments in transformers and other fundamental technologies, quantum computing. I guess we are not unlike those efforts as well. Dmitri, thank you. Yeah, thank you.