Transcript for:
Rivian Autonomy & AI Roadmap

Please welcome R.J. Scarge, founder and chief executive officer of Rivian. [Applause] [Music] We are incredibly excited to host everybody here today. We're in Palo Alto, which is the hub for our autonomy and our technology development teams. AI is enabling us to create technology and customer experiences at a rate that is completely different from what we've seen in the past. If we look forward 3 or four years into the future, the rate of change is an order of magnitude greater than what we've experienced in the last 3 or four years. directly controlling our network architecture and our software platforms and our vehicles has of course created an opportunity for us to deliver amazingly rich software. But perhaps even more importantly, this is the foundation of enabling AI across our vehicles and our business. I'd like to talk about autonomy first. The field of autonomy really started about 20 years ago and about uh let's say up until the early 2020s the approach was centered on a rulesbased environment where a set of perception sensors would identify and classify objects and hand those classified and vector associated objects to a planner that was built around a human-defined rulesbased framework. A few years ago, it became clear the approach to autonomy need to shift. With innovations around transformer-based encoding and the design of large parameter models, the approach has moved to building a neural net-like understanding of how to drive instead of following a classical rules-based approach. Recognizing this massive shift in how we approach autonomy, in early 2022, we began the process of a clean sheet design to our platform. This first embodiment of this work was in our Gen 2 R1 vehicles which we launched in mid 2024. With this updated P platform, our Gen 2 vehicles now have 55 megapixels of cameras, five radars, and run on an inference platform that was a 10x improvement over our Gen 1 vehicles. This platform was designed around an AIcentric approach. And with the deployment of our Gen 2 R1s, we began the process of our DA building our data flywheel to grow and build our large driving model. Now, because this AIcentric approach represents a model trained end to end through the millions and millions of miles driven on our vehicles, enhancing the perception platform or improving the compute is accretive to the capabilities of the model. meaning the model only continues to get better as the perception and as the compute platform improve. So you can sort of think of it like this. Um, if you learn to drive and you had bad vision and suddenly you're able to put on glasses and see much better and then go even further and supplement that with new perception modalities of radar and LAR and then your compute here metaphorically your brain was expanding in capability by an order of magnitude. You wouldn't forget the things you'd learned, the rules of the road, how to operate a vehicle. But your ability to understand nuance, to respond to complex situations, and perceive the world in environments with poor or limited visibility would improve dramatically. Our approach to building self-driving is really designed around this data flywheel. where deployed fleet has a carefully designed data policy that allows us to identify important and interesting events that we can use to train our large model offline before distilling it the model back down into the vehicle. Now, while our R1 gen two vehicles and the demos some of you are going to be trying today are using our Gen two sensor set and of course the associated compute platform, over the last few years, we've also been developing our Gen 3 substantially enhance enhanced platform and this will under underpin a massive leap forward with R2. Starting in late 2026, our Gen 3 autonomy platform will include 65 megapixels of cameras, a robust radar array and a front-facing long-range LAR. And at the core of this platform, our first in-house Rivian Autonomy processor. [Music] Now, the first iteration of our in-house inference platform includes a neural engine with 800 tops. It's optimized to support cameracentric AI in the physical world and enables a dramatic expansion of Rivian's autonomy capabilities. When integrated into what we call our Gen 3 autonomy compute platform, which video will go through later in detail, it will deliver 1,600 tops. The effectiveness and efficiency of our in-house processor has been a core focus in its development. Our Gen 3 computer is capable of processing 5 billion pixels per second. Now, you've heard me say this a couple of times. We've designed this entire architecture around an AIcentric approach where the data flywheel of our deployed fleet helps the model get better and better through reinforcement learning. Not only does this sensor set enable a much higher ceiling than what we have in our vehicles today, it also makes the platform much better to serve in building our model. We're going to continue to see improvements on our platform. Later this month, we'll be issuing an overtheair update to our R1 Gen 2 uh customers, and that will dramatically expand the existing hands-free capability, going from less than 150,000 miles of roads to more than 3 and a half million miles of roads in North America. And and again, this is just a step in a series of steps. Starting in 2026, we'll begin rolling out point-to-point capabilities in which the vehicle can drive address to address. And what that means is you can get into the vehicle at your house, plug in the address to where you're going, and the vehicle will completely drive you there. Now, some of you are going to experience this today in our R1 vehicles, and of course, this will be embedded in R2. Now, as I described, the Gen 3 hardware architecture launching in 2026 expands the ceiling of what we can achieve. The next major step beyond pointto-oint will be eyes off. Meaning, you can navigate pointtooint with your hands off the wheel, but importantly, your eyes off the road. And this gives you your time back. You can be on your phone or reading a book, no longer needing to be actively involved in the operation of the vehicle. And following eyes off, the next major step will be personal level four and and with this the vehicle will operate entirely on on its own. This means it can drop the kids off uh at school. It can pick you up from the airport. It allows it to to really be integrated into your life in ways that we can't even imagine and haven't seen yet today. Now, while our initial focus will be on personally owned vehicles, which today represent a vast majority of the miles driven United States, this also enables us to pursue opportunities in the ride share space. Now, beyond self-driving, we've also created what we call the Rivian unified intelligence. This AI backbone exists across our vehicle and across our entire business. And we've talked for a long time about softwaredefined vehicles which which are really the foundational building block for an AIdefined vehicle where every part of the the vehicle the experience and everything that's happening across across the vehicle is designed around AI from our Rivian assistant to enabling our direct to consumer sales and service model as well as our future manufacturing infrastructure. Now, as I said at the start, we could not be more excited about what we're building and we have a lot of details to show you here today. And with that, I'm excited to introduce Vidia to talk about our hardware platform. Thank you. Thank you. My name is Vidya Raja Gopalan and I lead the electrical hardware team here at Rivian. My team is responsible for electrical content in the vehicle ranging from our in-house 5nometer silicon that you just heard about that operates at voltages below a volt to the power electronics for electric motors that operate at 400 volts and lots of things in between. One common thread that runs across all these designs beyond the fact that it involves the transport of electrons is a Rivian ethos of vertical integration. At Rivian, we have chosen to vertically integrate critical pieces of technology that differentiate our that allow us to differentiate ourselves over time. We started this journey as a startup when we consciously chose to build our ECUs inhouse. Last year at investor day, we shared how this journey helped us to get to an in-house developed zonal architecture far ahead of other OEMs. Today, I'm here to talk to you about our autonomy hardware system, which is similarly very vertically integrated. As RJ shared earlier today, we will be launching our Gen 3 autonomy system late next year on the R2 vehicle platform. The hardware enabling it focuses on three main areas of leadership. Sensors, compute, and the overall product integration. Well, let's start with our sensors. At Rivian, we have a multimodel sensor strategy that provides a rich and diverse set of data for our AI models to operate on. On the screen behind me, you can see the feed from all our sensors on an R2 vehicle. On an R2 platform, much like the R1 before, we have 11 cameras providing a total of 65 megapixels of data. That is 10 megapixels over and above what we had in R1. The cameras pro provide an extremely rich set of two-dimensional data and help us see the world around us. But cameras alone have some shortcomings. They do not perform well under non ideal lighting conditions. This could be low light, excessive light, and fog. And so, much like the R1 platform, again, we still carry five radars. One front-facing imaging radar and four corner radars. By using radio frequencies, radars are able to see in total darkness while also providing the depth and velocity of objects in their path. Our corner radars on R2 are further improved. They support a dual r dual mode of operation short range and long range. In short range mode, they have very high spatial resolution which helps us delete the ultrasonic sensors in R2. That's right. We add sensors, but we also delete them when it makes sense. And now, for the first time in R2, we're adding a third sensor, the LAR. The LAR Thank you. The LAR is an optical sensor, but its strength comes from the fact that unlike the camera, the LAR has an active light source, enabling it to see it much better in the dark. Another advantage of the LAR is that it can provide a three-dimensional view of the world unlike cameras which we know provide a two-dimensional view requiring the AI models to infer the depth which they do but with a lot less accuracy. So in summary, camera is the main workhorse of our sensor sensor suite, generating the bulk of the data fed to the models, but the radar and LAR are critical to addressing the edge cases which would otherwise create the long tailor problem cases. So why do we choose to use introduce LAR? Now it turns out that there are three main factors that make this the right moment to incorporate LAR. They are cost, resolution, and size. About 10 years ago, lidars used to cost in the tens of thousands of dollars. Today, you can get a very good LAR for several hundreds of dollars. The resolution of LARS has similarly improved tremendously. The picture on the left here is a data product of a LAR from 2016, whereas the picture on the right, it shows the data actually from our R2 lighter, the one we use on our vehicle. As you can see, the R2 LAR data is much much richer. Today's automotive LAR have point cloud densities of the order of five million points per second, which is about 25 times better than what we could get 10 years ago. Finally, today's LAR mechanical spinning beasts of the past. The LAR of today is more compact and more easily integrated into a vehicle. So, let us take a look at the LAR integration in an R2. Here you go. This is one of our prototype R2 vehicles. From afar, it looks the same as the R2 many of you have seen and come to love. But if you zoom in closer and and look up front, you can see the lighter. What you see is a seamless integration and no signs of the unsightly taxi cab style bump or the tiara structure that is more commonly associated with LAR integrations. Our studio and design teams work very closely with the supplier to shape the face of the LAR in such a way so that it blends in beautifully with the R2. And now if you zoom out and look at the vehicle from the side, voila, you not even know it had a lighter. So, and by the way, this lighter integration is camera safe. It will not burn your phone camera. So, don't worry about it. So, let's move on to compute. Before we get too deep though, it's important to address why we chose to build in-house silicon. It's a non-trivial development effort. Those who've been involved or observed chip development efforts would know that it's time consuming and it requires a worldclass team. It ti the reason for doing it though ties back to the same reasons for building our own in-house ECUs which is velocity, performance, and cost. With our in-house silicon development, we're able to start our software development almost a year ahead of what we can do with supplier silicon. We actually had software running on our in-house hardware prototyping platform well ahead of getting first silicon. Our hardware and software teams are actually coll-located and they're able to develop at a rapid pace that is just simply not possible with supplier silicon. All of this means we're able to get to market sooner with the most cutting edge AI product. Secondly, we understand our application and our vehicle architecture thoroughly and are able to optimize our silicon for our use case. Note, we don't just design for today's use case, but we design with headroom for the models of the future. By building purpose-built silicon, we do not carry the overhead that comes from leveraging a design that was built for some other task and repurposed for autonomous driving. We built this silicon so it would do a really good job at autonomous driving and physical AI problems. All of this enables us to get the best performance per dollar spent. Finally, it's all about cost. When we design in-house, we're able to get the best cost point and the best power points. The cost reductions from our design come from the fact that this is optimized for our use case. Not just the chip use case, but we look at the whole vehicle use case and as well as the fact that there's a meaning meaningful reduction in supplier margins. Now, join me as we go look for the silicon inside the R2 vehicle. Our Gen 3 autonomy computer is the next step in our vertical integration journey and features our very own Riven design custom silicon. It is a highly integrated solution. As you can see, there is very little on the board beyond the two instances of Riven silicon power supply and passives. The hardware and software on this computer are fully designed and developed by Rivian. This computer achieves four times the peak performance of our Gen 2 computer while improving power efficiency by a factor of two and a half. So let's go take a look at the chip. The Rivian autonomy processor or RAP one as we call it is the first in a series of silicon built for physical AI. It is actually much more than one piece of silicon. It's a multi-chip module or MCM that integrates Rivian silicon and memory die. Our custom Rivian silicon is produced on a leading edge TSMC 5 nanometer automotive process. The star of the die of course is a Rivian design neural engine which is capable of 800 sparse int8 tops. TOPS stands for trillion operations per second and is a common measure used to assess AI performance. The chip was also designed with the intent of providing different cost and performance points. We can put multiple rap processors together in a system and they can talk to each other via custom high-speed link we call rib link. So this gets fun. Let's take the lid off wrap one. Look under it. What you see is the RAP one SOC in the middle surrounded by three memory drive spread across two sides. This allows for three independent LPDDR5 channels but more importantly allows for very tight integration between the SOC and memory enabling a very clean data eye between them which in turn then enables high memory bandwidth. With RAP one, we're one of the first to introduce multi-chip module packaging for highMP compute applications in automotive. This is not to be confused with systems and packages or SIPs which have existed in automotive for a very long time. So it is well known that memory bandwidth is key for AI applications and this tight coupling enables us to achieve a net bandwidth of 205 gigabytes per second. The MCM design also enables us to significantly simplify the design of the PCB. The PCB that it sits on no longer has to accommodate DRM chips with critical timing constraints, which means it can be smaller, simpler, and implemented with fewer layers. All of which results in a meaningful cost reduction. In summary, using an MCM style package enables us to achieve higher bandwidth and lower cost. The SOC itself is designed to solve the needs of autonomous driving. Looking ahead, as mentioned earlier, the star of the show is a Rivian design neural engine capable of 800 intake tops. In addition to neural engine, the SOC has a plethora of other blocks that are required to complete the system. The application processor complex is implemented using 14 power efficient ARM Cortex A720 AE cores. This allows us to leverage the rich open-source software ecosystem that comes along with ARM processors. We will be the first OEM to introduce the ARM V9 compute platform for automotive using the Cortex A720 AE in production vehicles. [Applause] In addition to an application processor complex, an SOC for autonomous driving applications needs a high availability safety island and compute that is real-time capable. Our SOC implements both a safety island and a real-time processing unit which are built using eight ARM Cortex R52 cores. And finally, we have all the other pieces you would need for sensor processing, including an image signal processor, encoder, GPUs, etc. The Rivian neural engine itself is designed to implement state-of-the-art deep learning models for perception, control, and planning. It is flexible and supports multiple supports mixed precision data formats. The neural engine has native support in hardware for the latest and greatest in AI model technology for inference. An inference chip today must support transformers and support them really well and that is what we do. We also include support for all types of attention such as multi-headed attention, deformable attention and more. But we don't just implement transformers. The hardware has a host of other capabilities. Some examples include support for nonlinear functions such as softmax and of course we do simple networks such as CNN's. We also have special hooks in the hardware to support LAR and radar processing both of which can be very unstructured unlike transformers and CMM. Earlier I talked about designing an MCMM package with a view to optimizing for memory access. Well, we didn't just stop there. We looked at memory very holistically. Our neural engine, for example, supports weight decompression as another way to relieve the pressure on memory bandwidth. And we don't just enable running one model at a time. We can support the concurrent execution up to four models at any given time. All of this would be irre irrelevant if we didn't have the software and tools to actually exploit the hardware. We made a significant investment in the development of tools and a middleware stack that can exploit the power of rep one. The entire software stack is fully developed inhouse. Our in-house tools include an in-house compiler that can take standard models and generate code targeting our neural engine. We also support profiling tools that can help users optimize their code. And finally, we have an in-house middleware stack that enables us to write application code that is target agnostic. We use that same middleware stack on our Gen 2 hardware platform and we will be using it again on our Gen 3 platform. So what makes silicon for physical AI different from general silicon targeting inference is the importance of functional safety. RAP one was designed from the get-go to factor factor in functional safety in every block of the design. We adhere to the ISO26262 scheme for risk classification which is also called automotive safety and integrity levels which is a mouthful which is why most people call it ASIL. Every block is designed in appropriate ASIL level based on the scheme and then hardware and software are implemented to ensure that the level is achieved. Even our chipto-chip interconnect the rib link is protected using the scheme. So in some cases it means that you actually put an extra redundant hardware in the chip uh which is does the same function in twice and you cross-check the results. In some other cases it means you put ECC on memories instead of parody which a lot of other chips would do. Um so we have a really a lot of hardware hooks and bells and whistles to make sure that this chip is really functionally safe. In addition we also have software. It doesn't stop with the hardware. We also have software that runs on the chip when the chip is actually working in the vehicle. We have software that runs at keyon to make sure the chip is still functionally safe and it runs periodically to complete the whole solution. The rock one chip is not meant to be just you know one incantation. It's one inc instantiation. It's really designed to be scalable. While the first instantiation is a twochip solution targeting autonomy in the R2 vehicle platform, it can be easily extended to solve other physical AI problems such as in robotics. It can scale down to a single chip solution for low cost or scale up to multiple chips for more performance. Rivlink was specifically designed to allow multiple RAP chips to talk to each other via a high bandwidth low latency interface at data rates of up to 128 gigabits per second. Rivlink allows sensor data from one SOC to be seamlessly shared with other SOC's. The scalability doesn't just end there. RAP one was designed to also be flexible in configuration. While the system to be deployed in R2 is liquid cooled, we have demonstrated that this can be con configured as an air cooled system. I'm happy to share that we have successfully demonstrated that our silicon is robust and meeting the performance goals we set out at the start of the project. While peak tops are useful to indicate the capability of the hardware, a more useful measure is perhaps the ability of the system to process sensor data. We have shown that a third gen autonomy hardware system is capable of processing five billion pixels per second of sensor data. [Applause] We are very proud to be at the leading edge of multimodality sensing and to be continuing our trajectory of vertical integration with our RAP one chip and Gen 3 autonomy computer. We expect that at launch in late 2026, this will be the most powerful combination of sensors and inference compute in consumer vehicles in North America. We are now actively testing the silicon and systems and vehicles. For those of you attending this attending this event in person, you'll get a chance to see some of our subsystem test box that test the entire hardware software configuration. We have also integrated the hardware into our R2 vehicles and are continuing to test it extensively. I will now hand it over to James who will show you how his team is continuing to improve the autonomy experience for our customers and how he plans to harness the power of Rap One to make autonomy better. Over to James. [Music] [Applause] Okay, thanks Vidia. So, Vidia just discussed all the amazing autonomy sensors and compute that we'll have on R2. And now I'll go into detail on some of the software that runs on them and powers the Rivian autonomy platform. Firstly, our large driving model is trained end to end from millions of miles of driving sequences collected across our fleets. That's directly from pixels, radar returns, and lightup points to retrajectories. This large driving model uses state-of-the-art techniques based on transformers, auto reggressive prediction, and reinforcement learning that turbocharges our velocity by allowing us to leverage innovation from the world of large language models. It's also built entirely in house and this gives us unprecedented flexibility and being able to change all parts of the stack. So that doesn't that means we don't need to coordinate with other tier ones and tier 2s to make changes. Consequently, our features improve with every update. Finally, and most importantly, the autonomy platform is built on a data flywheel where growth in vehicle fleets and feature adoption drives improvements in autonomy that compound over time. So, let's look under the hood now and discuss the data flywheel in more detail. We'll start with a multimodal onboard model that runs on every customer's vehicle. And the goal for our onboard sensing stack isn't just human level, it's superhuman level. And multiple modalities enable that, allowing our vehicle to see way beyond what a person can. By end-to-end training, the sensor data is early fused into a singular world model. A system where the sensors complement each other, they don't fight against each other. Just like being able to hear can make you a better driver, multiple sensors can make Rivian autonomy better, enabling enhanced precision and more confident predictions. And with more sensors comes richer and better fidelity data. More sensing modalities allows us to achieve the same level of accuracy as a unimodal system but with much less data or to surpass the unimodal system with the same amount of data. So it's a very efficient approach. Let's see how this works in practice by visualizing the output of early fusion. So of course we start with the cameras. Now as video mentioned our cameras are really good. Some of the highest combined megapixel counts and dynamic ranges of any vehicle on sale today. And what you can see here is that when cameras clearly sense things, the system works very well. In fact, cameras alone can handle most autonomy tasks most of the time. But for full autonomy, most of the time isn't enough. Autonomy needs to work all of the time on a moonless night, in the snow, and here in the fog. And in those case, in those cases, cameras alone don't cut it. If we can't sense something, we can't expect the system to handle it. And so, we add radar to the mix. This is where we are with every Gen 2 R1 vehicle today. You can see that we're now able to detect much more. The system can detect more occluded objects, can assign better depths to objects, and is better at estimating object velocities. The confidences and redundancies are also much greater than with a camera only system. But to unlock full autonomy, we need to go further. Adding LAR creates the ultimate sensing combination. It gives the most comprehensive 3D model of the space the vehicle is traveling through. And as Vidia showed previously, the combination of all three sensors identifies more objects and can detect things more quickly. This trinity of modalities enables autonomy features that R.J. was mentioning such as eyes off driving and personal L4 by increasing perception quality, safety, and redundancy. Now, our onboard network is designed from the ground up to flexibly and incorporate new modalities and new sensors. And together with our in-house silicon team, we're co-designing this network to run optimally on our chip. So over here you can see where the sensors and modalities are fed into the model. Each pixel, each radar return, each lighter return is encoded, projected, and then combined into a geometric feature space shown here in the center. And this is where the fusion happens optimally and automatically learned through end-to-end training. There's no extra complexity added. There's no handwritten rules that need to adjudicate. The network has figured out the best way to combine this information. So what we're visualizing here are the semantic projections of these fused feature maps. And although this this space is very abstract, you know, humans were never really designed to see it, you can already see elements here such as other agents and lane lines showing up. So now this whole fuse tensor is fed into our transformer based decoders to produce the world model. From the same input, the network is is trained to generate all of these different outputs for the world around the vehicle. So that includes other objects and agents shown here in the top left. It also includes dense 3D occupancy to handle short-range maneuvers and narrow negotiation. And here's where we estimate the local map around the vehicle. Note that this is quite different from typical robo taxi efforts where the map is pre-generated offline and then localized online. Those maps are expensive to generate and hard to maintain. But just like a human driver can navigate a road they've never seen before, our local map is produced and accumulated directly from what the vehicle perceives. And in the bottom right here, finally, we have estimated trajectories. These represent the model's best estimate of how to proceed through the scene. This output will be a key technology enabler for pointto-point driving. And I'll discuss more on this piece later. So now we're moving on to the next stop in our flywheel, the autonomy data recorder. This is the system that turns real world driving into data. Essentially, our entire Gen 2 fleet becomes a huge queryable dynamic database of driving scenarios. The data recorder runs trigger code that can fire off any set of signals seen by the world model. Jwalking pedestrians, red light runners, large animals in the road. All of these cases are examples of cases which you can mine with triggers. We can also run more general triggers such as finding divergences between the human-driven trajectory and our predicted trajectories. These might indicate areas, for example, where the model could be improved. We can even push new triggers live to our fleets outside of the usual OTAA cycle. This allows us to capture the data we need on demand with minimal turnaround time, greatly speeding up development. That's a huge driver of developer productivity. As soon as an engineer wants to find more scenarios of an event, they can mine for them immediately. And because ADR is so selective, it's also very efficient. The vast majority of boring driving data is never captured, never uploaded, and never trained on. Now, once a trigger fires, all the sensor data before and after the event is captured, tagged, compressed, and uploaded. And that data is then immediately available to engineers. These scenarios can be used for model training, for evaluation, or for replay through our simulator. Here you can see examples of scenario clusters found automatically by our ADR system. We have examples of environmental conditions such as dusk at night. We have examples of map-based uh scenarios such as tight turns and merges and also agent-based situations like animals, bike racks and semi-truckss. And after upload all of these curated scenarios are immediately available on the Rivian cloud. So as the feat expands and adoption increases the size of our knowledge base is growing and compounding. From next year this growth will be obviously further accelerated with the additional volume of R2. Now all of this data is stored securely and we don't associate any sequences with your bin and if you select it your home or your place of work and through ADR every sequence is already tagged without further processing. Finally, because we store all sensor modalities, the data is incredibly rich and complete. This allows us to autolel most sequences using large offline models, which would be too slow to run on board. In fact, the vast majority of our training data today is autolabeled. That's massively more efficient than using human annotators. So, now let's talk about ground truth fleets. You may have seen other autonomy providers ground truth vehicles. They all typically have multiple non-production sensors such as lidars and other things strapped to the roofs and sides of the vehicles. And these fleets are incredibly valuable for training perception systems. The LAR data is so crisp, it's essentially used as a ground truth that is used to train the other production sensor sets. But because their prototypes, these fleets are typically small, numbering the tens to 100s, represented by the dots on this slide. In contrast, for the lighter equipped R2, every vehicle will become a ground truthing vehicle. That's orders of magnitude more data than other OEMs. It's an incredible force multiplier for better and richer training data that massively accelerates our pro progress. So let's see how we use all of this data to benefit our customers. This happens in the large driving model. The large driving model or LDM is an end-to-end model from sensors to driving actions and it's based on many of the same technologies used in large language models. So LDM uses neural net transformers for processing just like with LLMs and the large driving model uses tokens for training also just like LLMs. But instead of thinking about these tokens as words, they're actually small parts of trajectories that are jointly predicted and assembled together. And the large driving model also uses reinforcement learning just like state-of-the-art large language models. But here, instead of aligning the output with human values and intentions, we align the large driving model output for safe, performant, and smooth driving. Because LDM is such a close cousin to an LLM, we can reap all of the advancements, investments, and innovations being made in improving generative AI and it comp and can apply them directly to our driving task. This makes the LDM incredibly costefficient to develop. So let's do a deep dive into how LDM is trained by reinforcement learning. Over here you can see the sensor data from a scenario. Here we're approaching a stop sign and that sensor data is fed into our transformer-based encoders behind me. Then we sample multiple trajectories from this model token by token and trajectory by trajectory. And the different tokens shape the trajectories in different ways. Once we've sampled all these trajectories, we then need to rank them. Now, the one on the left here is actually the most humanlike. The vehicle slowed almost to a stop, but then rolled it. A lot of our Rivian data is kind of like that. Um, the one in the middle here, they're clearly stopping too soon. And the one on the right is just right. We're stopping behind the line and correctly following the road rules. So here you can see three different trajectories that have navigated this stop intersection. Now what we're able to do is apply our road rule rankers that can then say the third here is the optimal trajectory and then we reorder them and then through back propagation the model is trained to produce more of these type types of trajectories in similar scenarios in the future. Now, that's obviously highly simplified and you're just looking at one scenario here, but imagine this process running millions of times a second across millions of scenarios with a whole database of road rule costs and losses. That's how LDM is trained. We can then distill this model into one that we can run on board. All of this work results in new models, continuous enhancements, refinements, and new features that we continuously deliver to our customers. But how do we know we can release? We've built a cloud-based simulator that runs the whole autonomy stack through millions of miles of real world scenarios on every release. And that allows us to measure safety, comfort, and performance in a statistically significant way without having to manually drive those miles. We also have an capability we call apprentice mode. And before we release features, we can launch them in the background of a previous release. We can then monitor the performance of that new version compared to the human-driven miles, but also compared to the pre the previous version of autonomy. That allows us to do an even bigger evaluation in the tens of millions of miles. And so through simulation and apprentice mode, we can rapidly build the confidence we need to ship new features and enhancements to customers. Because the system is developed entirely in house, we can update any part of the stack from the lowest level camera drivers all the way to the highest level motion planning code. That means that means that the whole stack is always improving with every release and we have a feature roadmap that stretches to the highest levels of autonomy. Now when we survey customers this year on the autonomy capabilities they wanted the most, the answer was resounding. They wanted more road coverage for hands-free highway assist. Now previously we supported 135,000 miles of divided highways. But as RJ mentioned, our map today is about to grow. [Applause] Universal hands-free unlocks over three and a half million miles of hands-free driving on roads across the US and Canada. If there's a painted line and it's clearly marked, you can now drive handsree. Universal Handsfree will be part of our paid tier bundled into one simple package, Autonomy Plus. It's a onetime fee or you can pay monthtomonth. And Autonomy Plus features will be available to all Gen 2 customers for free until March next year. And this is and this is just the beginning for Autonomy Plus. We have many exciting features on the way such as point-to-point automatic parking and enabled by the LAR on R2 eyes off. As our fleet continues to grow, as our adoption continues to increase, our data flywheel will continue to grow. And we've been thinking about this as a circle. But in fact, the system is better on every orbit. So a better analogy is an upward helix continually improving and compounding on itself. And with that, I'd like to like to hand over to Aim to discuss some of the other improvements being made in AI here at Rivian. [Music] Thank you, James. We have made significant progress in our AI enabled autonomy stack. But as James said, it doesn't stop here. AI runs through the core of everything we do. It's a profound platform shift which changes our product and everything we do at the company from the way we design, we develop, we manufacture and we service our cars. This is all made possible by the Rivian unified intelligence, a common AI foundation that understands our products, our operations as one continuous system and personalizes the experience for our customers. So, how does it work? We revamped our vehicle operating system to be AI ready. We developed an in-house multi- aent multi-lm multimodal intelligence platform. The platform is built on a robust data governance framework with security and privacy as main tenants. We have a suite of specialized agents. Every Rivian system from manufacturing, diagnostics, EVR planning, navigation becomes an intelligent node through MCP. And the beauty here is we can integrate third-party agents and this is completely redefining how apps in the future will integrate in our cars. We orchestrate multiple foundation models in real time, choosing the right model for each task. And we support memory and context, allowing us to offer advanced levels of personalized experience. And the architecture is natively multimodal using audio, vision, and text through the same unified layer. The beauty of our architecture is the seamless integration between the cloud and the edge. Edge AI with an embedded small language model allows us to achieve higher levels of performance, lower latency, and the best conversational experience. And wait till the R2 comes. R2 will have close to 100 tops AJI fully dedicated to INC cabin experience. This will allow us this will allow us to move most of the intelligence workloads from the cloud to the edge powering an incin AI experience fully available when the car is offline. [Applause] The Rivian unified intelligence is the connective tissue that runs through the very heart of Rivian's digital ecosystem. This platform enables targeted agentic solutions that drive value across our entire operation and our entire vehicle life cycle. Let's start with the factory. Our diagnostics agent is the ultimate example of unified intelligence in action. It instantly connects real time telemetry from vehicles on the assembly line and allows us to validate quality at production, identifying changes needed before the vehicle leaves the factory. The same unified intelligence is fundamentally helping us redefine service because the platform is grounded in real vehicle data. Every Rivian technician is now being empowered by AI systems trained on live software data, service manuals, repair histories, electrical diagrams, and supply chain logistics. This is accelerating service repair time by hours, helping us to dramatically improve technician efficiency. And it doesn't stop with our technicians. The same platform that in will enable in the future our customers to self troubleshoot and resolve minor issues directly from the Rivian removal app and from your Rivian car. [Applause] Now let me tell you more about the invehicle experience. This architecture will fundamentally reshape how we all interact with our Rivians. Today I am very excited to announce the Rivian assistant. [Applause] As you see from the beautiful wave behind me that our UX team has designed, the Rivian assistant is fully integrated into the user experience and our incar operating system. It's designed to understand you, your vehicle, and the context you're in. And the Rivian assistant will be available for all Rivian Gen 2 and Gen One customers in early 2026. Now, rather than tell you about all its features, we'll take some risk here and then we'll do a live demo [Applause] from the beautiful R1S Borealis which is over there. So Oscar will be our main driver today. Oscar is the product manager behind the assistant. Oscar, are we ready? Yes. Uh, I'm ready. Thank you. Wasim, hello everyone. Let me show you what the Rivian assistant can do. To initiate the Rivian assistant, you can either hold the left steering wheel button or just say, "Hey, Riven." The assistant is not just an integration of a chatbot on top of the vehicle UI. We built an agentic framework allowing us to integrate into the larger ecosystem and bring your own digital context to the car. We started with Google Calendar. We had a ton of fun collaborating with the Google team on this project. Oscar, can we see that in action? What's on my calendar today? What's on my calendar today? You have two events today. Call with Tim from 2:00 p.m. to 2:30 p.m. and meet up with Wasim from 3:00 p.m. to 400 pm. [Applause] The agentic integration allows us to not only connect with your calendar and read it, but also take actions and manage it. Oscar, can you show us that in action? Can you move my call with Tim to 5:00 p.m.? I've moved your call with Tim to 5 p.m. As you see, the same action would have probably taken me multiple tabs and clicks and swipes to get to the same result. The assistant can help you control your calendar in a much safer and easier way. And Google Calendar is just the beginning. The platform will expand to many more applications and many more thirdparty agents as they become available. Beyond the agentic integration, our AI platform can blend your personal context into your vehicle context. In this case, the Google calendar agent is connected to other vehicle applications. Oscar, can we see a life example? Let's go to my meet up with Wasim. Navigating to Ferry Building. You'll drive for 55 minutes and arrive around 10:50 a.m. [Applause] And the AI integration goes much deeper than navigation. Because this is an AI native operating system, it can connect with all vehicle apps, including our customers favorites, EVR planner or drive modes. Oscar, can we see that in action? How much battery will I have when I get to my destination? You will have 67% battery remaining, which is about 231 miles of range. Can you switch to a more efficient drive mode? [Applause] As you see, Oscar did not have to specify conserve mode. This is the future of vehicle UIs. The assistant allows user to perform use cases without knowing the exact vehicle command, without knowing where they are in the different menus in the UI, making it way easier for every user to interact with the Rivian. Oscar, can we see another example? Can you make the seats toasty for everyone except me? Let's check it this out. As you see, all seats except the driver's seat are warmed up. Imagine the possibilities that this opens up. The assistant takes the vehicle experience to the next level. Instead of having multiple UI commands, multiple taps on the touch screen, you can perform the whole task with just one natural language command. And one of the features that our community has requested and I promise that we will get it is messaging. But I really wanted to get messaging right. So, let me text Oscar. I'm getting a call right now. Lots of texts that people are seeing me on the live feed. Thank you. Read my last message from Wasim. Hey, I'm nearby. Can you find a good restaurant near the ferry building? Let me know. Do you want to reply to Wasim? So we did not only want to stop at reading your message even with this beautifully integrated UI. The messaging app with this new AI platform is fully integrated into the vehicle operating system and has access to every single application and control. Oscar, can we see an example? We're hungry. Can you find some restaurants near my destination? I found terine about 30 miles away in San Francisco. It has a 4.2 star rating. Would you like to go there or explore other options? Actually, can you send a text message to a seam and uh show him the top three options from this list? Ask him which one he wants to meet up at and include my ETA. Would you like to send a message, California, Eclipse Kitchen and Barcadero Center, San Francisco, California 3, Super Duper Burgers, 98 Mission Street, San Francisco. I think I would go for Super Duper. is 10:53 a.m. to Yeah, send it. Receive and Oscar, while I appreciate the invitation, I think I have something else to do right now. You just saw the difference that native integration makes. The assistant has memory, has context, it remembers the full story, who you're talking to, where you're going, and what you just searched for. And then it puts everything into a perfect message. What you've witnessed today is more than just a new voice assistant. This is a peak into the Rivian unified intelligence platform that powers the new foundation of Rivian's digital ecosystem. The gap between softwaredefined vehicles and traditional architectures is getting exponentially wider with AI. Rivian is uniquely positioned to move from a softwaredefined vehicle and bring to the world an AI definfed vehicle. With this, let me welcome ARJ back to the stage for some closing comments. [Applause] That was so much fun to watch. Um, so much time and effort have gone into building the platforms, designing the architectures. Um, of course building the teams, growing the teams, organizing the teams, all the work that goes into these really complex systems. And I often describe it as as if you're building the plumbing. And if you're building a house, you don't start with the finished house. It takes years of time of planning. You then have to do foundation work. You then have to do wiring and plumbing that go into the house. And then at the very end, it all comes together. And so, and you saw our our work that we talked about on our in-house processor, this is something that has been years in the making. And you know, the amount of effort and the amount of time that's gone into it. And by the way, the amount of effort that went into this not leaking, which is amazing, um, is just so inspiring. And I spent some time last night with the team, you know, talking about this right before we're about to show it today. And one of the lead engineers looked at me and said, "Boy, this is like we've been working on this for years, and I haven't been able to talk about it. It's so cool. Tomorrow I can start to talk about what I do every day, all day long." But, you know, between the work that we put into the processor, the large driving model that you heard us all talk about, how that feeds our data flywheel and this this large flywheel approach to building a data set that continues to improve our model and then as you just heard from wasim all the work that got went into first building a software definfined architecture, developing and building all the electronics that go into the vehicle and then using that as the foundation for enabling an AI definfed vehicle. This is coming together today. You're seeing the house start to form in front of you. And um what customers are going to see on our Gen 2 R1 vehicles starting very soon is a lot of these features. Uh as I said later this month, we're going to be growing the amount of miles you can access universal hands-free from just under 150,000 to 3 and a half million miles. 2026 pointto-oint navigation. Shortly thereafter that hands off, eyes off. So, we're very very excited and we appreciate all of you being here today. Now, for those that are here, we have demos. Uh I I know there's a lot of people here, so not everybody will get to do a demo, but for those that are doing demo, you'll get to see uh the pointto-oint navigation, the work that's gone to that. We have a bunch of great displays that show some of the hardware. Uh but for those that aren't here, uh we appreciate you listening along and we appreciate your support and enthusiasm for building. Thanks, everyone. See you later.