Elon Musk Podcast

Why Tesla Must Retrofit Four Million Cars

16 min
Apr 25, 20263 days ago
Listen to Episode
Summary

Tesla faces a critical hardware limitation affecting 4 million vehicles that lack the computing power for unsupervised full self-driving. The episode explores how the shift to end-to-end neural networks has exposed memory bandwidth constraints in older hardware, forcing Tesla to pursue retrofits, trade-in programs, and custom chip development while competitors adopt NVIDIA's all-in-one approach.

Insights
  • Memory bandwidth, not raw processing power, is the primary bottleneck limiting autonomous driving capability in current vehicle hardware
  • The industry-wide shift from modular software to end-to-end neural networks fundamentally changed hardware requirements, making older architectures obsolete
  • Tesla's strategy involves custom-designed, vision-optimized silicon rather than general-purpose computing, creating a fundamentally different approach than competitors
  • Hardware obsolescence in autonomous vehicles may force consumers to replace cars every few years rather than keep them as long-term assets
  • Vertical integration of semiconductor manufacturing (TerraFab) represents a strategic pivot to control the entire hardware-software iteration cycle
Trends
Autonomous driving hardware becoming a primary differentiator between automakers rather than software aloneMemory bandwidth emerging as the critical constraint in AI-driven vehicle systems, not CPU speedIndustry split between custom-optimized silicon (Tesla) versus all-in-one supercomputer approaches (NVIDIA competitors)Vertical integration of semiconductor fabrication as competitive advantage for autonomous vehicle manufacturersTriple modular redundancy and shadow mode testing becoming standard safety architecture in next-generation autonomous systemsPseudo-LiDAR (3D depth from 2D cameras) driving extreme memory bandwidth requirements in vision-only autonomous systemsHardware upgrade cycles creating planned obsolescence concerns in consumer vehiclesHumanoid robotics driving next-generation chip architecture requirements beyond autonomous driving needsSupply chain control through in-house chip fabrication becoming strategic priority for major automakersTrade-in programs emerging as primary strategy to manage hardware obsolescence rather than large-scale retrofits
Companies
Tesla
Primary subject; facing 4M vehicle retrofit challenge due to Hardware 3 limitations for unsupervised full self-driving
NVIDIA
Competitor offering DriveThor platform adopted by BYD, Lucid, and Xiaomi as alternative to Tesla's custom silicon
BYD
Automaker adopting NVIDIA's DriveThor platform instead of developing custom autonomous driving hardware
Lucid
Luxury automaker using NVIDIA DriveThor platform for autonomous driving and vehicle compute systems
Xiaomi
Tech company adopting NVIDIA DriveThor platform for autonomous vehicle development
SpaceX
Partnering with Tesla and Intel to construct TerraFab semiconductor fabrication facility in Texas
Intel
Partnering with Tesla and SpaceX on TerraFab semiconductor manufacturing facility development
ARM
Cortex-A72 CPU architecture used in Tesla's custom autonomous driving chips
People
Elon Musk
Admitted that 4 million Tesla vehicles lack hardware for unsupervised full self-driving capability
Quotes
"If you own a hardware 3 vehicle and you are offered a compelling enough discount on a brand new car that already has the advanced hardware, you are highly likely to take the upgrade."
HostMid-episode
"You're trying to build the perfect box for an intelligence that hasn't finished growing."
HostLate-episode
"The race for full autonomy is no longer just a software problem solved by clever code. It is a brutal battle defined by physical silicon limits, the realities of hardware manufacturing, and just how fast you can push data through a wire."
HostConclusion
"Hardware 3 relies on older LPDDR4 memory. It simply cannot move the data fast enough to feed the demands of these large transformer models."
HostEarly-mid episode
"With three chips, they can compare their results. If chip A hallucinates a wall, but chips B and C both agree the road is completely clear, the system votes to ignore chip A."
HostMid-episode
Full Transcript
Elon Musk recently admitted that approximately 4 million Tesla vehicles currently on the road lack the hardware to achieve unsupervised full self-driving. Yeah, and that physical limitation, I mean, it forces this enormous retrofit effort for older cars. It really reveals a hidden ceiling in the tech world. We're looking at how this insatiable demand for processing power is, you know, physically altering in-car computers. It's forcing companies to scrap their roadmaps and creating a hard divide in how different automakers approach the future of driving. Right. So how does a company pivot when its core autonomous driving software suddenly outgrows the silicon powering its fleet? To answer that, you really have to look closely at the physical limitations of the computers already welded inside these vehicles. Like if you look at Tesla's older hardware, which they call hardware three, there is a severe physical constraint happening right on the board itself. OK, what kind of constraint? Well, its memory bandwidth is only one-eighth that of its successor, Hardware 4. Wait, one-eighth. So memory bandwidth, we're talking about the actual speed at which data travels back and forth between the memory chips and the main processor, right? Precisely. And the reason memory bandwidth is causing such an emergency right now comes down to how autonomous driving software is currently being built. The entire industry is abandoning old methods and moving toward what are called end-to-end neural networks. If you look at how self-driving software worked in the past, it was highly modular. You had one specific piece of software tasked with looking at camera pixels and, say, identifying a stop sign. Right. And then a totally separate piece of software applied a hard-coded rule written by a human that told the car to apply the brakes if a stop sign was detected. Yeah, exactly. But end-to-end neural networks throw out those hard-coded rules entirely. They take raw camera pixels and map them directly to driving actions like steering and braking. So the software just learns by watching millions of hours of human driving video rather than following a rigid list of instructions typed out by a programmer in an office. That is the fundamental shift. And to execute that kind of learning, these systems rely on large transformer models. And these models require massive amounts of extremely high-speed memory to store what engineers call activation maps. Activation maps. Okay. You could visualize an activation map as the car's short-term working memory. When you approach a busy, chaotic intersection, you naturally track moving objects, right? Yeah, of course. The computer needs to do the exact same thing. It needs to remember that a pedestrian stepped off the curb three seconds ago, even if that pedestrian is currently hidden behind a delivery truck that just pulled forward. Oh, I see. So it is less about how fast the processor calculates the math and more about how much data can physically fit through the pipe at any given microsecond. Yes. It sounds like trying to drink from a fire hose with a coffee stir. That analogy perfectly captures the bottleneck. Hardware 3 relies on older LPDDR4 memory. It simply cannot move the data fast enough to feed the demands of these large transformer models. And the processor itself might have the capability to do the math. But it's just sitting there waiting. Exactly. It sits there, starved for data, waiting for the memory to catch up and deliver the next frame of the environment. And this physical traffic jam on the circuit board dictates the overall size of the neural networks the car is allowed to run. Because if the pipe is too small, you cannot run the bigger, smarter network. Right. Which totally prevents true unsupervised autonomy. The car literally cannot hold enough of the surrounding environment and its active memory to safely navigate without a human sitting there ready to grab the wheel. Wait, back up a second. Sure. If those four million hardware three cars currently sitting in driveways cannot do unsupervised driving, what happens to them? I mean people paid thousands of dollars for software they were told would eventually drive them to work while they slept Yeah well Tesla plans to establish specialized microfactories in major urban areas to physically swap out the computers and the cameras in those hardware three vehicles. Microfactories. Yeah, and while they try to figure out the logistics of that, they are offering a software compromise called FSD version 14 Lite. It's an interim solution designed to run on the older hardware until the cars can be physically rebuilt. Setting up entirely new micro factories just to swap out computers, that sounds like an astronomical expense. You have to secure real estate in expensive cities, hire specialized technicians, build isolated supply chains strictly for retrofitting older models. I mean, why not just send people to the regular service centers? Because standard service centers are just too inefficient for a complex swap like this. You have to remember, you are not just unbolting a metal box and plugging in a new one. Right. You have to carefully remove the old cameras, install new high-resolution 5-megapixel cameras, integrate the new AI4 compute module, and then, this is the hard part, precisely align every single sensor so the new software receives perfect, undistorted visual data. Oh, wow. Yeah, service centers are designed for standard maintenance. They fix brakes, align tires, replace cracked windshields. Exactly. They're not built to perform major surgical overhauls on the sensory nervous system of a complex machine. Relying on the existing service center network would clog up normal operations for years, and it would just create an impossible financial burden. But even with specialized microfactories, the cost of ripping apart millions of cars just to fulfill a past promise seems financially ruinous anyway. Every hour a technician spends doing surgery on a four-year-old car is an hour they aren't building a new one. That financial reality is exactly why offering heavily discounted trade-ins to new vehicles equipped with hardware 4 might end up being their primary strategy. Oh, to avoid the physical retrofits entirely. Right. Think about the psychology of the consumer. If you own a hardware 3 vehicle and you are offered a compelling enough discount on a brand new car that already has the advanced hardware, you are highly likely to take the upgrade. Win-win, sort of. The customer gets a new car, and the company completely avoids the agonizingly slow, expensive process of tearing down and rebuilding a used vehicle in a temporary microfactory. But meanwhile, they are continuing to change the hardware on the production line right now. Tesla is already shipping a modified three-chip computer, which is labeled AP45, some recent Model Ys. And they just announced a future upgrade called AI4 Plus that doubles the system RAM to 64 gigabytes. Yeah, so the transition from a two-chip architecture to a three-chip architecture introduces a really vital concept called triple modular redundancy. Okay, what does that mean? Well, if you look at their previous computers, they used two chips for redundancy. If chip A failed completely, chip B took over. Simple. But a three-chip layout allows the system to actively vote on reality. Vote on reality, like an election for what the car should do next. In a way, yes. Imagine a scenario where, I don't know, a cosmic ray strikes the silicon and physically flips a bit from a zero to a one, causing a momentary glitch. Okay. Or perhaps the software on one chip briefly misinterprets a weird shadow on the road as a concrete barrier. If you only have two chips and they disagree, the system panics and hands control back to the human. Right, because it doesn't know which one is right. Exactly. But with three chips, they can compare their results. If chip A hallucinates a wall, but chips B and C both agree the road is completely clear, the system votes to ignore chip A. The outlier is outvoted. Wow. This ensures continuous operation and prevents the car from phantom braking or violently disengaging on the highway Three silicon brains debating the safest action in real time That is wild It is And a three setup serves as another critical function too It allows one of the chips to run highly experimental software in the background In the industry, this is known as shadow mode. Oh, so the car physically drives using the proven, stable software running on two chips. Yep. While the third chip silently tests next-generation code against the real-world conditions the driver is experiencing. Exactly. It learns and processes the environment without ever risking passenger safety, because its outputs are physically disconnected from the steering wheel. And what about the AI4 Plus upgrade? Because doubling the RAM to 64 gigabytes seems entirely focused on that memory pipe issue we discussed earlier. It totally addresses the core problem of neural network weights. The upcoming AI4 Plus hardware increases the raw compute speed and the memory bandwidth by roughly 10%, which is, you know, a nice bump. Sure. But the critical change is that massive increase in RAM. End-to-end neural networks are not static. They're constantly growing. The weights, which are the billions of mathematical parameters the artificial intelligence learns and refines during its training phase in the data center, they take up physical space in the car's memory. Right. As the models get smarter and more capable of handling complex edge cases, those weights get heavier. Doubling the RAM prevents the current fleet from hitting a computational wall as the software inevitably becomes more complex over the next couple of years. Hold on. If they are upgrading the hardware again right now, are current hardware 4 owners going to face the exact same obsolescence issues as hardware 3 owners in a few years? I mean, if the neural networks just keep getting fatter, a 64 gigabyte limit will eventually be reached too. You are identifying the central tension in the entire autonomous driving space right now. A company can claim their current hardware is capable of unsupervised driving, but when they simultaneously release incremental upgrades with double the memory and extra processors, well, it reveals a lack of certainty. Yeah, no kidding. It really highlights the immense difficulty of predicting exact hardware requirements for software that hasn't actually been fully solved yet. You're trying to build the perfect box for an intelligence that hasn't finished growing. And the hardware roadmap extends even further, right? Tesla's next-generation AI-5 chip has officially completed its design phase, but according to the sources, it will actually be prioritized for their Optimus humanoid robot and their data centers, rather than being immediately deployed into cars. Yeah, the compute requirements for robotics are just on an entirely different level than driving. Well, a humanoid robot requires unstructured 3D spatial reasoning and constant tactile feedback from its environment. Look, navigating a two-dimensional road is certainly difficult, but roads have clear rules, painted lines, and a general flow of traffic. The environment is somewhat constrained. Exactly. Now, picture a robot walking through a cluttered human kitchen. It has to recognize a fragile glass sitting precariously on a counter, calculate the exact pressure needed to pick it up without shattering it, navigate around a dog sleeping on the floor, and place the glass into a dishwasher. That is a fundamentally different engineering challenge. Uses orders of magnitude more compute-intensive than staying between two white lines on a highway. Because a humanoid robot is dealing with infinite variations of human environments, rather than a standardized public road system. Exactly. So to handle the immense processing load of physical AI, the AI-5 chip offers five times the memory bandwidth of the current AI-4 architecture. Five times, wow. Yeah, and the physical manufacturing of these chips is becoming a bottleneck in itself. To secure the massive volume of silicon required for both millions of cars and millions of robots, Tesla is partnering with SpaceX and Intel to construct a massive semiconductor fabrication facility in Texas They calling it TerraFab TerraFab This facility basically opens up the ability for them to manufacture their own logic chips and memory completely under one roof It sounds like building a silicon fortress Instead of relying entirely on external foundries spread across the globe, they are attempting to control the physical manufacturing of their brains from the ground up. Right, because the standard practice across the tech industry is fables design. You design the chip architecture on a computer in California, and you pay another company in Asia to actually bake the silicon and manufacture it. Right. But relying on external supply chains creates delays. Building a research-grade wafer fab in Texas that integrates mask fabrication, logic chips, memory, and packaging in the exact same physical building, it creates the fastest possible iteration cycle. Oh, I see. You can design a new chip architecture on Tuesday, test it, and scale it into production without ever waiting in line behind other tech giants at an external foundry. Precisely. Well, we should look at the rest of the industry because the approaches are really splitting here. While Tesla develops this highly specialized, vision-focused silicon relying on older ARM Cortex-A72 CPUs and GDDR6 memory competitors like BYD, Lucid, and Xiaomi are adopting NVIDIA's DriveThor platform. Yeah, and NVIDIA's Thor platform represents a completely different engineering philosophy. How so? Thor utilizes cutting-edge manufacturing processes and server-grade CPUs to deliver 2,000 teraflops of raw compute power. That's Loretta math. It is. And it does not just focus on driving. It is designed to run the entire vehicle. Thor handles the complex autonomous driving neural networks, but it also has the power to run the infotainment screen, process the digital dashboard, render 3D graphics for passengers, and manage in-car voice assistants simultaneously. So it is an all-in-one supercomputer for the car, just brute force doing everything at once. Yes, and that brute force requires immense power. Tesla's custom architecture, on the other hand, deliberately sacrifices general-purpose CPU power to maximize one specific thing, memory bandwidth. Back to the fire hose. Exactly. They use GDDR6 memory, which is the type of memory typically found in high-end graphics cards for video games. They use it to heavily prioritize the rapid, continuous transfer of high-resolution video streams. You have to remember, they do not use LiDAR lasers or radar sensors anymore. Right, they rely entirely on flat cameras. Right, so to understand depth, to know exactly how many feet away a vehicle or a pedestrian is, the computer has to constantly analyze flat, two-dimensional pixels and calculate 3D geometry in real time. They call that pseudo-LiDAR, right? Deriving 3D depth from 2D images. Exactly. And running pseudo-LIDAR on eight high-resolution camera streams simultaneously requires moving a tremendous amount of video data without a single millisecond of lag. That is why they willingly accept older, less powerful CPU cores in their custom chips. They trade general processing capability in exchange for maximum memory bandwidth, ensuring that video data never, ever stops flowing into the neural network. So does a custom, hyper-specialized design focus purely on video data give an automaker an edge? Or does NVIDIA's massive all-in-one compute power offer a safer bet for companies trying to build the cars of the future? You know, the race for full autonomy is no longer just a software problem solved by clever code. It is a brutal battle defined by physical silicon limits, the realities of hardware manufacturing, and just how fast you can push data through a wire. Right. It leaves you wondering, will the constant escalating need for more processing power turn modern cars into temporary tech gadgets that inevitably age out and require replacement every few years rather than long-term assets that sit in our garages for decades? If you're not subscribed yet, take a second and hit follow on whatever app you're using. It helps us keep making this. We appreciate you being here. Also, check out our YouTube channel for more business and tech updates. There's a link in the description.