Robohub.org
 

Why Tesla’s Autopilot and Google’s car are entirely different animals


by
28 October 2015



share this:

Google_Tesla_robocar_In the buzz over the Tesla autopilot update, a lot of commentary has appeared comparing this Autopilot with Google’s car effort and other efforts and what I would call a “real” robocar — one that can operate unmanned or with a passenger who is not paying attention to the road. We’ve seen claims that “Tesla has beaten Google to the punch,” but while the Tesla release is a worthwhile step forward, the two should not be confused as all that similar.

Tesla’s autopilot isn’t even particularly new. Several car makers have had similar products in their labs for several years, and some have released it to the public, at first in a “traffic jam assist” mode, but reportedly in full highway cruise mode outside the USA. The first companies to announce it were Cadillac with the “Super Cruise” and VW’s “Temporary Autopilot” — but they delayed that until much later.

Remarkably, Honda showed off a car ten years ago doing this sort of basic autopilot (without lane change) and sold only in the UK. They decided to discontinue the project, however. That this was actually promoted as an active product ten years ago will give you some clue as to how different this was from the bigger efforts.

Cruise products like these require constant human supervision. With regular cruise control, you could take your feet off the pedals, but might have to intervene fairly often either by using the speed adjust buttons or full control. Interventions could be several times a minute. Later came “Adaptive Cruise Control”, which still required you to steer and fully supervise, but would rarely require intervention on the pedals while driving on the highway — a few times an hour might be acceptable.

The new autopilot systems allow you to take your hands off the wheel but demand full attention. Users report needing to intervene rarely on some highways, but frequently on other roads. Once again, if you only need to intervene once an hour, the product could make your drive more relaxing.

Now consider a car that drives without supervision …

Human drivers have minor accidents about every 2,500 to 6,000 hours, depending on what figures you are using — that would be about once every 10 to 20 years of driving. A fatal accident takes place every 2,000,000 hours of driving — about once every 10,000 years for the typical driver, thankfully a much longer span than a person’s lifetime.

If a full robocar needs human intervention, logic tells you that it’s going to have an accident because there is nobody there to intervene. Just like with human drivers, most of the errors that would cause an accident are minor: running off the road, fender benders. Not every mistake that could cause a crash or a fatality causes one. Indeed, humans make mistakes that might cause a fatality far more often than every 2,000,000 hours, because we “get away” with many of them.

Even so, the difference is staggering. A cruise autopilot (such as Tesla’s) is a workable product if you have to correct it a few times an hour, whereas a full robocar product is only workable if you need to correct it only after decades or even lifetimes of driving. This is not a difference of degree, it is a difference of kind. It is why there is probably not an evolutionary path from the cruise/autopilot systems based on existing ADAS technologies to a real robocar. Doing many thousands times better will not be done by incremental improvement. It almost surely requires a radically different approach, and probably very different sensors.

To top it all off, a full robocar doesn’t just need to be great at avoiding accidents. If it’s running unmanned, with no human to help it at all, it needs a lot of other features and capabilities too.

The mistaken belief in an evolutionary path also explains why some people imagine robocars are many decades away. If you wanted evolutionary approaches to take you to 100,000x better, you would expect to wait a long time. When an entirely different approach is required, what you learn from the old approach doesn’t help you predict how the other approaches — including unknown ones — will do.

It does teach you something, though. By simply being on the road, Tesla will encounter all sorts of interesting situations its developers weren’t expecting, and they will use this data to train new generations of software that do better. They will learn things that help them make the revolutionary unmanned product they hope to build in the 2020s. This is a good thing.

Google and others have also been out learning that, and soon more teams will.

This post originally appeared on Robocars.com.


If you liked this article, you may also be interested in:

See all the latest robotics news on Robohub, or sign up for our weekly newsletter.



tags: , , , , ,


Brad Templeton, Robocars.com is an EFF board member, Singularity U faculty, a self-driving car consultant, and entrepreneur.
Brad Templeton, Robocars.com is an EFF board member, Singularity U faculty, a self-driving car consultant, and entrepreneur.





Related posts :



Using generative AI to diversify virtual training grounds for robots

  24 Oct 2025
New tool from MIT CSAIL creates realistic virtual kitchens and living rooms where simulated robots can interact with models of real-world objects, scaling up training data for robot foundation models.

Robot Talk Episode 130 – Robots learning from humans, with Chad Jenkins

  24 Oct 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Chad Jenkins from University of Michigan about how robots can learn from people and assist us in our daily lives.

Robot Talk at the Smart City Robotics Competition

  22 Oct 2025
In a special bonus episode of the podcast, Claire chatted to competitors, exhibitors, and attendees at the Smart City Robotics Competition in Milton Keynes.

Robot Talk Episode 129 – Automating museum experiments, with Yuen Ting Chan

  17 Oct 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Yuen Ting Chan from Natural History Museum about using robots to automate molecular biology experiments.

What’s coming up at #IROS2025?

  15 Oct 2025
Find out what the International Conference on Intelligent Robots and Systems has in store.

From sea to space, this robot is on a roll

  13 Oct 2025
Graduate students in the aptly named "RAD Lab" are working to improve RoboBall, the robot in an airbag.

Robot Talk Episode 128 – Making microrobots move, with Ali K. Hoshiar

  10 Oct 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Ali K. Hoshiar from University of Essex about how microrobots move and work together.

Interview with Zahra Ghorrati: developing frameworks for human activity recognition using wearable sensors

and   08 Oct 2025
Zahra tells us more about her research on wearable technology.



 

Robohub is supported by:




Would you like to learn how to tell impactful stories about your robot or AI system?


scicomm
training the next generation of science communicators in robotics & AI


 












©2025.05 - Association for the Understanding of Artificial Intelligence