Robohub.org
 

Is Apple building a robocar? Maybe. Or maybe not.


by
05 February 2015



share this:

Dodge caravan self driving car in brooklyn new york_Apple_robocar?There is great buzz about some sensor-laden vehicles being driven around the USA, which have been discovered to be owned by Apple. The vehicles have cameras, LIDARs, and GPS antennas, and many are wondering is this an Apple Self-Driving Car? See also speculation from cult of Mac.

Here’s a video of the vehicle driving around the East Bay (50 miles from Cupertino), but they have also been seen in New York.

We don’t see the front of the vehicle, but it sure has plenty of sensors. On the front and back you can see two Velodyne 32E Lidars. These are 32 plane LIDARS that cost about $30K. You can also see two GPS antennas and what appears to be cameras in all directions. Unfortunately, you can’t see the front of the vehicle in these pictures, which is where the most interesting sensors will be.

So is this a robocar, or is this just a fancy mapping car? Rumours about Apple working on a car have been swirling for a while, but the absence of sightings of cars like this call them into question. You can’t have an active program without also testing cars on roads. There are ways to hide LIDARS and even cameras to a degree (and Apple is super secretive, so they might), but this vehicle hides little.

Most curious are the Velodynes. They are tilted down significantly. The 32E unit sees from about 10 degrees up to 30 degrees down. Tilting them this much means you don’t see out horizontally, which is not at all what you want if this is for a self-driving car. These LIDARs are densely scanning the road area close to the car, and higher items in the opposite direction. The rear LIDAR will be seeing out horizontally, but it’s placed just where you wouldn’t place it to see what’s in front of you. A GPS antenna is blocking the direct forward view, so if the goal of the rear LIDAR is to see ahead, it makes no sense.

We don’t see the front, so there might be another LIDAR up there, along with radars (often hidden in the grille) and these would be pretty important for any research car.

For mapping, these strange angles and blind spots are not an issue. You are trying to build a 3D and visible light scan of the world. What you don’t see from one point, you might see from another. For street mapping, what’s directly in front and behind is generally the road, and not especially interesting. But what’s to the side could be really interesting.

The car also has an accurate encoder on its wheel to give improved odemetry. Both robocars and mapping cars are interested in precise position information.

Evidence that this is a robocar:

  • The Velodynes are expensive, high-end and more than you need for mapping (though if cost is no object, they are a decent choice).
  • Apple knows it’s being watched, and might try to make their robocar look like a mapping car.
  • There are other sensors that we can’t see

Evidence that this is a mapping car:

  • As noted, the Velodynes are titled in a way that really suggests mapping. (Ford uses tilted ones, but paired with horizontal ones.)
  • The cameras are aimed at the corners, not forward.
  • They are driving in remote locations, which eventually you want to do, but initially you are more likely to get to the first stage close to home. Google has not done serious testing outside the Bay Area in spite of their large project.
  • The lack of streetview is a major advantage Google has over Apple, so it is not surprising they might make their own.

I can’t make a firm conclusion, but the evidence so far leans toward it being a mapping car. Seeing the front (which I am sure will happen soon) will tell us more. Another option is that it could be a mapping car building advanced maps for a different, secret, self-driving car.



tags: , ,


Brad Templeton, Robocars.com is an EFF board member, Singularity U faculty, a self-driving car consultant, and entrepreneur.
Brad Templeton, Robocars.com is an EFF board member, Singularity U faculty, a self-driving car consultant, and entrepreneur.





Related posts :



Robohub highlights 2025

  29 Dec 2025
We take a look back at some of the interesting blog posts, interviews and podcasts that we've published over the course of the year.

The science of human touch – and why it’s so hard to replicate in robots

  24 Dec 2025
Trying to give robots a sense of touch forces us to confront just how astonishingly sophisticated human touch really is.

Bio-hybrid robots turn food waste into functional machines

  22 Dec 2025
EPFL scientists have integrated discarded crustacean shells into robotic devices, leveraging the strength and flexibility of natural materials for robotic applications.

Robot Talk Episode 138 – Robots in the environment, with Stefano Mintchev

  19 Dec 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Stefano Mintchev from ETH Zürich about robots to explore and monitor the natural environment.

Artificial tendons give muscle-powered robots a boost

  18 Dec 2025
The new design from MIT engineers could pump up many biohybrid builds.

Robot Talk Episode 137 – Getting two-legged robots moving, with Oluwami Dosunmu-Ogunbi

  12 Dec 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Oluwami Dosunmu-Ogunbi from Ohio Northern University about bipedal robots that can walk and even climb stairs.

Radboud chemists are working with companies and robots on the transition from oil-based to bio-based materials

  10 Dec 2025
The search for new materials can be accelerated by using robots and AI models.



 

Robohub is supported by:




Would you like to learn how to tell impactful stories about your robot or AI system?


scicomm
training the next generation of science communicators in robotics & AI


 












©2025.05 - Association for the Understanding of Artificial Intelligence