Elon Musk Was Right: Cheap Cameras Could Replace Lidar on Self-Driving Cars, Researchers Find

Illustration for article titled Elon Musk Was Right: Cheap Cameras Could Replace Lidar on Self-Driving Cars, Researchers Find
Screenshot: Cornell University

Among the many tidbits of wisdom that Elon Musk dropped at a Tesla company investor event on Monday was the revelation that Lidar, a laser-based scanning technology that images objects in 3D, was “friggin’ stupid,” and that “...anyone relying on LiDAR is doomed.” It seemed a grandiose claim given how many autonomous car initiatives rely on the tech, but Cornell researchers have just backed up Musk’s predictions with a new method for self-driving cars to see the world in 3D using a pair of cheap cameras.

Being able to visualize and detect objects around a vehicle in three dimensions is crucial for autonomous cars to safely operate in a world where roads are shared with other vehicles, cyclists, and often pedestrians. As a driver, every time you turn your head to scan what’s around your car, your brain is instantly visualizing your surroundings in 3D and assessing potential hazards. Using cheap sensors to simply detect objects near a self-driving car isn’t enough. When it’s cruising down the road at 60 MPH, it needs to see what’s ahead and be able to plan for avoiding hazards.

That’s why you’ll often see Lidar (Light Detection and Ranging) systems perched atop autonomous vehicles. Using spinning lasers they scan a vehicle’s surroundings and generate 3D images of objects near and far, allowing the software to analyze the results and pinpoint things to avoid. Lidar’s expensive, though, often adding $10,000 worth of components to a car’s price tag, and it needs to be perched atop a vehicle for the best vantage point. In a time when we’re trying to maximize the range of both gas and electric vehicles, a Lidar upgrade adds a lot of drag to a car’s aerodynamics and its performance.


In a paper that will be presented at the 2019 Conference on Computer Vision and Pattern Recognition in June, Pseudo-LiDAR from Visual Depth Estimation: Bridging the Gap in 3D Object Detection for Autonomous Driving, Cornell researchers detail a potential breakthrough for autonomous vehicles. Cameras have typically been considered an inferior technology to Lidar given that they’re often installed at low angles, near a vehicle’s bumper, resulting in images that tend to distort objects in the distance which confuses neural networks trying to process and interpret the data.

But by placing a pair of cheap cameras on either side of a vehicle behind its windshield, stereoscopic images are produced which can be converted to 3D data. Because the images are being generated from a higher vantage point, closer to where Lidar systems are typically installed, the 3D data that was generated from the cameras was found to be nearly as precise as what laser scanners are able to generate, without distortion, and at a fraction of the cost.

It will probably be a long time before this research makes its way into self-driving vehicles, however. Lidar is still reliable and incredibly accurate, and companies working on autonomous vehicles are more concerned about safety and liabilities at the moment, instead of costs. But as the technologies improve, the software improves, and restrictions limiting where and when autonomous cars can roam are lifted, self-driving will soon be a big selling point for consumers buying new vehicles—and they do care about costs. Cornell’s approach will make it much cheaper to implement self-driving features on a car, and it could eventually make Lidar obsolete. So maybe Musk was right?

3D point cloud data extracted from images generated by a car’s stereo cameras are rotated to provide a more accurate bird’s eye view of an autonomous vehicle’s surroundings.
Gif: Cornell University

Update, April 25, 11:06 a.m. EST/EDT: The Cornell researchers reached out to us to clarify how their solutions differs to other approaches when it comes to visualizing the surroundings of an autonomous vehicle. The images from the stereo cameras are used to generate a 3D point cloud which is then rotated in 3D (as demonstrated in the above GIF) to produce a top-down perspective of a vehicle’s surroundings. This allows for improved accuracy that puts their approach on par with Lidar solutions.

Share This Story

Get our `newsletter`


Any word on how these systems handle the hay bale in the road problem? You’re going down the road and you can see through the windshield of the cars in front of you that there’s a big ass hay bale in the road and you need to dodge via change lanes or slow down. An AI powered car doesn’t see that hay bale (yet, I guess) and has a much smaller window of opportunity to make that decision. Granted it can make that decision much faster than any human but sometimes physics wins...

The point I’m making is that there is a lot of intuition with driving and I wonder how AI driving systems account for that.

In the far future where every car has driving systems I don’t see this being an issue. With interconnectedness a car miles up the road can spot a hazard, relay that information to your vehicle, and decisions be made before you ever see the problem.