Columnists

Testing driverless cars

AS the computers that operate driverless cars digest the rules of the road, some engineers think it might be nice if they can learn from mistakes made in virtual reality rather than on real streets.

Companies like Toyota, Uber and Waymo have discussed at length how they are testing autonomous vehicles, but what is not known is that they are also testing vehicles inside computer simulations.

Virtual cars, equipped with the same software as the real thing, spend thousands of hours driving in digital worlds.

Think of it as a way of identifying flaws in the way the cars operate without endangering real people. If a car makes a mistake on a simulated drive, engineers can tweak its software accordingly, laying down new rules of behaviour.

Researchers are also developing methods that would allow cars to actually learn new behaviour from these simulations, gathering skills more quickly than human engineers could ever lay them down with explicit software code.

“Simulation is a tremendous thing,” said Gill Pratt, chief executive of the Toyota Research Institute, one of the artificial intelligence labs exploring this kind of virtual training for autonomous vehicles and other robotics.

These methods are part of a sweeping effort to accelerate the development of autonomous cars through machine learning.

When Google designed its first self-driving cars nearly a decade ago, engineers built most of the software line by line, carefully coding each tiny piece of behaviour. But increasingly, thanks to recent improvements in computing power, autonomous carmakers are embracing complex algorithms that can learn tasks on their own, like identifying pedestrians on the roadways or predicting events.

“This is why we think we can move fast,” said Luc Vincent, who recently started an autonomous vehicle project at Lyft, Uber’s main rival.

“This stuff didn’t exist 10 years ago when Google started.”

There are still questions hanging over this research.

Most notably, because these algorithms learn by analysing more information than any human ever could, it is sometimes difficult to audit their behaviour and understand why they make particular decisions.

But, in the years to come, machine learning will be essential to the continued progress of autonomous vehicles.

Today’s vehicles are not nearly as autonomous as they may seem.

After 10 years of research, development and testing, Google’s cars are poised to offer public rides on the streets of Arizona.

Waymo, which operates under Google’s parent company, is preparing to start a taxi service near Phoenix, and unlike other services, it will not put a human behind the wheel as a backup.

But, its cars will still be on a tight leash. Waymo and many of its rivals have embraced deep neural networks, complex algorithms that can learn tasks by analysing data.

By analysing photos of pedestrians, for example, a neural network can learn to identify a pedestrian. These kinds of algorithms are also helping to identify street signs and lane markers, predict what will happen next on the road, and plan routes forward.

The trouble is that this requires enormous amounts of data collected by cameras, radar and other sensors that document real-world objects and situations.

And, humans must label this data, identifying pedestrians, street signs and the like.

Gathering and labeling data describing every conceivable situation is an impossibility. Data on accidents, for instance, is hard to come by. This is where simulations can help.

Recently, Waymo unveiled a roadway simulator it calls Carcraft.

Today, the company said, this simulator provides a way of testing its cars at a scale that is not possible in the real world.

Its cars can spend far more time on virtual roads than the real thing. Presumably, like other companies, Waymo is also exploring ways its algorithms can learn new behaviour from this kind of simulator.

Pratt said Toyota is using images of simulated roadways to train neural networks, and this approach has yielded promising results.

In other words, the simulations are similar enough to the physical world to reliably train the systems that operate the cars.

Part of the advantage with a simulator is that researchers have complete control over it.

They need not spend time and money labeling images — and potentially making mistakes with these labels.

“You have ground truth,” Pratt explained.

“You know where every car is. You know where every pedestrian is. You know where every cyclist is. You know the weather.”

Bridging the gap between the virtual and the physical is no easy task, Pratt said. And, companies must also ensure that algorithms don’t learn unexpected or harmful behaviour while learning on their own. That is a big worry among artificial intelligence researchers.

For this and other reasons, companies like Toyota and Waymo are not building these cars solely around machine learning. They also hand-coded software in more traditional ways in an effort to guarantee certain behaviour. Waymo cars don’t learn to stop at stop lights, for example. There is a hard and fast rule that they stop.

But, the industry is headed towards more machine learning, not less. It provides a better way to train the car to do tasks like identifying lane makers, said Waymo’s vice president of engineering, Dmitri Dolgov.

But, it becomes even more important, he explained, when a car needs a much deeper understanding of the world around it.

“Robotics and machine learning go hand in hand,” he said.

Cade Metz is a reporter withThe New York Times, covering artificial intelligence, driverless cars, robotics and other ways the world is changing.

Most Popular
Related Article
Says Stories