Human factor a self-driving hitch

Experts fear boredom, wandering minds will raise crash risk

Kaushik Raghu, senior staff engineer at Audi, demonstrates an Audi self-driving vehicle on an interstate in Virginia on Friday. Automakers are developing automated systems that can operate cars in some or most circumstances, but they still require a driver’s attention.
Kaushik Raghu, senior staff engineer at Audi, demonstrates an Audi self-driving vehicle on an interstate in Virginia on Friday. Automakers are developing automated systems that can operate cars in some or most circumstances, but they still require a driver’s attention.

WASHINGTON -- Experts say the development of self-driving cars over the coming decade depends on an unreliable assumption by many automakers: that the humans in them will be ready to step in and take control if the car's system fails.

Instead, experience with automation in other modes of transportation like aviation and rail suggests that the strategy will lead to more deaths like that in May of a Tesla driver in Florida.

Decades of research shows that people have a difficult time keeping their minds on boring tasks like monitoring systems that rarely fail and hardly ever require them to take action. The human brain continually seeks stimulation. If the mind isn't engaged, it will wander until it finds something more interesting to think about. The more reliable the system, the more likely it is that attention will wane.

Automakers are in the process of adding increasingly automated systems that essentially drive cars in some or most circumstances, but still require the driver as a backup in case the vehicle encounters a situation unanticipated by its engineers.

Tesla's Autopilot, for example, can steer itself within a lane and speed up or slow down based on surrounding traffic or on the driver's set speed. It can change lanes with a flip of its signal, automatically apply brakes, or scan for parking spaces and parallel park on command.

Joshua Brown, a 40-year-old tech company owner from Canton, Ohio, who was an enthusiastic fan of the technology, was killed when neither he nor his Tesla Model S sedan's Autopilot braked for a truck making a left turn on a highway near Gainesville, according to federal investigators and the automaker.

Tesla warns drivers to keep their hands on the wheel even though Autopilot is driving, or the vehicle will slow to a stop automatically. A self-driving system Audi plans to introduce in its 2018 A7, which the company says will be the most advanced on the market, monitors drivers' head and eye movements, and automatically slows the car if the driver's attention is diverted.

But Brown's failure to brake means he either didn't see the truck in his path or saw it too late to respond -- an indication he was relying on the automation and his mind was elsewhere, said Missy Cummings, director of Duke University's Humans and Autonomy Laboratory. The truck driver said he had heard a Harry Potter video playing in the car after the crash.

"Drivers in these quasi- and partial modes of automation are a disaster in the making," Cummings said. "If you have to rely on the human to see something and take action in anything less than several seconds, you are going to have an accident like we saw."

Operators -- an airline pilot, a train engineer or car driver -- can lose awareness of the environment when they turn control over to automation, said Rob Molloy, the National Transportation Safety Board's chief highway-crash investigator.

He pointed to the crash of Air France Flight 447 into the Atlantic Ocean while flying from Brazil to France in 2007. A malfunction in equipment used to measure airspeed caused the plane's autopilot to disconnect, catching pilots by surprise. Confused, they caused an otherwise flyable plane to stall and fall from the sky, killing 228 people.

Planes and trains have had automation "for 20, 30 years and there are still times when they're like, 'Wow, we didn't expect that to happen,'" Molloy said.

Part of the problem is overconfidence in the technology causes people to think they can check out. Not long after Tesla introduced its Autopilot system, people were posting videos of cars with the self-driving mode engaged cruising down tree-lined roads or even highways with no one in the driver's seat. Brown, for example, had posted videos lauding the Autopilot system and demonstrating it in action.

"There is a tendency of people to take one ride in one of these vehicles and then conclude that because they have not crashed over the course of 10 minutes that the system must be ready," said Bryant Walker Smith, a University of South Carolina professor who studies the technology.

Some experts think the ability of people to monitor autonomous systems may be getting worse. With the advent of smartphones, people are accustomed to having their desire for mental stimulation satisfied immediately.

"Go into Starbucks, for example," said Cummings. "No one can just patiently wait in line, they're all doing something on their phones. It's kind of pathetic."

Some automakers may be rethinking their approach. Two years ago, General Motors announced that it would start selling a Cadillac in the fall of 2016 that almost would drive itself on freeways. But last week the company confirmed that the project has been delayed for an unspecified reason.

At previous briefings, company executives said they were waiting to perfect methods of assuring that the driver pays attention to the road even when the system is on.

The system, called "Super Cruise," will use cameras and radar to keep the car in the center of a lane and also stay a safe distance behind cars in front of it. The system will stop the car without driver action if traffic halts, and it can keep the car going in stop-and-go traffic. But it's designed for use only on limited-access divided highways.

Google, meanwhile, is aiming for a car that's fully self-driving and may not even have a steering wheel or brake pedals.

U.S. Transportation Secretary Anthony Foxx said Tuesday that he wants government regulators and the auto industry to work more closely together to test self-driving technology.

Foxx told about 1,200 people at a self-driving convention in San Francisco that a more rigorous review of robotic controls is needed to reassure consumers that autonomous vehicles are safe before they are cleared to hit the road.

He plans to propose guidelines for self-driving vehicles later this summer.

Information for this article was contributed by Tom Krisher of The Associated Press.

Business on 07/20/2016

Upcoming Events