Tesla Crashes Highlight ‘Black Box’ Challenge for Investigations

Two collisions involvingTesla Inc. vehicles left three people dead on Sunday, and police in Indiana and California are still trying to answer an unfamiliar question: who, or what, was in control at the time of the crash?

A Tesla Model S ran a red light shortly after midnight in Gardena, California, and slammed into a Honda Civic, killing both passengers inside the latter car. Hours later, a Tesla Model 3 slammed into a fire truck parked across the left-hand lane of an Indiana highway, killing a woman and injuring her husband.

Police have yet to determine whether the drivers were using Tesla’s suite of automated driver-assist features known as Autopilot, which havedrawn the interest of federal safety officials, or if they were manually driving their vehicles. They say they expect to learn that in the coming days.

The distinction will be important in determining the cause of the crashes and in answering questions regulators have posed about the industry’s rapid shift to vehicles that take over some of the functions of a driver.

Conventional vehicles have for years been outfitted with so-called event data recorders, the “black box” that logs critical information needed to piece together a crash sequence, such as braking, air bag deployment and other measurements that can be downloaded via widely available tools.

Data detailing the performance of automated-driving technologies in the seconds before a crash, however, are only accessible by vehicle manufacturers. That means investigators seeking to understand whether an automated system was active during a crash must turn to the company that manufactured the car.

“We should not ever have to rely on the manufacturer to translate this sort of data, because they are potentially liable for product defects and they have an inherent conflict of interest,” saidSam Abuelsamid, principal analyst for Navigant Research in Detroit.

He said a standard method should be in place to determine what control modes are active before and during a crash.

“As we deploy more vehicles with these partial automation systems, it should be mandatory to record information about the state of automation and the driver,” he said.

After a 2016 Tesla crash, theNational Transportation Safety Board called for the Transportation Department to define what data should be collected, but this hasn’t happened yet.

“As more manufacturers deploy automated systems on their vehicles, to improve system safety, it will be necessary to develop detailed information about how the active safety systems performed during, and how drivers responded to, a crash sequence,” the NTSB warned in its report onthe incident. “Manufacturers, regulators, and crash investigators all need specific data in the event of a system malfunction or crash.”

Tesla did not respond to emails seeking comment. The company has said drivers are ultimately responsible for controlling their vehicle while using Autopilot, and that they must remain attentive with their hands on the wheel at all times. Tesla has also pushed back forcefully against criticism that the system is unsafe, pointing to quarterly data released by the company that it says show drivers using Autopilot are safer than those operating without it.

TheNational Highway Traffic Safety Administration said it was reviewing the crash in California. An agency spokesman declined to comment on whether Autopilot was suspected in that crash.

Raul Arbelaez, vice president of theInsurance Institute for Highway Safety’s Vehicle Research Center, said the the inability to readily access that data impedes the ability of researchers to understand how automated driver aids are performing in the field, especially in less-severe crashes that account for the vast majority of traffic collisions.

“How do people interact with these technologies? What conditions do they tend to work in? Do they work really poorly at night with snow and rain, or are they excellent under those conditions?” he said.

“I’m sure it is very useful for the auto manufacturers to help them improve their products down the line, but in terms of understanding how the current fleet is performing, we really don’t have access to that in these crashes that are happening very quickly without working with the manufacturers.”

That was the case in 2016, when the NTSB and NHTSAprobed the first fatal crash involving Tesla’s Autopilot system. In that collision, a 2016 Tesla Model S driver was killed after driving under a semi-trailer that was crossing a highway, shearing off the Tesla’s roof.

Florida Crash

The vehicle involved in that Florida crash didn’t have a traditional event data recorder that could be read by widely available tools, the NTSB said. Instead, the Tesla vehicle collected substantial data that the company provided to the NTSB that revealed the car’s driver was using Autopilot at the time of the crash.

And even if the Tesla did have an event data recorder, the 15 data parameters required by the 2006 rule governing the devices “are inadequate to comprehend even the simplest questions of who/what controlled an automated vehicle at the time of a crash,” the NTSB wrote in the report detailing its investigation.

Since at least 2018, Tesla vehicles have had event data recorders and the company hasoffered tools to the public to download crash data.

The NTSB is investigating other Tesla vehicles crashes that occurred while Autopilot was in use, including aMarch 2018 fatal collision in Mountain View, California. The NHTSA, meanwhile, has launched probes into 13 crashes that it believes may have occurred while drivers were using Autopilot, including a Dec. 7 crashin Connecticut in which a Tesla driver rear-ended a parked police cruiser.

Real-World Data

The number of inquiries opened by the agency shows it is taking interest in how the new technology is being used in the field, said Frank Borris, a former director of the Office of Defects Investigation at NHTSA. Borris said the role of agency crash investigators is to gather real-world crash examples to better understand about areas of potential traffic safety risk, including new technologies.

Borris said he has long been concerned that Tesla dubbing its system “Autopilot” could lead drivers to rely on it too heavily and in scenarios for which it is not designed to operate. He also said little empirical data exists documenting real-world performance of Autopilot and other automated-driver assist technologies.

Meanwhile, the need for data by investigators is likely to grow more urgent as automated systems become more widespread and advanced.

InApril 2019, during an Autonomy Day for investors at the company’s Palo Alto, California headquarters, Tesla Chief Executive Elon Musk said that will deploy its “first operating robotaxis — with no one in them — next year.”

During anearnings call in October, Musk said Tesla should be able to upload software enabling its cars to become a robotaxis by the end of 2020.

He clarified that acceptance by regulatory authorities will vary. “But that transition, that sort of flipping the switch from a car that is from not robotaxi to robotaxi, I think, will probably be the biggest step change increase inasset value in history by far,” he said.

— With assistance by Dana Hull

Source: Read Full Article

Comments

Popular posts from this blog

Here's what you can get for $800,000 in Charlotte, North Carolina

Why Late-Deciding Democratic Voters Swung To Joe Biden

WhatsApp creates secret 'self-deleting messages' feature – wiping texts forever