Detroit — Eleven additional people were killed in U.S. crashes involving vehicles that were using automated driving systems during a four-month period earlier this year, according to newly released government data, part of an alarming pattern of incidents linked to the technology.
Ten of the deaths involved vehicles made by Tesla, though it is unclear from the National Highway Traffic Safety Administration’s data whether the technology itself was at fault or whether driver error might have been responsible. The 11th death involved a Ford pickup truck.
The deaths included four crashes involving motorcycles that occurred during the spring and summer: Two in Florida and one each in California and Utah. Safety advocates note that the deaths of motorcyclists in crashes involving Tesla vehicles using automated driver-assist systems such as Autopilot have been increasing.
The new fatal crashes are documented in a database that NHTSA is building in an effort to broadly assess the safety of automated driving systems, which, led by Tesla, have been growing in use. Tesla alone has more than 830,000 vehicles on U.S. roads with the systems. The agency is requiring auto and tech companies to report all crashes involving self-driving vehicles as well as autos with driver assist systems that can take over some driving tasks from people.
The 11 new fatal crashes, reported from mid-May through September, were included in statistics that the agency released Monday. In June, the agency released data it had collected from July of last year through May 15.
The figures that were released in June showed that six people died in crashes involving the automated systems, and five were seriously hurt. Of the deaths, five occurred in Teslas and one a Ford. In each case, the database says that advanced driver assist systems were in use at the time of the crash.
Michael Brooks, executive director of the nonprofit Center for Auto Safety, said he is baffled by NHTSA’s continued investigations and by what he called a general lack of action since problems with Autopilot began surfacing back in 2016.
“I think there’s a pretty clear pattern of bad behavior on the part of Tesla when it comes to obeying the edicts of the (federal) safety act, and NHTSA is just sitting there,” he said. “How many more deaths do we need to see of motorcyclists?”
Brooks noted that the Tesla crashes are victimizing more people who are not in the Tesla vehicles.
“You’re seeing innocent people who had no choice in the matter being killed or injured,” he said.
A message was left Tuesday seeking a response from NHTSA.
Tesla’s crash number may appear elevated because it uses telematics to monitor its vehicles and obtain real-time crash reports. Other automakers lack such capability, so their crash reports may emerge more slowly or may not be reported at all, NHTSA has said.
NHTSA has been investigating Autopilot since August of last year after a string of crashes since 2018 in which Teslas collided with emergency vehicles parked along roadways with flashing lights on. That investigation moved a step closer to a recall in June, when it was upgraded to what is called an engineering analysis.
In documents, the agency raised questions about the system, finding that the technology was being used in areas where its capabilities are limited and that many drivers weren’t taking steps to avoid crashes despite warnings from the vehicle.
NHTSA also reported that it has documented 16 crashes in which vehicles with automated systems in use hit emergency vehicles and trucks that were displaying warning signs, causing 15 injuries and one death.
The National Transportation Safety Board, which also has investigated some of the Tesla crashes dating to 2016, has recommended that NHTSA and Tesla limit Autopilot’s use to areas where it can safely operate. The NTSB also recommended that NHTSA require Tesla to improve its systems to ensure that drivers are paying attention. NHTSA has yet to act on the recommendations. The NTSB can make only recommendations to other federal agencies.
Messages were left Tuesday seeking comment from Tesla. At the company’s artificial intelligence day in September, CEO Elon Musk asserted that, based on the rate of crashes and total miles driven, Tesla’s automated systems were safer than human drivers — a notion that some safety experts dispute.
“At the point of which you believe that adding autonomy reduces injury and death, I think you have a moral obligation to deploy it,” Musk said. “Even though you’re going to get sued and blamed by a lot of people. Because the people whose lives you saved don’t know that their lives were saved. And the people who do occasionally die or get injured, they definitely know, or their state does, that it was, whatever, there was a problem with Autopilot.”
Teslas with automated systems have driven more than 3 million vehicles on the road, Musk said.
“That’s a lot of miles driven every day. And it’s not going to be perfect. But what matters is that it is very clearly safer than not deploying it.”
In addition to Autopilot, Tesla sells “Full Self-Driving” systems, though it says the vehicles cannot drive themselves and that motorists must be ready to intervene at all times.
The number of deaths involving automated vehicles is small compared with the overall number of traffic deaths in the U.S. Nearly 43,000 people were killed on U.S. roads last year, the highest number in 16 years, after Americans returned to the roads as the pandemic eased. Authorities blamed reckless behavior such as speeding and driving while impaired by drugs or alcohol for much of the increase.