Tesla Autopilot Was Uniquely Risky—and May Still Be
A federal report published today found that Teslaâs Autopilot system was involved in at least 13 fatal crashes in which drivers misused the system in ways the automaker should have foreseenâand done more to prevent. Not only that, but the report called out Tesla as an âindustry outlierâ because its driver assistance features lacked some of the basic precautions taken by its competitors. Now regulators are questioning whether a Tesla Autopilot update designed to fix these basic design issues and prevent fatal incidents has gone far enough.
These fatal crashes killed 14 people and injured 49, according to data collected and published by the National Highway Traffic Safety Administration, the federal road-safety regulator in the US.
At least half of the 109 âfrontal planeâ crashes closely examined by government engineersâthose in which a Tesla crashed into a vehicle or obstacle directly in its pathâinvolved hazards visible five seconds or more before impact. Thatâs enough time that an attentive driver should have been able to prevent or at least avoid the worst of the impact, government engineers concluded.
In one such crash, a March 2023 incident in North Carolina, a Model Y traveling at highway speed struck a teenager while he was exiting a school bus. The teen was airlifted to a hospital to treat his serious injuries. The NHTSA concluded that âboth the bus and the pedestrian would have been visible to an attentive driver and allowed the driver to avoid or minimize the severity of this crash.â
Government engineers wrote that, throughout their investigation, they âobserved a trend of avoidable crashes involving hazards that would have been visible to an attentive driver.â
Tesla, which disbanded its public affairs department in 2021, did not respond to a request for comment.
Damningly, the report called Tesla âan industry outlierâ in its approach to automated driving systems. Unlike other automotive companies, the report says, Tesla let Autopilot operate in situations it wasnât designed to, and failed to pair it with a driver engagement system that required its users to pay attention to the road.
Regulators concluded that even the Autopilot product name was a problem, encouraging drivers to rely on the system rather than collaborate with it. Automotive competitors often use âassist,â âsense,â or âteamâ language, the report stated, specifically because these systems arenât designed to fully drive themselves.
Last year, California state regulators accused Tesla of falsely advertising its Autopilot and Full Self-Driving systems, alleging that Tesla misled consumers into believing the cars could drive themselves. In a filing, Tesla said that the stateâs failure to object to the Autopilot branding for years constituted an implicit approval of the carmakerâs advertising strategy.
The NHTSAâs investigation also concluded that, compared to competitorsâ products, Autopilot was resistant when drivers tried to steer their vehicles themselvesâa design, the agency wrote in its summary of an almost two-year investigation into Autopilot, that discourages drivers from participating in the work of driving.
These crashes occurred before Tesla recalled and updated its Autopilot software via an over-the-air update earlier this year. But along with closing this investigation, regulators have also opened a fresh probe into whether the Tesla updates, pushed in February, did enough to prevent drivers from misusing Autopilot, from misunderstanding when the feature was actually in use, or from using it in places where it is not designed to operate.
The review comes after a Washington state driver last week said his Tesla Model S was on Autopilotâwhile he was using his phoneâwhen the vehicle struck and killed a motorcyclist.
The agency will now also look into Teslaâs statements that drivers can opt in to parts of the recall fix, and are easily able to reverse parts of it.
According to Tesla, the post-recall fixes included stricter driver attentiveness requirements while using Autopilot, larger driver-monitoring alerts (âPlease pay attention to the roadâ), and a suspension policy that restricts driversâ use of the feature if the system finds they are using it improperly.
âTesla should get serious about driver monitoring, and should limit its use to roads itâs built to work on, for crying out loud,â says Phil Koopman, an engineering professor at Carnegie Mellon University whose research includes self-driving car safety. Without more serious intervention, Koopman says, a cycle of investigations and recalls could continue for years.
Car safety experts with the publication Consumer Reports found in February that Teslaâs recall fixes did not prevent Autopilot misuse.
The new probe comes at a bad time for the electric-auto maker, which is facing its worst sales and revenue growth numbers in years. Tesla also revealed in regulatory filings last fall that it faces an investigation by the US Department of Justice into its Autopilot feature and how it has represented the driving ranges of its battery-powered vehicles.
The timing is also exceedingly awkward for Tesla because itâs pushing even further into autonomy. On an earnings call with investors Tuesday, CEO Elon Musk seemed unperturbed by the companyâs recent downturn, focusing instead on Tesla's work in autonomy and its plans to operate a fully autonomous ride-hail service. âReally, we should be thought of as an AI or robotics company,â he said. âIf you value Tesla as just like an auto company ⦠fundamentally, it's just the wrong framework.â
Musk has said the company will hold an unveiling event for a purpose-built Tesla robotaxi, called the Cybercab, in August. Just this week, Tesla slashed the price of its more advanced driver assistance feature, called Full Self-Driving, by a third, to $8,000. The company began offering car customers free 30-day trials of FSD last month. Now, its original, flagship automated feature is under new scrutiny.
© 2024 Condé Nast. All rights reserved. WIRED may earn a portion of sales from products that are purchased through our site as part of our Affiliate Partnerships with retailers. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of Condé Nast. Ad Choices