Tesla Inc. failed to fix limitations in its Autopilot system following a gruesome Florida crash that killed a driver in 2016, company engineers said in a family’s lawsuit over a very similar 2019 fatal collision that’s headed to a jury trial.
(Bloomberg) — Tesla Inc. failed to fix limitations in its Autopilot system following a gruesome Florida crash that killed a driver in 2016, company engineers said in a family’s lawsuit over a very similar 2019 fatal collision that’s headed to a jury trial.
The electric-car maker didn’t make any changes to its driver-assistance technology to account for crossing traffic in the nearly three years between two high-profile accidents that killed Tesla drivers whose cars slammed into the side of trucks, according to newly revealed testimony from multiple engineers.
After years of touting autonomous driving as the way of the future, Tesla and Chief Executive Officer Elon Musk are under legal pressure from consumers, investors, regulators and federal prosecutors who are questioning whether the company has over-hyped its progress toward self-driving vehicles during the last eight years. Tesla also is in the cross-hairs of multiple investigations by the National Highway Traffic Safety Administration over possible defects in Autopilot linked to at least 17 deaths since June 2021.
The trial set for October, the first for the company over a death blamed on Autopilot, will pit Musk’s repeated assertion that Teslas are the safest cars ever made against technology experts expected to testify that the company’s marketing has lulled drivers into a false sense of security.
Read More: Tesla Fatal-Crash Lawsuit to Test Musk’s Autopilot Claims
Tesla’s attorneys representing the company didn’t immediately respond to requests for comment.
The automaker contends it has been transparent about Autopilot’s limitations, including challenges with detecting traffic crossing in front of its cars. Tesla warns in its owner’s manual and car screens that drivers must be alert and ready to take control of vehicles at any time.
Tesla prevailed earlier this year in its first trial over a non-fatal Autopilot crash when a Los Angeles jury cleared the company of wrongdoing over a woman’s claim that the driver-assistance feature in her Model S caused her to veer into the center median of a city street.
The case set to be presented to a jury in Palm Beach County, Florida, was brought by the family of Jeremy Banner, a 50-year-old father of three who had switched on Autopilot 10 seconds before his Model 3 plowed into the underbelly of a tractor-trailer in 2019. An investigation by the National Transportation Safety Board found that Banner probably didn’t see the truck crossing a two-lane highway on his way to work. Autopilot apparently didn’t see it either.
Despite the company’s knowledge “that there’s cross traffic or potential for cross traffic, the Autopilot at the time was not designed to detect that,” according to testimony given in 2021 by company engineer Chris Payne that was excerpted in a recent court filing. Engineer Nicklas Gustafsson provided a similar account in a 2021 deposition.
Last week, Banner’s widow revised her complaint to seek punitive damages, raising the stakes for Tesla at trial. She argues the company should have re-programmed Autopilot so that it would shut off in dangerous circumstances after Tesla driver Joshua Brown crashed into the side of a truck in 2016.
“There is evidence in the record that the defendant Tesla engaged in intentional misconduct and/or gross negligence for selling a vehicle with an Autopilot system which Tesla knew to be defective and knew to have caused a prior fatal accident,” the Banner family said in the amended complaint.
Read More: Tesla Can’t Perfect Autopilot Without a Few Deadly Crashes
One of the expert witnesses retained by the Banner family brought is Mary “Missy” Cummings, who recently served as an adviser to the National Highway Traffic Safety Administration. Cummings, a Duke University professor and vocal skeptic of Autopilot, said in a court filing that Tesla “is guilty of intentional misconduct and gross negligence” for failing to test and enhance Autopilot between the Brown and Banner crashes.
Tesla made “public statements that its Autopilot technology is far more capable than it actually is,” Cummings wrote.
Trey Lytal, a lawyer representing the Banner family, didn’t immediately respond to a request for comment.
More stories like this are available on bloomberg.com
©2023 Bloomberg L.P.