A representative for Tesla sent Ars the following statement: “Today’s verdict is wrong and only works to set back automotive safety and jeopardize Tesla’s and the entire industry’s efforts to develop and implement life-saving technology. We plan to appeal given the substantial errors of law and irregularities at trial. Even though this jury found that the driver was overwhelmingly responsible for this tragic accident in 2019, the evidence has always shown that this driver was solely at fault because he was speeding, with his foot on the accelerator—which overrode Autopilot—as he rummaged for his dropped phone without his eyes on the road. To be clear, no car in 2019, and none today, would have prevented this crash. This was never about Autopilot; it was a fiction concocted by plaintiffs’ lawyers blaming the car when the driver—from day one—admitted and accepted responsibility.”
So, you admit that the company’s marketing has continued to lie for the past six years?
There are other cars on the market that use technology that will literally override your input if they detect that there is a crash imminent. Even those cars do not claim to have autopilot and Tesla has not changed their branding or wording which is a lot of the problem here.
I can’t say for sure that they are responsible or not in this case because I don’t know what the person driving then assumed. But if they assumed that the “safety features” (in particular autopilot) would mitigate their recklessness and Tesla can’t prove they knew about the override of such features, then I’m not sure the court is wrong in this case. The fact that they haven’t changed their wording or branding of autopilot (particularly calling it that), is kind of damning here.
Autopilot maintains speed and heading or flight path in planes. But the average person doesn’t know or understand that. Tesla has been using the pop culture understanding of what autopilot is and that’s a lot of the problem. Other cars have warning about what their “assisted driving” systems do, and those warnings pop up every time you engage them before you can set any settings etc. But those other car manufacturers also don’t claim the car can drive itself.
Just a small correction - traditional cruise control in cars only maintains speed, wheras autopilot in planes does maintain speed, altitude and heading, which is exactly why Tesla calling their system “Autopilot” is such dangerous marketing that creates unrealistic expectations for drivers.
To me having the car be able to override your actions sounds more dangerous than being to override the autopilot.
I had one rental truck that drove me insane and scared the shit out of me because it would slam on the brakes when I tried to reverse into grass that was too tall.
What if I were trying to avoid something dangerous, like a train or another vehicle, and the truck slammed on the brakes for me because of some tree branches in the way? Potentially deadly.