Self-Driving Teslas Are Fatally Striking Motorcyclists More Than Any Other Brand: New Analysis
-
Unless it's a higher rate than human drivers per mile or hours driven I do not care. Article doesn't have those stats so it's clickbait as far as I'm concerned
-
The fact that the other self driving brands logged zero motorcyclist fatalities means the technology exists to prevent more deaths. Tesla has chosen to allow more people to die in order to reduce cost. The families of those five dead motorcyclists certainly care.
-
I'm not going to lie, I almost had a stroke writing it...
-
Thanks, 'Satan.
Do you know the number of miles driven by Tesla's self-driving tech? Because I don't, Tesla won't say, they're a remarkably non-transparent company where their tech is concerned. Near as I can tell, nobody does (other than folks locked up tight with NDAs). If the ratio of accidents-per-mile-driven looked good, you know as a flat fact that Elon would be Tweeting all about it.
Sorry you didn't find the death of 5 Americans newsworthy. I'll try harder for the next one.
-
I upvoted every comment in this sub-thread shitshow and hated all of it.
-
Yer welcome!
-
That's not good though, right? "We have the technology to save lives, it works on all of our cars, and we have the ability to push it to every car in the fleet. But these people haven't paid extra for it, so..."
-
As someone who likes the open sky feeling, this is why I drive a convertible instead.
-
In this case, does it matter? Both are supposed to follow a vehicle at a safe distance
-
the cybertruck is sharp enough to cut a deer in half, surely a biker is just as vulnerable.
-
they originally had lidar, or radar, but musk had them disabled in the older models.
-
Same goes for the other vehicles. They didn’t even try to cover miles driven and it’s quite likely Tesla has far more miles of self-driving than anyone else.
I’d even go so far as to speculate the zero accidents of other self-driving vehicles could just be zero information because we don’t have enough information to call it zero
-
I wonder if a state court judge could mandate its use as unsafe?
-
Well, only 1 or 2 of those were in a time frame where I'd consider FSD superior to AP, it's a more recent development where that's likely the case.
But to your point, at some point I expect Tesla to use the FSD software for AP for the exact reasons you mentioned. My guess is they'd just do something like disable making turns, so you wouldn't be able to use it outside of straight stretches like AP today.
-
Cybertrucks have 17 times the mortality rate of the ford pinto.
https://www.motherjones.com/politics/2025/02/report-cybertruck-safety-ford-pinto/
-
Why not? It’s got multiple cameras so could judge distances the same way humans do.
However there have been both hardware and software updates since most of those, so the critical question is how much of a problem is it still? The article had no info or speculation on that
-
In this case, does it matter? Both are supposed to follow a vehicle at a safe distance
I think it does matter, while both are supposed to follow at safe distances, the FSD stack is doing it in a completely different way. They haven't really been making any major updates to AP for many years now, all focus has been on FSD.
AP is looking at the world frame by frame, each individual camera on it's own, while FSD is taking the input of all cameras, turning into 3d vector space, and then driving based off that. Doing that on city streets and highways is only a pretty recent development. Updates for doing it this way on highway and streets only went out to all cars in the past few months. For along time it was on city streets only.
I’d be more interested in how it changes over time, as new software is pushed.
I think that's why it's important to make a real distinction between AP and FSD today (and specifically which FSD versions)
They're wholly different systems, one that gets older every day, and one that keeps getting better every few months. Making an article like this that groups them together muddies the water on what / if any progress has been made.
-
I mean yeah, I just said above that someone almost killed me. They were probably a human driver. But that's a "might happen, never know." If self driving cars are rear-ending people, that's an inherent artifact of it's programming, even though it's not intentionally programmed to do that.
So it's like, things were already bad. I already do not feel safe doing any biking anymore. But as self driving cars become more prevalent, that threat upgrades to a kind of defacto, "Oh, these vast stretches of land are places where only cars and trucks are allowed. Everything else is roadkill waiting to happen."