this post was submitted on 03 Sep 2023
210 points (100.0% liked)

Technology

37717 readers
457 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 1 year ago (2 children)

The AV-sceptics are really hard to please…

I do not care about situations that they work in, I care about what situations they will fail at. That's what matters and that's what no company will tell you. As said, we learn about the capabilities of self driving cars one crash at a time, and that's just unacceptable when you could figure all of that out years ago in simulation.

So far none of the self-driving incidences I have seen were some kind of unforeseen freak situation, it was always some rare, but standard thing, fog, pedestrian crossing the road, road blocked by previous crash, etc.

[–] [email protected] 6 points 1 year ago (1 children)

Humans get into accidents all the time. Is that not unacceptable for you?

I feel like people apply standards to self driving cars that they don't to human driven ones. It's unreasonable to expect a self driving system never to fail. It's unreasonable to imagine you can just let it practice in simulation untill it's perfect. This is what happens when you just narrowly focus on one aspect of self driving cars (individual accidents) - you miss the big picture.

[–] [email protected] 1 points 1 year ago (2 children)

I feel like people apply standards to self driving cars that they don’t to human driven ones.

Human drivers need to pass driving test, self-driving cars do not. Human drivers also have a baseline of common sense that self-driving cars do not have, so they really would need more testing than humans, not less.

It’s unreasonable to expect a self driving system never to fail.

I don't expect them to never fail, I just want to know when they fail and how badly.

It’s unreasonable to imagine you can just let it practice in simulation untill it’s perfect.

What's unreasonable about that?

individual accidents

They are only "individual" because there aren't very many self-driving cars and because not every fail ends up deadly.

[–] [email protected] 6 points 1 year ago* (last edited 1 year ago) (1 children)

Tesla on FSD could easily pass the driving test that's required for humans. That's a nonsensical standard. Most people with fresh license are horribly incompetent drivers.

[–] [email protected] 1 points 1 year ago (1 children)

Tesla on FSD could easily pass the driving test that’s required for humans

So why don't we check it? Right now we are blindly trusting the claims of companies.

[–] [email protected] 2 points 1 year ago (1 children)

What are these claims we're blindly trusting exaclty? Do you have any direct quotes?

[–] [email protected] 3 points 1 year ago

Have you used it? It's not very good. It tries to run red lights, makes random swerves and inputs, and generally drives like someone on sedatives.

They've had to inject a ton of map data to try to make up for the horrendously low resolution cameras, but "HD MaPs ArE a CrUtCh" right?

No radar or lidar means the sun can blind it easily, and there's a blind spot in front of the car where cameras cannot see.

Is what they've made impressive? Sure, but it's nowhere near safe enough to be on public roads in customer's cars. At all.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago) (1 children)

I don’t expect them to never fail, I just want to know when they fail and how badly.

"Over 6.1 million miles (21 months of driving) in Arizona, Waymo’s vehicles were involved in 47 collisions and near-misses, none of which resulted in injuries"

How many human drivers have done millions of miles of driving before they were allowed to drive unsupervised? Your assertion that these systems are untested is just wrong.

"These crashes included rear-enders, vehicle swipes, and even one incident when a Waymo vehicle was T-boned at an intersection by another car at nearly 40 mph. The company said that no one was seriously injured and “nearly all” of the collisions were the fault of the other driver."

According to insurance companies, human driven cars have 1.24 injuries per million miles travelled. So, if Waymo was "as good as a typical human driver" then there would have been several injuries. They had zero serious injuries.

The data (at least from reputable companies like Waymo) is absolutely available and in excruciating detail. Go look it up.

[–] [email protected] 0 points 1 year ago

The data (at least from reputable companies like Waymo) is absolutely available and in excruciating detail. Go look it up.

As already said, I want to know where they fail, preferably in the simulator, not on actual roads. Having vehicles drive in circles on carefully selected roads and making a lot of miles is no big accomplishment and not comparable with humans that have to drive on all the roads under all the conditions.

[–] [email protected] 4 points 1 year ago (1 children)

As a software developer, that’s not how testing works. QA is always trying to come up with weird edge cases to test, but once it’s out in the wild with thousands (or more) of real-world users, there’s always going to be something nobody ever tried to test.

For example, there was a crash where an unmarked truck with exactly the same color as the sky was 90° sideways on the highway. This is just something you wouldn’t think of in lab conditions.

[–] [email protected] 2 points 1 year ago (1 children)

there’s always going to be something nobody ever tried to test.

That's not what is happening. We don't see weird edge cases, we see self driving cars blocking emergency vehicles and driving through barriers.

For example, there was a crash where an unmarked truck with exactly the same color as the sky was 90° sideways on the highway.

The sky is blue and the truck was white. Testing the dynamic range of the camera system is absolutely something you do in in lab situation. And a thing blocking the road isn't exactly unforeseen either.

Or how about railroad crossing, Tesla can't even the difference between a truck and a train. Trucks blipping in out of existence, even changing direction, totally normal for Tesla too.

I don't expect self driving cars to be perfect and handle everything, but I expect the manufacturers to be transparent about their abilities and they aren't. Furthermore I expect the self driving system to have a way to react to unforeseen situations, crashing in fog is not acceptable when the fact that there was fog was plainly obvious.

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago)

And a thing blocking the road isn’t exactly unforeseen either.

Tesla's system intentionally assumes "a thing blocking the road" is a sensor error.

They have said if they don't do that, about every hour or so you'd drive past a building and it would slam on the brakes and stop in the middle of the road for no reason (and then, probably, a car would crash into you from behind).

The good sensors used by companies like Waymo don't have that problem. They are very accurate.