• SuperSleuth@lemm.ee
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    1
    ·
    1 year ago

    Should a self-driving car face more rigorous tests than actual human drivers? Honest question.

      • stopthatgirl7@kbin.social
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        1 year ago

        Yes, because when there’s an accident with a person driving, you usually know exactly who is legally to blame in an accident. With self-driving, if the car accidentally hits and kills someone, who do you charge for it? There’s no one person you can point to for responsibility for if something goes wrong, like you can for a person responsible for an accident.

    • FoxBJK@midwest.social
      link
      fedilink
      English
      arrow-up
      10
      ·
      1 year ago

      Human drivers should be facing more rigorous testing regardless. It’s horrifically easy to get a license… and then they never test you again for the rest of your life. That’s just insane when you think about it. My test was in 2002. Feels like I should have to retake it at some point.

    • IphtashuFitz@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      1 year ago

      Yes. A human brain can handle edge cases it’s never encountered before. Can a self driving car?

      • Ever stop at a red light only to have a police officer wave you through?

      • Ever encounter a car driving the wrong way down a one way street?

      • Ever come across a flooded out stretch of road? (if the road has no lines and the water is still it can be very deceptive looking)

      These are a tiny number of things I’ve encountered over the past few years. I’m sure plenty of other drivers can provide other good examples. I’d want to know how a self driving car would handle itself in situations like these.

        • NeoNachtwaechter@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 year ago

          How will the bot car handle itself out in the country? Dirt roads?

          They don’t go there. They have their limits. Simple as that.

          But when the police has ordered them there (for example, the good road must be emptied because of an emergency) then the trouble starts… now imagine not just one or two, but hundreds of them.

    • snooggums@kbin.social
      link
      fedilink
      arrow-up
      5
      ·
      1 year ago

      Yes because each person must learn on their own and have limited experience relative to the general public as a whole.

      Self driving cars can ‘learn’ from all self driving cars and don’t get tired, forget, or anything like that. While they shouldn’t be held to perfection, they should absolutely be held to a higher standard than a human.

    • nxfsi@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      1 year ago

      Only Tesla self driving cars need to have more rigorous tests. Other brands are fine as it is because they have lidar.

      • skymtf@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        I feel like all them do, have you seen wayze nearly getting black people killed cause it didn’t stop for s cop. And it can’t recognize construction zones.

      • sky@codesink.io
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        Five LiDAR sensors hasn’t stopped Cruise from running into a bus, multiple cars, and a fire truck. Maybe self-driving is a myth?

        Maybe we should just build buses and trains and pay people good salaries to operate them??