• skymtf@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    65
    arrow-down
    4
    ·
    1 year ago

    I feel like the NTSB need to draft a min spec for self driving cars and a testing course that involves some of the worst circtimstances to get approved. I feel like all self driving cars should have to have lidar, and other sensors. Computer vision really isn’t working out.

    • echo64@lemmy.world
      link
      fedilink
      English
      arrow-up
      46
      arrow-down
      4
      ·
      1 year ago

      You build a benchmark and tesla will train on that benchmark, says nothing about real world use but gets them signed off.

      But yes western society is currently in a hellscape of refusing to do even basic regulation of any new technology so it’ll probably be a good 20 years of murder robots on the streets before anything gets written down.

      • FoxBJK@midwest.social
        link
        fedilink
        English
        arrow-up
        23
        ·
        1 year ago

        By “western society” do you mean the US? Because the EU doesn’t seem to have any qualms about regulating new technologies. That seems to be a uniquely American thing.

      • ayaya@lemdro.id
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        1
        ·
        edit-2
        1 year ago

        To be fair we already have giant metal murder boxes zooming around on the streets. If AI kills even a single person everyone flips out even though over 40,000 people die every year in the US from car accidents. And that is just the deaths, not including injuries. Yet I don’t really see anyone calling for more regulations on driving tests for humans.

        People want AI to somehow be perfect when in reality as long as AI is even 1% better than humans that’s saving over 400 lives per year. AI doesn’t get sleepy, distracted, drunk, etc. so it probably already is at least 1% better in most situations. Humans are horrible drivers.

      • originalucifer@moist.catsweat.com
        link
        fedilink
        arrow-up
        5
        arrow-down
        1
        ·
        1 year ago

        But yes western society is currently in a hellscape of refusing to do even basic regulation

        US regulations are only written in blood or money. the united states was built on the backs of slaves, and then wage-slaves. literal graveyards filled with workers.

        im not disagreeing with you, i just found this comically disparate to history… ie, its always been a regulation hellscape.

      • NeoNachtwaechter@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        But yes western society is currently in a hellscape of refusing to do even basic regulation

        Only the Usamerican country.

        We Europeans are scratching our heads already for very long: why are they letting these guys do just everything they want?

    • SuperSleuth@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      1 year ago

      Should a self-driving car face more rigorous tests than actual human drivers? Honest question.

        • stopthatgirl7@kbin.social
          link
          fedilink
          arrow-up
          8
          arrow-down
          1
          ·
          1 year ago

          Yes, because when there’s an accident with a person driving, you usually know exactly who is legally to blame in an accident. With self-driving, if the car accidentally hits and kills someone, who do you charge for it? There’s no one person you can point to for responsibility for if something goes wrong, like you can for a person responsible for an accident.

      • FoxBJK@midwest.social
        link
        fedilink
        English
        arrow-up
        10
        ·
        1 year ago

        Human drivers should be facing more rigorous testing regardless. It’s horrifically easy to get a license… and then they never test you again for the rest of your life. That’s just insane when you think about it. My test was in 2002. Feels like I should have to retake it at some point.

      • IphtashuFitz@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        ·
        1 year ago

        Yes. A human brain can handle edge cases it’s never encountered before. Can a self driving car?

        • Ever stop at a red light only to have a police officer wave you through?

        • Ever encounter a car driving the wrong way down a one way street?

        • Ever come across a flooded out stretch of road? (if the road has no lines and the water is still it can be very deceptive looking)

        These are a tiny number of things I’ve encountered over the past few years. I’m sure plenty of other drivers can provide other good examples. I’d want to know how a self driving car would handle itself in situations like these.

          • NeoNachtwaechter@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 year ago

            How will the bot car handle itself out in the country? Dirt roads?

            They don’t go there. They have their limits. Simple as that.

            But when the police has ordered them there (for example, the good road must be emptied because of an emergency) then the trouble starts… now imagine not just one or two, but hundreds of them.

      • snooggums@kbin.social
        link
        fedilink
        arrow-up
        5
        ·
        1 year ago

        Yes because each person must learn on their own and have limited experience relative to the general public as a whole.

        Self driving cars can ‘learn’ from all self driving cars and don’t get tired, forget, or anything like that. While they shouldn’t be held to perfection, they should absolutely be held to a higher standard than a human.

      • nxfsi@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        1 year ago

        Only Tesla self driving cars need to have more rigorous tests. Other brands are fine as it is because they have lidar.

        • skymtf@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          1
          ·
          1 year ago

          I feel like all them do, have you seen wayze nearly getting black people killed cause it didn’t stop for s cop. And it can’t recognize construction zones.

        • sky@codesink.io
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          1
          ·
          1 year ago

          Five LiDAR sensors hasn’t stopped Cruise from running into a bus, multiple cars, and a fire truck. Maybe self-driving is a myth?

          Maybe we should just build buses and trains and pay people good salaries to operate them??

    • markr@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Everyone would build to pass the test track. This does get at the problem though: the permutations of scenarios an L5 system has to correctly process is a huge number. Trying to build a system that can do that appears to be beyond anyone’s av system right now. This is why the most advanced deployments are all geofenced. That way at least the traffic signs and signals, lane markings, etc all understood and tested. Even then ‘shit happens’. Untested scenarios still occur. Also the maps are always out of date.

      The problem really requires AGI, and nobody has one of those, or if they do it’s a secret.

    • Cheers@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Throw I some pot holes and child pedestrian crossing the street, etc and they’d even come out with a powerful marketing ad.

    • soEZ@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      1 year ago

      It’s not really a sensor issue, as much as having software that can interpret the sensor data and act on it. Cameras and lidar effectively provide same thing, distance to objects in 2d/3d. But u need software to process that data and identify where the road is, where little jonny is, and what to do…arguably, the distance measuring problem has been solved for a while with lidar or with cameras, it’s object identification and reaction to that info that’s not solved. You can’t really solve it with traditional if/else programming, while AI gives you only a probability of what something is or what action to do…so the problem is hard.

      But ntsb/dmv whatever needs to come up with a way to test and classify autonomous driving software…probably doing real world test and identifying edge cases where it fails.

    • jonne@infosec.pub
      link
      fedilink
      English
      arrow-up
      21
      ·
      1 year ago

      Yeah, fun stuff happens when the AI tries to interpret vision of sunlight shining straight into the lens.

    • wagoner@infosec.pub
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      1
      ·
      1 year ago

      They did but then Musk had the genius idea to stop installing them. I still have it in my older model but they changed the software not to use them anymore. Like I said, genius…

    • TigrisMorte@kbin.social
      link
      fedilink
      arrow-up
      10
      arrow-down
      1
      ·
      1 year ago

      Don’t you know those things cost money!?

      Sadly cost cutting MBAs seldom concern themselves with silly things like function or necessity.

    • Thorny_Thicket@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      4
      ·
      1 year ago

      For an autonomous vehicle without radars or LiDAR they do still drive pretty darn well. AI DRIVR makes really good videos about FSD on YouTube and love it or hate it, it’s quite impressive how well it does despite the the lack of these sensors.

      • NeoNachtwaechter@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        1
        ·
        edit-2
        1 year ago

        For an autonomous vehicle without radars or LiDAR they do still drive pretty darn well.

        That’s a bit like saying “For an old hound dog with only two legs left he is running pretty darn fast”… :-)

  • Uriel238 [all pronouns]@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    2
    ·
    1 year ago

    When a service is willing to take responsibility for collisions and driving violations, then we know it works. If the guy asleep at the wheel (which he allegedly can do in an autonomous car) is still the one held responsible, then were not there yet.

    That said end-to-end AI totally sounds like equivocal marketing buzz.

    • danhab99@programming.dev
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      2
      ·
      1 year ago

      When a service is willing to take responsibility for collisions and driving violations

      Devil’s advocate: it’s kinda hard to pin the responsibility on Tesla when at the end of the day there was a person driving and the driver’s always responsible.

      I’m not disagreeing with you, I’m on team ban-human-drivers

      • Uriel238 [all pronouns]@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        Ideally, we’d get to the point where the driver merely directs the vehicle to where it wants to go, and then the computer system works out all the pathfinding and maneuvering, so that yes, any instance where a vehicle avoidably collides with another thing can be regarded as a malfunction.

    • Corkyskog@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      I wonder what happens when the car is on a collision course with a golden retriever and the only way not to hit it would be to damage the car. Or same scenario, but the only way not to hit it, is it to hit an 07 Carolla parked on the side of the road. Not saying humans have superior judgement… just wondering if it will be programmed by the theory of actuarial of philosophical science.

      • ramblinguy@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        That makes me think- will the AI see a kid that’s about to run out from behind a parked car? As a human, if I see a kid run from the house into a row of parked cars, I know he’s still there and will slow down before I get there. But would self driving make that same leap of logic? I’m not sure what the range and capabilities of self driving cars are right now in terms of scanning, but hopefully it would be smart enough to take preventative measures

        • Uriel238 [all pronouns]@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 year ago

          Right now, car AI has trouble both with kids and non-white persons. That said, when it comes to the things it is good at detecting, the cars respond much more quickly. This came up when an official asked about how it detects brake lights, and the project advisor (from Google, I think) explained that the car doesn’t worry about break lights but instantly detects when a car ahead of it rapidly decelerates, and responds immediately.

          I’m pretty sure we can get cars smart enough and sharp enough to drive better than humans. But the recent incident in San Francisco where Cruise driverless taxis blocked an ambulance with a patient in critical condition (resulting in their death), suggests to me we underestimated the layers of logistics necessary to make cars truly autonomous.

          Randal Munroe listed a few more incidents we can expect (Obligatory XKCD).

        • Corkyskog@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          According to other commentors, the need will never arise because the AI cars will be programmed so well it’s impossible to have accidents 🙄… now I see why FSD will never become a reality.

        • ramjambamalam@lemmy.ca
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 year ago

          Good question. Neural networks are modelled after how brains learn and process information, so it’s certainly theoretically possible for a neural network (or other machine learning algorithm) to make inferences like that, just like how you’ve learned them with years of experience.

          The biggest challenge in any machine learning is finding enough labelled training data. In fact, a friend of mine contributed to a paper in which (no joke) GTA V was used to generate labelled training data for an automous vehicle. Because it’s a game engine, every object in the game is already digitized, and the 3D modelling is accurate enough to be useful, at least. This vehicle used LIDAR so the actual shaders and such didn’t matter as much as the 3D point cloud.

      • Uriel238 [all pronouns]@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I doubt it. Germany has already implemented (considering implementing) regulations regarding the ethics of autonomous vehicles. As it is, cars are simply trying not to collide with anything and given their reflexes and perception are way faster and more accurate than human beings, they have a better chance of saving both the dog and the other car.

        That said, one of the problems we’re seeing with smart devices (that is devices that are software run rather than controlled by simple mechanics) is that companies are keen to abuse the power that gives them, hence the whole John Deere tractors debacle and the development of right-to-repair laws. Also, some BMWs require rental of some of their features (such as seat warmers) which seems to me as less than ethical.

        So I hope we’ll get to a point where not only is it anyone’s right to jailbreak their devices (including a self-driving car) but there will be several FOSS options we can choose from. And that means someone who programmed them may actually find a process-layer in which hazard prioritization or victim prioritization is considered.

        It is certainly an entertaining idea of speculative fiction that an aggressive driver package is developed, gets popular and then causes a rise in traffic accidents. More likely would be software packages that allow the vehicle to operate despite self-test failures, again leading to a higher traffic collision rate.

      • iminahurry@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        There are many such hypothetical scenarios based on the trolley problem, but the real answer is that a good self driving system will never end up in that situation in the first place.

        So as a dev, you just program to not let that situation arise, then you won’t need to program a solution for that.

  • arefx@lemmy.ml
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    3
    ·
    1 year ago

    Yeah I’m not getting in a car driven by AI. The tech ain’t where it needs to be for that.

    • DSX@lemm.ee
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      1
      ·
      1 year ago

      Especially Tesla. I am very into computer vision research but I would never trust a vehicle that relies on only that with 0 LIDAR or other sensing technologies in place.

      • Patches@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Great news. You don’t have to. The other people around you moving thousands of pounds of steel will be relying completely on that, and there is nothing you can do to stop it.

    • CoderKat@lemm.ee
      link
      fedilink
      English
      arrow-up
      16
      ·
      1 year ago

      It needs to be regulated to hold manufacturers responsible when their software isn’t good enough. My understanding is that there already probably is enough regulation and government agencies just need to hold Tesla accountable.

      Personally, I’m all for cars driven by AI iff it’s better and safer than a human driver. Human drivers make a lot of mistakes and driving is the most dangerous everyday activity many people do. But if the AI isn’t better than a human, that’s a problem. I don’t need AI drivers to be flawless, as that’s an unrealistic bar. I just need them to be undeniably better than humans. Everything I’m hearing about Tesla’s self driving is that they aren’t.

    • merc@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      3
      ·
      1 year ago

      I’m convinced that the self-driving AI from competent companies is better than human drivers already. The bar isn’t perfection, the bar is the average driver, and the average driver is bad.

      Having said that, I’d never get in a car with Tesla’s self-driving solution. Musk polluted the term “full self-driving” to cars that definitely weren’t.

      • batmaniam@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 year ago

        No, the bar is accountability. If the FSD screws up either:

        1. It’s claimed the driver was responsible, in which case it’s absolutely not FSD
        2. Someone other than the driver is investigated and cleared/convicted of wrong doing.

        It absolutely does not matter that it’s better than an “average” driver. If the best driver in the world screws up, they are still held accountable.

  • twhite@lemmy.ml
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    2
    ·
    edit-2
    1 year ago

    Not a fan of Tesla or Musk, but can we differentiate the broad public understanding of the term AI from machine learned control systems? People anthropomorphize the situation into thinking there is an I, Robot style driver enough as it is.

    Counterpoint, though, maybe doing so encourages skepticism of Tesla’s capabilities.

  • hark@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    3
    ·
    1 year ago

    That’s right, all the mistakes it makes are now powered purely by AI. Isn’t that fantastic?

  • BeautifulMind ♾️@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    1
    ·
    edit-2
    1 year ago

    As a cyclist I really do look forward to the day where good AI is consistently better than the average-to-worst drivers out there; the bar is depressingly low and the stakes are high.

    I write (and test) software for a living and my experience with Tesla as a consumer device is that it’s many generations away from being something I would trust.

    Also, I’ve seen what happens to product quality when management overrides its engineers in the way elon does- we get pre-alpha quality out there in the wild, being tested on a public that didn’t sign up for that shit