How stupid do you have to be to believe that only 8% of companies have seen failed AI projects? We can’t manage this consistently with CRUD apps and people think that this number isn’t laughable? Some companies have seen benefits during the LLM craze, but not 92% of them. 34% of companies report that generative AI specifically has been assisting with strategic decision making? What the actual fuck are you talking about?

I don’t believe you. No one with a brain believes you, and if your board believes what you just wrote on the survey then they should fire you.

  • sugar_in_your_tea@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    I interviewed a candidate for a senior role, and they asked if they could use AI tools. I told them to use whatever they normally would, I only care that they get a working answer and that they can explain the code to me.

    The problem was fairly basic, something like randomly generate two points and find the distance between them, and we had given them the details (e.g. distance is a straight line). They used AI, which went well until it generated the Manhattan distance instead of the Pythagorean theorem. They didn’t correct it, so we pointed it out and gave them the equation (totally fine, most people forget it under pressure). Anyway, they refactored the code and used AI again to make the same mistake, didn’t catch it, and we ended up pointing it out again.

    Anyway, at the end of the challenge, we asked them how confident they felt about the code and what they’d need to do to feel more confident (nudge toward unit testing). They said their code was 100% correct and they’d be ready to ship it.

    They didn’t pass the interview.

    And that’s generally my opinion about AI in general, it’s probably making you stupider.

    • deweydecibel@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      6 months ago

      I’ve seen people defend using AI this way by comparing it to using a calculator in a math class, i.e. if the technology knows it, I don’t need to.

      And I feel like, for the kind of people whose grasp of technology, knowledge, and education are so juvenile that they would believe such a thing, AI isn’t making them dumber. They were already dumb. What the AI does is make code they don’t understand more accessible, which is to say, it’s just enabling dumb people to be more dangerous while instilling them with an unearned confidence that only compounds the danger.

    • Excrubulent@slrpnk.net
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      6 months ago

      Wait wait wait so… this person forgot the pythagorean theorem?

      Like that is the most basic task. It’s d = sqrt((x1 - x2)^2 + (y1 - y2)^2), right?

      That was off the top of my head, this person didn’t understand that? Do I get a job now?

      I have seen a lot of programmers talk about how much time it saves them. It’s entirely possible it makes them very fast at making garbage code. One thing I’ve known for a long time is that understanding code is much harder than writing it, and so asking an LLM to generate your code sounds like it’s just creating harder work for you, unless you don’t care about getting it right.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 months ago

        Yup, you’re hired as whatever position you want. :)

        Our instructions were basically:

        1. randomly place N coordinates on a 2D grid, and a random target point
        2. report the closest of those N coordinates to the target point

        It was technically different (we phrased it as a top-down game, but same gist). AI generated manhattan distance (abs(x2 - x1) + abs(x2 - x1)) probably due to other clues in the text, but the instructions were clear. The candidate didn’t notice what it was doing, we pointed it out, then they asked for the algorithm, which we provided.

        Our better candidates remember the equation like you did. But we don’t require it, since not all applicants finished college (this one did). We’re more concerned about code structure, asking proper questions, and software design process, but math knowledge is cool too (we do a bit of that).

        • frezik@midwest.social
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 months ago

          College? Pythagorean Theorem is mid-level high school math.

          I did once talk to a high school math teacher about a graphics program I was hacking away on at the time, and she was surprised that I actually use the stuff she teaches. Which is to say that I wouldn’t expect most programmers to know it exactly off the top of their head, but I would expect they’ve been exposed to it and can look it up if needed. I happen to have it pretty well ingrained in my brain.