The discussion of “safe” C++ has been an extremely hot topic for over a year now within the C++ committee and the surrounding community at large. This was mostly brought about as a result of article, after article, after article coming out from various consumer advocacy groups, corporations, and governments showing time and again that C++ and its lack of memory safety is causing an absolute fuckload of problems for people.

And unfortunately, this means that WG21, the C++ committee, has to take action because people are demanding it. Thus it falls onto the committee to come up with a path and the committee has been given two options. Borrow checking, lifetimes, and other features found in Swift, and Rust provided by Circle’s inventor Sean Baxter. Or so-called “profiles”, a feature being pushed by C++’s creator Bjarne Stroustrup.

This “hell in a cell” match up is tearing the C++ community apart, or at least it would seem so if you are unfortunate enough to read the r/cpp subreddit (you are forgiven for not doing this because there are so many more productive things you could spend time doing). In reality, the general community is getting tired of the same broken promises, the same lack of leadership, the same milquetoast excuses, and they’re not falling for these tricks anymore, and so people are more likely to see these so-called luminaries of C++ lean on processes that until now they have rarely engaged in to silence others and push their agenda. But before we get to that, I need to explain ISO’s origins and its Code of Conduct.

  • FizzyOrange@programming.dev
    link
    fedilink
    arrow-up
    3
    ·
    1 month ago

    Not really. It will predict more vulgar output but that is fixed by fine tuning. It’s not going to “poison” it in any meaningful sense.

    • Boomkop3@reddthat.com
      link
      fedilink
      arrow-up
      3
      arrow-down
      1
      ·
      1 month ago

      No, it won’t malfunction. It’s just not very useful as training data without extra work

      • lad@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        I’m afraid, LLMs are gone a bit further from the state when such ‘poisoning’ made sense.

        I’m afraid that soon this may reach a point where it will be easier for LLM to make sense of the text, than for a human, if this idea gets further development.

        • Boomkop3@reddthat.com
          link
          fedilink
          arrow-up
          1
          ·
          1 month ago

          llm’s might be able to go trough more content. But they won’t develop any sense any time soon

          • lad@programming.dev
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            I meant ‘make sense’ to mean ‘could rewrite without garbage’. Maybe I was wrong, anyway

            • Boomkop3@reddthat.com
              link
              fedilink
              arrow-up
              1
              ·
              1 month ago

              Ah, I’m not so sure about that. You’d be feeding the model it’s own partial work. Which should work, but nowhere near what pure human data would’ve been.