Related:

This is in a PR where Shougo, another long-time contributor, communicates entirely in walls of unparseable AI slop text: https://github.com/vim/vim/pull/19413

Thank you for the detailed feedback! I’ve addressed all the issues:

Thank you for the feedback! I agree that following the Vim 8+ naming convention makes sense.

Thank you for the feedback on naming!

Thanks for the suggestion! After thinking about this more, I believe repeat_set() / repeat_get() is the right choice:

Thank you for the feedback. A brief clarification.

https://hachyderm.io/@AndrewRadev/116176001750596207

@AndrewRadev@hachyderm.io

  • hperrin
    link
    fedilink
    English
    arrow-up
    160
    arrow-down
    1
    ·
    1 day ago

    I spent literally all day yesterday working on this:

    https://sciactive.com/human-contribution-policy/

    I’ve started to add it to my projects. Eventually, it will be on all of my projects. I made it so that any project could adopt it, or modify it to their needs. It’s got a thorough and clear definition of what is banned, too, so it should help any argument over pull requests.

    Hopefully more projects will outright ban AI generated code (and other AI generated material).

      • hperrin
        link
        fedilink
        English
        arrow-up
        8
        ·
        17 hours ago

        Ok, yeah, I’ll make a post for it.

        Feel free to share it anywhere. :)

      • hperrin
        link
        fedilink
        English
        arrow-up
        74
        ·
        1 day ago

        Basically the best you can do is continue as normal, and if someone submits something that says it is or obviously is AI, point to this policy and reject it. Just having the policy should be a decent deterrent.

      • Jankatarch@lemmy.world
        link
        fedilink
        arrow-up
        19
        ·
        23 hours ago

        Same mindset as “You don’t need a perfect lock to protect your house from thieves, you just need one better than what your neighbors have.”

        If a vibecoder sees this they will not bother with obfuscation and simply move onto the next project.

      • Retail4068@lemmy.world
        link
        fedilink
        arrow-up
        3
        arrow-down
        67
        ·
        1 day ago

        No, it’s a prejudiced hot take that’s completely and utterly unenforceable which will be seen as some Luddite behavior in 10 years when everyone is using the tooling.

          • Retail4068@lemmy.world
            link
            fedilink
            arrow-up
            2
            arrow-down
            46
            ·
            1 day ago

            I did. And you’re worried about clankers being able to comprehend as well as a human 🤣, good Lord the bar is low.

            • Scubus@sh.itjust.works
              link
              fedilink
              arrow-up
              1
              ·
              11 hours ago

              Ok that’s really funny and I do agree with you, but I think you might be coming at this a little… unhinged. The issue with this is that it is unenforceable and honestly somewhat pointless. If AI tools are not up to scratch, then that will always be reflected in the quality of the code. Bad code is bad code, it doesn’t matter what made it. A lot of people seem to think AI is synonomous with bad code, and if that is the case, simply ban bad code.

              The issue they are going to run into is twofold:

              Firstly, what qualifies as “using AI”? Admittedly I haven’t actually read their licensing, but I’m just going to take a guess and say that it bans all forms of AI used anywhere in production. Almost every compiler I use these days has auto predict. It’s rarely useful, but if it does happen to guess the rest of the code I was already going to type, and I accept that, did I use AI to assist my coding? Back in the day before it was an llm the auto predict was actually decent, so not all of them use AI. How would you even know whether your is AI or not?

              The second issue is an issue of foresight. When the AI tools do become up to scratch, that will be reflected in the quality of their code. Suddenly AI generated code is faster, more efficient, and easier to understand all simultaneously. Anyone using this license is effectively admitting that theirs is the inferior option.

              It’s always hilarious to me when people ask whether something is AI slop. I dunno man, has your ability to detect whether something is good been reduced to AI slop? If it’s good, it’s good. If it’s not, it’s not. Either you like it or you don’t. Feels very similar to transphobes saying they can always tell. If that’s true, and AI really is always going to worse, you should never have to ask whether something is AI slop, you should just be able to tell. Otherwise it’s just slop, no ai necessary.

          • Retail4068@lemmy.world
            link
            fedilink
            arrow-up
            1
            arrow-down
            2
            ·
            30 minutes ago

            Yes it does. Folks who just want to screech went crazy. Like, two of you actually engaged and brought valid concerns. Y’all are a CRAZY prejudiced bunch and hate being called out just as much as the next shit flinging monkey tribe.

            You actually think Lemmy is better behaved 🤣🤣🤣🤣

    • thethunderwolf@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      2
      ·
      17 hours ago

      “AI generated” means that the subject material is in whole, or in meaningful part, the output of a generative AI model or models, such as a Large Language Model. This does not include code that is the result of non-generative tools, such as standard compilers, linters, or basic IDE auto-completions. This does, however, include code that is the result of code block generators and automatic refactoring tools that make use of generative AI models.

      As “artificial intelligence” is not that well defined, you could clarify what the policy defines “AI” as by specifying that “AI” involves machine learning.

      • hperrin
        link
        fedilink
        English
        arrow-up
        9
        ·
        17 hours ago

        “Generative AI model” is a pretty well defined term, so this prohibits all of those things like ChatGPT, Gemini, Claude Code, Stable Diffusion, Midjourney, etc.

        Machine learning is a much more broad category, so banning all outputs of machine learning may have unintended consequences.