• vrighter@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    7
    ·
    edit-2
    23 hours ago

    well the answer is in the first sentence. They did not train a model. They fine tuned an already trained one. Why the hell is any of this surprising anyone? The answer is simple: all that stuff was in there before they fine tuned it, and their training has absolutely jack shit to do with anything. This is just someone looking to put their name on a paper

    • floofloofOP
      link
      fedilink
      English
      arrow-up
      13
      ·
      edit-2
      21 hours ago

      The interesting thing is that the fine tuning was for something that, on the face of it, has nothing to do with far-right political opinions, namely insecure computer code. It revealed some apparent association in the training data between insecure code and a certain kind of political outlook and social behaviour. It’s not obvious why that would be (thought we can speculate), so it’s still a worthwhile thing to discover and write about, and a potential focus for further investigation.

        • floofloofOP
          link
          fedilink
          English
          arrow-up
          6
          ·
          21 hours ago

          And it’s interesting to discover this. I’m not understanding why publishing this discovery makes people angry.

            • floofloofOP
              link
              fedilink
              English
              arrow-up
              5
              ·
              21 hours ago

              It’s research into the details of what X is. Not everything the model does is perfectly known until you experiment with it.

              • vrighter@discuss.tchncs.de
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                5
                ·
                20 hours ago

                we already knew what X was. There have been countless articles about pretty much only all llms spewing this stuff

    • OpenStars@piefed.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      22 hours ago

      Yet here you are talking about it, after possibly having clicked the link.

      So… it worked for the purpose that they hoped? Hence having received that positive feedback, they will now do it again.