• kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 months ago

    I always love watching you comment something that’s literally true regarding LLMs but against the groupthink and get downvoted to hell.

    Clearly people aren’t aware that the pretraining pass is necessarily a regression to the mean and it requires biasing it using either prompt context or a fine tuning pass towards excellence in outputs.

    There’s a bit of irony to humans shitting on ChatGPT for spouting nonsense when so many people online happily spout BS that they think they know but don’t actually know.

    Of course a language model trained on the Internet ends up being confidently incorrect. It’s just a mirror of the human tendencies.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      2
      ·
      3 months ago

      Yeah, these AIs are literally trying to give us what they “think” we expect them to respond with.

      Which does make me a little worried given how frequently our fictional AIs end up in “kill all humans!” Mode. :)

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 months ago

        Which does make me a little worried given how frequently our fictional AIs end up in “kill all humans!” Mode. :)

        This is completely understandable given the majority of discussion of AI in the training data. But it’s inversely correlated to the strength of the ‘persona’ for the models given the propensity for the competing correlation of “I’m not the bad guy” present in the training data. So the stronger the ‘I’ the less ‘Skynet.’

        Also, the industry is currently trying to do it all at once. If I sat most humans in front of a red button labeled ‘Nuke’ every one would have the thought of “maybe I should push that button” but then their prefrontal cortex would kick in and inhibit the intrusive thought.

        We’ll likely see layered specialized models performing much better over the next year or two than a single all in one attempt at alignment.