• CileTheSane
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    2
    ·
    6 hours ago

    Because it’s not AI, it’s LLMs, and all LLMs do is guess what word most likely comes next in a sentence. That’s why they are terrible at answering questions and do things like suggest adding glue to the cheese on your pizza because somewhere in the training data some idiot said that.

    The training data for LLMs come from the internet, and the internet is full of idiots.