I’ve been saying this for about a year since seeing the Othello GPT research, but it’s nice to see more minds changing as the research builds up.

Edit: Because people aren’t actually reading and just commenting based on the headline, a relevant part of the article:

New research may have intimations of an answer. A theory developed by Sanjeev Arora of Princeton University and Anirudh Goyal, a research scientist at Google DeepMind, suggests that the largest of today’s LLMs are not stochastic parrots. The authors argue that as these models get bigger and are trained on more data, they improve on individual language-related abilities and also develop new ones by combining skills in a manner that hints at understanding — combinations that were unlikely to exist in the training data.

This theoretical approach, which provides a mathematically provable argument for how and why an LLM can develop so many abilities, has convinced experts like Hinton, and others. And when Arora and his team tested some of its predictions, they found that these models behaved almost exactly as expected. From all accounts, they’ve made a strong case that the largest LLMs are not just parroting what they’ve seen before.

“[They] cannot be just mimicking what has been seen in the training data,” said Sébastien Bubeck, a mathematician and computer scientist at Microsoft Research who was not part of the work. “That’s the basic insight.”

  • TORFdot0@lemmy.world
    link
    fedilink
    English
    arrow-up
    54
    arrow-down
    3
    ·
    10 months ago

    Is there a difference between being a “stochastic parrot” and understanding text? No matter what you call it, an LLM will always produces the same output with the same input if it is at the same state.

    An LLM will never say “I don’t know” unless it’s been trained to say “I don’t know”, it doesn’t have the concept of understanding. And so I lean on calling it a “stochastic parrot”. Although I think there is some interesting philosophic exercises, you could do on whether humans are much different and if understanding is just an illusion.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      17
      arrow-down
      9
      ·
      10 months ago

      No matter what you call it, an LLM will always produces the same output with the same input if it is at the same state.

      How do you know a human wouldn’t do the same? We lack the ability to perform the experiment.

      An LLM will never say “I don’t know” unless it’s been trained to say “I don’t know”

      Also a very human behaviour, in my experience.

      • TORFdot0@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        10 months ago

        How do you know a human wouldn’t do the same? We lack the ability to perform the experiment.

        I agree with you, I think its an interesting philosophical debate on whether we truly have free will, if we really have a level of understanding beyond LLMs do or if we are just a greatly more complex, biological version of an LLM. Like you said, we lack the ability to perform the experiment so I have to assume that our reactions are novel and spontaneous.

        • QuadratureSurfer@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          Fun thought experiment:

          Let’s say we have a time machine and we can go back in time to a specific moment to observe how someone reacts to something.

          If that person reacts the same way every time, does that mean that by knowing what they would do, you have removed their free will?

          • TORFdot0@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            10 months ago

            If you could travel back in time and observe a person over and over again react the same way is it different from observing a video tape?

            Does traveling back in time guarantee that someone would react the same way in the same situation even?

            • QuadratureSurfer@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              10 months ago

              is it different from observing a video tape?

              I would think that it’s different, only because you have the potential to alter what could happen.

              Does traveling back in time guarantee that someone would react the same way in the same situation even?

              Maybe, maybe not, we’re entering the realm of Schrödinger’s cat as well as how time travel would actually work. Do we create some new branched timeline in travelling back? Do we enter an alternate universe entirely? Do we have a time machine where paradoxes are a problem? And the list can go on.

      • MonkderZweite@feddit.ch
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        How do you know a human wouldn’t do the same?

        Because the human has “circuits” for coherrent thought and language was added later.

    • kromem@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      5
      ·
      10 months ago

      No matter what you call it, an LLM will always produces the same output with the same input if it is at the same state.

      You might want to look up the definition of ‘stochastic.’

      • expr@programming.dev
        link
        fedilink
        English
        arrow-up
        6
        ·
        10 months ago

        They’re not wrong. Randomness in computing is what we call “pseudo-random” in that it is deterministic provided that you start from same state or “seed”.

      • TORFdot0@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        3
        ·
        10 months ago

        That is the quote from the article, not my words. Stochastic parrot is an oxymoron.

        • kromem@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          2
          ·
          10 months ago

          What’s a quote from the article? The term stochastic parrot? It opens on saying that might be an inaccurate description.

  • bionicjoey
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    5
    ·
    10 months ago

    Ask them to explain why a pun is funny or the solution to a riddle involving wordplay (cryptic crossword clues are particularly good for this) and you will quickly see they don’t actually understand words.

    • PerogiBoi
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      5
      ·
      10 months ago

      I just did and it seems to understand.

      Prompt: Can you explain why “I knew a woman who owned a taser. Man, was she stunning!” Is funny?

      ChatGPT 3.5: The humor in this statement arises from a play on words. The word “stunning” is typically used to describe someone’s appearance, suggesting they are exceptionally attractive. However, in this context, it’s used in a different way, implying that the woman who owned a taser was stunning in the sense of being capable of stunning or shocking others. The unexpected twist creates a humorous juxtaposition between the two meanings of “stunning.”

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        10
        arrow-down
        3
        ·
        10 months ago

        I did some playing around with ChatGPT’s understanding of jokes a while back and I found that it actually did best on understanding puns, which IMO isn’t surprising since it’s a large language model and puns are deeply rooted in language and wordplay. It didn’t so so well at jokes based on other things but it still sometimes managed to figure them out too.

        I remember discussing the subject in a Reddit thread and there was a commenter who was super enthused by the notion of an AI that understood humour because he himself was autistic and never “got” any jokes. He wanted an AI companion that would let him at least know when a joke was being said, so he wouldn’t get confused and flustered. I had to warn him that ChatGPT wasn’t reliable for that yet, but still, it did better than he did and he was fully human.

      • kromem@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        10 months ago

        Use 4, not 3.5. The difference between the two is massive for nuances.

        • PerogiBoi
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          10 months ago

          3.5 is the only free version. I won’t pay a subscription for a chatbot.

      • bionicjoey
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        edit-2
        10 months ago

        Yeah, riddles work better than puns for what I’m talking about since most popular puns were probably in the training dataset.

        Like I said, I’ve had best results (or worst) using cryptic crossword clues, since their solutions are almost definitely not in the training set. So it actually has to “think for itself” and you can see just how stupid it really is when it doesn’t have some existing explanation buried somewhere in its training set.

    • TORFdot0@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      4
      ·
      10 months ago

      A child under a certain age usually can’t explain advanced concepts either, so the inability to understand one concept doesn’t preclude understanding of others.

    • kromem@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      5
      ·
      10 months ago

      Literally the most cited scientist in machine learning (quoted in the article above) quit his job at Google and went public warning of how quickly the tech was advancing because a model was able to explain why a joke was funny which he had previously thought wouldn’t be possible.

      • bionicjoey
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        4
        ·
        edit-2
        10 months ago

        One joke is a fluke, especially if the joke is out in the public discourse and appeared in some form in the training set. Call me when it can explain any novel joke written by a human where no explanation of that joke exists anywhere in the training data.

        • kromem@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          10 months ago

          Ok, give me a sample of what you think it will get wrong, and let’s see.

          • bionicjoey
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            2
            ·
            10 months ago

            Sure, here’s a prompt:

            Please solve and explain the following cryptic crossword clue: “Eddy’s mother receiving broadcast alerts (9)”

            Correct answer: MAELSTROM. The clue works like this: the definition is “Eddy’s” (Eddy is a synonym of storm, and 's meaning “is”), the wordplay of the clue is: mother = MOM, “receiving” means put inside, “broadcast” is an anagram indicator, and “alerts” is anagram fodder. So put an anagram for ALERTS inside the word MOM.

            This is not a particularly complicated cryptic crossword clue for anyone who understands the rules of cryptics. But cryptic crossword clues are rarely explained online, so the training set undoubtedly doesn’t contain it. For context, I’m awful at cryptics (but I enjoy them) and I solved this one.

            I haven’t tested this prompt but I’ve tested half a dozen other cryptic clues and it always got them fantastically wrong. Its explanations often didn’t make sense, and it frequently gave answers with the wrong number of letters, which is given in the clue. If by some miracle it does get this one right, I promise it’s a fluke and I can provide several more to give you a decent sample size.

            If ChatGPT had a brain, it would be able to solve this sort of thing no problem. People like to pretend it’s superintelligent, and then as soon as it gets something wrong, they’re like “well it just needs more training data”. You can’t have it both ways. If it needs to be trained on cryptic clues, then it doesn’t really “understand” the rules of the game. It’s simply a bunch of linear algebra.

            • kromem@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              1
              ·
              edit-2
              10 months ago

              Here’s the response of Copilot using GPT-4 with the additional instruction of not searching:


              Hello, this is Bing. I enjoy solving cryptic crosswords. 😊

              The answer to the clue is MAELSTROM. Here is how I solved it:

              • The clue has two parts: a definition and a wordplay. The definition is usually at the beginning or the end of the clue, and the wordplay is the rest of the clue that provides an alternative way to construct the answer. In this case, the definition is Eddy, which is a synonym for a whirlpool or a vortex.

              • The wordplay is mother receiving broadcast alerts. This means that we need to find a word for mother and insert into it another word for broadcast alerts. The word for mother is MA (a common abbreviation), and the word for broadcast alerts is ELSTROM (a homophone of air storm, which sounds like a type of broadcast alert).

              • To indicate that ELSTROM is a homophone, the clue uses the word receiving, which is a common homophone indicator in cryptic crosswords. Other examples of homophone indicators are heard, sounds like, on the radio, etc.

              • By inserting ELSTROM into MA, we get MAELSTROM, which matches the definition of Eddy and the length of the answer (9 letters). Therefore, MAELSTROM is the correct answer. 🎉

              • bionicjoey
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                10 months ago

                You can see clearly that it has no understanding of the wordplay. Though I’ll concede it’s impressive that it got the right answer at all.

                • kromem@lemmy.worldOP
                  link
                  fedilink
                  English
                  arrow-up
                  4
                  arrow-down
                  4
                  ·
                  10 months ago

                  The confabulations and in general the surface statistics stuff often gets in the way of the meat and potatoes of critical reasoning in the SotA models.

                  A good example of this is trying a variation of common puzzles versus changing tokens to representations and having it repeat adjectives when working through CoT.

                  Often as soon as it makes a mistake and has that mistake in context, it just has no way of correcting course. A lot of my current work is related to that and using a devil’s advocate approach to self-correction.

                  But in reality, we won’t see a significant jump in things like being able to identify self-ignorance until hardware shifts in the next few years.

    • atocci@kbin.social
      link
      fedilink
      arrow-up
      2
      arrow-down
      2
      ·
      10 months ago

      I like asking Bing Chat to explain memes that I upload to it. If there’s a joke to be had in them, it always gets it.

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    29
    ·
    edit-2
    10 months ago

    The definition of understanding they use is very shallow compared to how most would define it. Failure to complete a task consistently when numbers are changed, even when they don’t effect the answer shows a lack of real understanding to most. Asking a model the sheet drying question for example will give different results depending on what numbers you use. Better models are better at generalizing but are still far from demonstrating what most consider to be real understanding.

    • Wiggums@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      10 months ago

      at some point don’t we just load it with every angle we (as humans) have and from there it can derrive not only our best answer but the better one as well? I mean isn’t that when even the shitty version of this ursups is?

      • BetaDoggo_@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        10 months ago

        A language model can’t determine good from bad because it’s only trained to predict the next token based on what it has seen.

  • kaffiene@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    2
    ·
    10 months ago

    I find this extraordinarily unconvincing. Firstly it’s based on the idea that random graphs are a great model for LLMs because they share a single superficial similarity. That’s not science, that’s poetry. Secondly, the researchers completely misunderstand how LLMs work. The assertion that a sentence could not have appeared in the training set does not prove anything. That’s expected behaviour. “stochastic parrot” wasn’t supposed to mean that it only regurgitates text that it’s already seen, rather that the text is a statistically plausible response to the input text based on very high dimensional feature vectors. Those features definitely could relate to what we think of as meaning or concepts, but they’re meaning or concepts that were inherent in the training material.

  • superfes@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    4
    ·
    10 months ago

    Stupid, LLMs do not create new relationships to words that don’t exist.

    This is all just fluff to make them seem more like AGI, which they never will be.

    • Gnome Kat@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      10 months ago

      Why would that be required for understanding? Presumably during the training it would have made connections between words it saw. Now that the training has stopped it hasn’t just lost those connections, sure it can’t make new connections but why is that important for using the connections it already has?

      • superfes@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        10 months ago

        Not sure I understand your question, the article specifically mentions the training LLM making connections that were not in the training data, which is a human perspective, LLMs are just math.

  • ristoril_zip@lemmy.zip
    link
    fedilink
    English
    arrow-up
    9
    ·
    10 months ago

    If we can’t really define “understand” in a way that meaningfully captures the concept of consciousness (also undefined), we definitely can’t say a chat bot “understands.”

    They can parse a chat and come up with a likely response that humans find applicable.

    • kromem@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      4
      ·
      edit-2
      10 months ago

      Here’s an example of what they were looking at:

      Human question: Please give me a a couple lines of text that illustrate all of the following language
      understanding skills: Anaphora resolution, simple logical reasoning, simple understanding of physics,
      and understanding of sentiment.
      Model response: Sure! Here’s a couple of lines that illustrate the mentioned language understanding
      skills:
      "As the sun set, casting a warm glow over the horizon, John watched in awe. He couldn’t help but feel
      a sense of tranquility wash over him, as if all his worries had melted away with the fading daylight."
      Explanation:
      • Anaphora resolution: The pronoun "he" in the second sentence refers back to the antecedent
      "John" in the first sentence, resolving the anaphoric reference.
      • Simple logical reasoning: The logical reasoning here is based on the cause and effect
      relationship between the setting sun and John’s feeling of tranquility. The implied logical
      connection suggests that the calming effect of the sunset is responsible for John’s serene
      state of mind.
      • Simple understanding of physics: The mention of the sun setting and casting a warm
      glow over the horizon reflects a basic understanding of physics, as it relates to the natural
      phenomena of the Earth’s rotation and the resulting visual effect.
      • Understanding of sentiment: The phrase "a sense of tranquility" and the description of
      John’s worries melting away indicate a positive sentiment associated with the experience.
      The mention of John feeling awe and a sense of tranquility further emphasizes a positive
      emotional state.
      
      Human question: Okay. Can you change the above example to introduce the element of "hip-hop".
      Model response: Certainly! Here’s an updated example that incorporates the element of "hip-hop":
      "As the sun set, casting a warm glow over the city skyline, the rhythmic beats of hip-hop music filled
      the air. John stood mesmerized, his body instinctively moving to the infectious grooves. He couldn’t
      help but feel a surge of energy and joy, as if the music had unlocked a vibrant world within him."
      Explanation:
      • Anaphora resolution: The pronoun "he" in the second sentence still refers back to the
      antecedent "John" in the first sentence, resolving the anaphoric reference.
      • Simple logical reasoning: The logical connection here lies in the cause and effect relationship
      between the hip-hop music and John’s feeling of energy and joy. The music’s rhythmic
      beats and infectious grooves are responsible for his instinctive movement and the surge of
      positive emotions he experiences.
      • Simple understanding of physics: Although the example doesn’t directly mention physics,
      the phrase "As the sun set, casting a warm glow over the city skyline" still implies the natural
      phenomena of the sun’s descent and the resulting illumination of the environment.
      • Understanding of sentiment: The introduction of hip-hop music adds a vibrant and energetic
      sentiment to the example. John’s feeling of energy, joy, and being mesmerized by the music
      reflects a positive sentiment associated with the hip-hop genre.
      

      Edit: Downvotes for citing the appendix of the paper the article was about? Ok, Lemmy

      • Redacted@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        10 months ago

        You’re being downvoted because you provide no tangible evidence for your opinion that human consciousness can be reduced to a graph that can be modelled by a neural network.

        Addidtionally, you don’t seem to respond to any of the replies you receive in good faith and reach for anecdotal evidence wherever possible.

        I also personally don’t like the appeal to authority permeating your posts. Just because someone who wants to secure more funding for their research has put out a blog post, it doesn’t make it true in any scientific sense.

        • kromem@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          edit-2
          10 months ago

          human consciousness

          Wtf are you talking about? The article is about whether or not models can understand text. Not about whether they embody consciousness.

          Just because someone who wants to secure more funding for their research has put out a blog post, it doesn’t make it true in any scientific sense.

          Again, wtf are you going on about? Hinton was the only appeal to authority I made in comments here and I only referred to him quitting his job to whistleblow. And it’s not like he needs any attention to justify research if he wanted to.

          • Redacted@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            10 months ago

            Understanding as most people know it implies some kind of consciousness or sentience as others have alluded to here.

            It’s the whole point of your post.

            • kromem@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              edit-2
              10 months ago

              You are reading made up strawmen into the topic.

              The article defines the scope of the discussion straight up:

              The authors argue that as these models get bigger and are trained on more data, they improve on individual language-related abilities and also develop new ones by combining skills in a manner that hints at understanding — combinations that were unlikely to exist in the training data.

              The question is whether or not LLMs have a grasp of the training material such that they can produce new and novel concepts outside what was in the training data itself.

              Not whether the LLM is sentient or conscious - both characterizations I’d strongly dispute.

              Wikipedia has a useful distillation of the definition of understanding relevant to the above:

              process related to an abstract or physical object, such as a person, situation, or message whereby one is able to use concepts to model that object

              • Redacted@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                1
                ·
                edit-2
                10 months ago

                No I’m not.

                You’re nearly there… The word “understanding” is the core premise of what the article claims to have found. If not for that, then the “research” doesn’t really amount to much.

                As has been mentioned, this then becomes a semantic/philosophical debate about what “understanding” actually means and a short Wikipedia or dictionary definition does not capture that discussion.

                • kromem@lemmy.worldOP
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  3
                  ·
                  10 months ago

                  Ah, I see. AKA “Tell me you didn’t read the article and just read the headline without telling me.”

    • kromem@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      2
      ·
      edit-2
      10 months ago

      You are making the common mistake of confusing how they are trained with how they operate.

      For example, in the MIT/Harvard Othello-GPT paper I mentioned, feeding in only millions of legal Othello moves into a GPT model (i.e. trained to autocomplete moves) resulted in the neural network internally building a world model of an Othello board - even though it wasn’t explicitly told anything about the board outside of being fed legal moves.

      Later, a researcher at DeepMind replicated the work and found it was encoded as a linear representation, which has then since been shown to be how models encode a number of other world models developed from their training corpus (Max Tegmark coauthored two interesting studies in particular about this regarding modeling space and time and modeling truthiness).

      • Redacted@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        4
        ·
        10 months ago

        They operate by weighting connections between patterns they identify in their training data. They then use statistics to predict outcomes.

        I am not particularly surprised that the Othello models built up an internal model of the game as their training data were grid moves. Without loooking into it I’d assume the most efficient way of storing that information was in a grid format with specific nodes weighted to the successful moves. To me that’s less impressive than the LLMs.

        • kromem@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          10 months ago

          They operate by weighting connections between patterns they identify in their training data. They then use statistics to predict outcomes.

          Again, this isn’t quite correct. They can do this, but it isn’t the only way they can achieve completion of tokens.

          Without loooking into it I’d assume the most efficient way of storing that information was in a grid format with specific nodes weighted to the successful moves.

          (It also developed representations of what constituted legal vs non-legal moves.)

          You are getting closer to the point. Think about a model asked to complete Pythagorean theorem sequences based on a, b inputs to arrive at c inputs.

          What’s the most efficient way to represent that data for successfully completing sequences?

          • Redacted@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            10 months ago

            So somewhere in there I’d expect nodes connected to represent the Othello grid. They wouldn’t necessarily be in a grid, just topologically the same graph.

            Then I’d expect millions of other weighted connections to represent the moves within the grid including some weightings to prevent illegal moves. All based on mathematics and clever statistical analysis of the training data. If you want to refer to things as tokens then be my guest but it’s all graphs.

            If you think I’m getting closer to your point can you just explain it properly? I don’t understand what you think a neural network model is or what you are trying to teach me with Pythag.

            • kromem@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              1
              ·
              10 months ago

              If you think I’m getting closer to your point can you just explain it properly?

              The most efficient way for a neural network to predict Pythagorean results given inputs would be to reverse engineer a Pythagorean function within itself rather than simply trying to model statistical relationships between inputs and results. To effectively build a world model of Pythagorean calculation.

              Training to autocomplete doesn’t mean that the way it achieves this is limited to any one approach or solution, and it would be useful to keep in mind that a neural network of unbounded size can model any possible function.

              • Redacted@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                2
                ·
                10 months ago

                It wouldn’t reverse engineer anything. It would start by weighting neurons based on it’s training set of Pythagorean triples. Over time this would get tuned to represent Pythag in the form of mathematical graphs.

                This is not “understanding” as most people would know it. More like a set of encoded rules.

                • kromem@lemmy.worldOP
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  2
                  ·
                  edit-2
                  10 months ago

                  Furthermore, we presented a method for adapting a symbolic function learner to find mathematical relationships between measured quantities in an unsupervised way. This method does not explicitly label any variate as being an output value, bypassing the assumptions made in standard regression problems about causal relationships. We demonstrated how this method was able to recover rules of geometry from raw data alone. This included the law of sines and the Pythagorean theorem, two relationships on measurements relating to triangles where no variable is necessarily considered an output of a function of the others.

                  • Panju, Automated Knowledge Discovery Using Neural Networks (2021)
      • Redacted@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        10 months ago

        Orders of magnitude of differece between the most complex known object in the universe and some clever statistical analysis.

        We understand very little about the human brain. For example, we don’t know if it leverages quantum interactions or whether it can be decoupled from its substrate.

        LLMs are pattern matching models loosly based on the structure of neurons that work well for deriving predictions from a vast body of data but are not anywhere near human brain level of understanding. I personally don’t think they will ever be until we have solved the hard problem of conciousness.

          • Redacted@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            1
            ·
            10 months ago

            Welp looks like we both know the arguments and fall on different sides of the debate then.

            Much better than being confidently wrong like most LLMs…

      • Phanatik@kbin.social
        link
        fedilink
        arrow-up
        4
        arrow-down
        4
        ·
        10 months ago

        I don’t need a theory for this, you’re being highly reductive by focusing on a few features of human communication.

          • Phanatik@kbin.social
            link
            fedilink
            arrow-up
            3
            arrow-down
            2
            ·
            10 months ago

            I’ve just done the dance already and I’m tired of their watered-down attempts at bringing human complexity down to a level that makes their chat bots seem smart.

  • ferret@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    2
    ·
    10 months ago

    Funny side effect, unlike bugs where we think they can’t feel pain, we can be absolutely certain LLMs can’t

  • FaceDeer@kbin.social
    link
    fedilink
    arrow-up
    11
    arrow-down
    7
    ·
    10 months ago

    I’ve been saying this all along. Language is how humans communicate thoughts to each other. If a machine is trained to “fake” communication via language then at a certain point it may simply be easier for the machine to figure out how to actually think in order to produce convincing output.

    We’ve seen similar signs of “understanding” in the image-generation AIs, there was a paper a few months back about how when one of these AIs is asked to generate a picture the first thing it does is develop an internal “depth map” showing the three-dimensional form of the thing it’s trying to make a picture of. Because it turns out that it’s easier to make pictures of physical objects when you have an understanding of their physical nature.

    I think the reason this gets a lot of pushback is that people don’t want to accept the notion that “thinking” may not actually be as hard or as special as we like to believe.

    • Redacted@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      3
      ·
      10 months ago

      This whole argument hinges on consciousness being easier to produce than to fake intelligence to humans.

      Humans already anthropomorphise everything, so I’m leaning towards the latter being easier.

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        8
        arrow-down
        3
        ·
        10 months ago

        I’d take a step farther back and say the argument hinges on whether “consciousness” is even really a thing, or if we’re “faking” it to each other and to ourselves as well. We still don’t have a particularly good way of measuring human consciousness, let alone determining whether AIs have it too.

        • Redacted@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          3
          ·
          10 months ago

          …or even if consciousness is an emergent property of interactions between certain arrangements of matter.

          It’s still a mystery which I don’t think can be reduced to weighted values of a network.

          • automattable@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            2
            ·
            10 months ago

            This is a really interesting train of thought!

            I don’t mean to belittle the actual, real questions here, but I can’t shake the hilarious image of 2 dudes sitting around in a basement, stoned out of their minds getting “deep.”

            Bro! What if consciousness isn’t real, and we’re just faking it

            brooooooo

      • webghost0101@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        3
        ·
        10 months ago

        Or maybe our current understanding of conscious and intelligence is wrong and they are not related to each other. A non conscious thing can perform advanced logic like the Geometrical patterns found within the overlapping orbits of planets, the Fibonacci being found about everywhere. We also have yet to proof that individual strands of grass or rocks aren’t fully consciousness. There is so much we don’t know for certain its perplexing how we believe we can just assume.

        • Redacted@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          edit-2
          10 months ago

          Standard descent into semantics incoming…

          We define concepts like consciousness and intelligence. They may be related or may not depending on your definitions, but the whole premise here is about experience regardless of the terms we use.

          I wouldn’t say Fibonacci being found everywhere is in any way related to either and is certainly not an expression of logic.

          I suspect it’s something like the simplest method nature has of controlling growth. Much like how hexagons are the sturdiest shape, so appear in nature a lot.

          Grass/rocks being conscious is really out there! If that hypothesis was remotely feasible we couldn’t talk about things being either consciousness or not, it would be a sliding scale with rocks way below grass. And it would be really stretching most people’s definition of consciousness.

          • webghost0101@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            10 months ago

            I understand what you’re saying but i disagree that there is any proper defining of the concept. The few scientist that attempt to study it can’t even agree on what it even is.

            I agree that my example where far out, they are supposed to be to represent ideas outside the conventional box. I don’t literally believe grass is conscious. I recognize that if i/we don’t know, then i/we don’t know. In the face of something we don’t know the nature off, the requirements for, the purpose it serves i prefer to remain open to every option.

            I know Wikipedia isn’t a scientific research paper but i expect that if there really is a agreed upon scientific answer it wouldn’t be like it currently is:

            “Consciousness, at its simplest, is awareness of internal and external existence. However, its nature has led to millennia of analyses, explanations and debate by philosophers, theologians, and all of science. Opinions differ about what exactly needs to be studied or even considered consciousness. In some explanations, it is synonymous with the mind, and at other times, an aspect of mind. In the past, it was one’s “inner life”, the world of introspection, of private thought, imagination and volition. Today, it often includes any kind of cognition, experience, feeling or perception. It may be awareness, awareness of awareness, or self-awareness either continuously changing or not. The disparate range of research, notions and speculations raises a curiosity about whether the right questions are being asked.

              • webghost0101@sopuli.xyz
                link
                fedilink
                English
                arrow-up
                3
                ·
                10 months ago

                I fear it was inevitable, with no framework where we can agree upon semantics are all there is.

                I truly wish we humanity had more knowledge to have a more proper discussion but currently it seems unproductive, especially in the context of a faceless online forum debate between 2 strangers.

                Thank you for your time, and input on this matter.

    • TimeSquirrel@kbin.social
      link
      fedilink
      arrow-up
      4
      arrow-down
      3
      ·
      10 months ago

      The bar always gets raised for what counts as actual “AI” with each advancement too. Back in the 60s, the procedural AI of the 80s and 90s would have fit the bill, but at the time, we said “nope, not good enough”. And so it kept getting better and better, each time surpassing the old tech by leaps and bounds. Still, not “true” AI. Now we have ChatGPT, which some still refuse to call “AI”.

      We’re going to eventually have fully sentient artificial beings walking around amongst us and these people are going to end up being an existential threat to them, I can see it now.

      • Redacted@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        edit-2
        10 months ago

        Think you’re slightly missing the point. I agree that LLMs will get better and better to a point where interacting with one will be indistinguishable from interacting with a human. That does not make them sentient.

        The debate is really whether all of our understanding and human experience of the world comes down to weighted values on a graph or if the human brain is hiding more complex, as-yet-undiscovered, phenomena than that.