• jarfil@beehaw.org
      link
      fedilink
      arrow-up
      1
      ·
      7 months ago

      It’s a “push as much data as a baby gets to train its NN” step, which is several orders of magnitude more, and more focused, than any training dataset in existence right now.

      Even with diminishing returns, it’s bound to get better results.