• Kachajal@lemmy.ml
    link
    fedilink
    arrow-up
    9
    arrow-down
    4
    ·
    edit-2
    5 months ago

    In the same way that computers are basically advanced abaci.

    Don’t confuse a simplification made to demonstrate the basic functioning to a layman with how things actually work.

    LLM’s are neural networks, which are based on a model of brain function. There’s little reason to believe that we cannot eventually reach similar levels of effectiveness as human brains.

    Hell - reaching the levels of pigeon brains would already be absurdly useful.

    • pelotronOP
      link
      fedilink
      English
      arrow-up
      6
      ·
      5 months ago

      The problem is they’re already talking about needing trillions of dollars worth of hardware to make it happen. It’s absurd.

    • chayleaf@lemmy.ml
      link
      fedilink
      arrow-up
      6
      ·
      5 months ago

      While I agree that LLMs can achieve human-tier efficiency at most tasks eventually (some architectural changes will be necessary, but the core approach seems sound), it’s wrong to say it’s modeled after the human brain. We have no idea how brains work as they’re super complex, we’re building artificial neural networks from the ground up. AI uses centuries’ worth of math, but with our current maths knowledge the code isn’t too complicated. Human brains aren’t like that, they can’t be summed up in a few lines of code because DNA is a huge mess that contains so much more than just “learning”, so many inactive or redundant bits and pieces. We’re building LLMs with knowledge of how languages work, not how brains work.

      • jsomae@lemmy.ml
        link
        fedilink
        arrow-up
        1
        ·
        5 months ago

        Transformers are not built with our knowledge of language. That’s a gross approximation – it would honestly be more accurate to say they’re modelled after the human brain than that they’re built with our understanding of language. A big problem is that the connection between AI and language is poorly understood – we can’t even understand what the word2vec axes are.

        • chayleaf@lemmy.ml
          link
          fedilink
          arrow-up
          1
          ·
          5 months ago

          i’m not talking about knowing about how humans perceive/learn languages, i’m talking about language structure. Perhaps it’s wrong to call it “how languages work”

          • jsomae@lemmy.ml
            link
            fedilink
            arrow-up
            1
            ·
            5 months ago

            That’s what I meant, yes. They’re not built based on any linguistic field

            • chayleaf@lemmy.ml
              link
              fedilink
              arrow-up
              2
              ·
              edit-2
              5 months ago

              different neural network types excel at different tasks - image recognition was invented way before LLMs, not only for lack of processing power, but also because the previous architectures didn’t work with languages. New architectures don’t appear out of thin air, they are created with a rough idea of what we could need to make the network do a certain task (e.g. NLP) better. Even tokenization isn’t blind codepoint separation but is based on an analysis of languages. But yes, natural languages aren’t “parsed” for neural networks, they don’t even have a formal grammar.