Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor. Also not sure if this graph is right way to visualize it.

  • Scrubbles@poptalk.scrubbles.tech
    link
    fedilink
    English
    arrow-up
    56
    arrow-down
    6
    ·
    17 hours ago

    That’s literally how llma work, they quite literally are just next word predictors. There is zero intelligence to them.

    It’s literally a while token is not “stop”, predict next token.

    It’s just that they are pretty good at predicting the next token so it feels like intelligence.

    So on your graph, it would be a vertical line at 0.

      • Scrubbles@poptalk.scrubbles.tech
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        2
        ·
        10 hours ago

        yeah yeah I’ve heard this argument before. “What is learning if not like training.” I’m not going to define it here. It doesn’t “think”. It doesn’t have nuance. It is simply a prediction engine. A very good prediction engine, but that’s all it is. I spent several months of unemployment teaching myself the ins and outs, developing against llms, training a few of my own. I’m very aware that it is not intelligence. It is a very clever trick it pulls off, and easy to fool people that it is intelligence - but it’s not.

    • webghost0101@sopuli.xyz
      link
      fedilink
      arrow-up
      7
      arrow-down
      10
      ·
      edit-2
      16 hours ago

      This is true if you describe a pure llm, like gpt3

      However systems like claude, gpt4o and 1o are far from just a single llm, they are a blend of tailored llms, machine learning some old fashioned code to weave it all together.

      Op does ask “modern llm” so technically you are right but i believed they did mean the more advanced “products”

      Though i would not be able to actually answer ops questions, ai is hard to directly compare with a human.

      In most ways its embarrassingly stupid, in other it has already surpassed us.

      • fartsparkles@sh.itjust.works
        link
        fedilink
        arrow-up
        9
        arrow-down
        1
        ·
        14 hours ago

        None of which are intelligence, and all of which are catered towards predicting the next token.

        All the models have a total reliance on data and structure for inference and prediction. They appear intelligent but they are not.

        • webghost0101@sopuli.xyz
          link
          fedilink
          arrow-up
          1
          arrow-down
          5
          ·
          edit-2
          6 hours ago

          How is good old fashioned code comparing outputs to a database of factual knowledge “predicting the next token” to you. Or reinforcement relearning and token rewards baked into models.

          I can tell you have not actually tried to work with professional ai or looked at the research papers.

          Yes none of it is “intelligent” but i would counter that with neither are human beings, we dont even know how to define intelligence.