Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor. Also not sure if this graph is right way to visualize it.

  • Max-P@lemmy.max-p.me
    link
    fedilink
    arrow-up
    7
    arrow-down
    1
    ·
    10 hours ago

    They’re still much closer to token predictors than any sort of intelligence. Even the latest models “with reasoning” still can’t answer basic questions most of the time and just ends up spitting back out the answer straight out of some SEO blogspam. If it’s never seen the answer anywhere in its training dataset then it’s completely incapable of coming up with the correct answer.

    Such a massive waste of electricity for barely any tangible benefits, but it sure looks cool and VCs will shower you with cash for it, as they do with all fads.

  • hotatenobatayaki@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    10 hours ago

    You’re trying to graph something that you can’t quantify.

    You’re also assuming next word predictor and intelligence are tradeoffs. They could as well be the same.

    • Randomgal@lemmy.ca
      link
      fedilink
      arrow-up
      3
      ·
      16 hours ago

      I think you point out the main issue here. Wtf is intelligence as defined by this axis? IQ? Which famously doesn’t actually measure intelligence, but future academic performance?

    • Todd Bonzalez@lemm.ee
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      15 hours ago

      Human intelligence created language. We taught it to ourselves. That’s a higher order of intelligence than a next word predictor.

      • Sl00k@programming.dev
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 hours ago

        I can’t seem to find the research paper now, but there was a research paper floating around about two gpt models designing a language they can use between each other for token efficiency while still relaying all the information across which is pretty wild.

        Not sure if it was peer reviewed though.

      • sunbeam60@lemmy.one
        link
        fedilink
        arrow-up
        2
        ·
        14 hours ago

        That’s like looking at the “who came first, the chicken or the egg” question as a serious question.

  • gandalf_der_12te@lemmy.blahaj.zone
    link
    fedilink
    arrow-up
    1
    ·
    11 hours ago

    Are you interested in this from a philosophical perspective or from a practical perspective?

    From a philosophical perspective:

    It depends on what you mean by “intelligent”. People have been thinking about this for millennia and have come up with different answers. Pick your preference.

    From a practical perspective:

    This is where it gets interesting. I don’t think we’ll have a moment where we say “ok now the machine is intelligent”. Instead, it will just slowly and slowly take over more and more jobs, by being good at more and more tasks. And just so, in the end, it will take over a lot of human jobs. I think people don’t like to hear it due to the fear of unemployedness and such, but I think that’s a realistic outcome.

  • nickwitha_k (he/him)@lemmy.sdf.org
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    edit-2
    9 hours ago

    Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor.

    They are good at sounding intelligent. But, LLMs are not intelligent and are not going to save the world. In fact, training them is doing a measurable amount of damage in terms of GHG emissions and potable water expenditure.

  • sunbeam60@lemmy.one
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    14 hours ago

    I hold a very strong hypothesis, which I’ve not seen any data contradict yet, that intelligence is only possible with formal language and symbolics and therefore formal language and intelligence is very hard to separate. I don’t think one created the other; they evolved together.

  • LarmyOfLone@lemm.ee
    link
    fedilink
    arrow-up
    3
    ·
    20 hours ago

    The way I would classify it is if you could somehow extract the “creative writing center” from a human brain, you’d have something comparable to to a LLM. But they lack all the other bits, and reason and learning and memory, or badly imitate them.

    If you were to combine multiple AI algorithms similar in power to LLM but designed to do math, logic and reason, and then add some kind of memory, you probably get much further towards AGI. I do not believe we’re as far from this as people want to believe, and think that sentience is on a scale.

    But it would still not be anchored to reality without some control over a camera and the ability to see and experience reality for itself. Even then it wouldn’t understand empathy as anything but an abstract concept.

    My guess is that eventually we’ll create a kind of “AGI compiler” with a prompt to describe what kind of mind you want to create, and the AI compiler generates it. A kind of “nursing AI”. Hopefully it’s not about profit, but a prompt about it learning to be friends with humans and genuinely enjoy their company and love us.

  • Scrubbles@poptalk.scrubbles.tech
    link
    fedilink
    English
    arrow-up
    70
    arrow-down
    7
    ·
    2 days ago

    That’s literally how llma work, they quite literally are just next word predictors. There is zero intelligence to them.

    It’s literally a while token is not “stop”, predict next token.

    It’s just that they are pretty good at predicting the next token so it feels like intelligence.

    So on your graph, it would be a vertical line at 0.

      • Scrubbles@poptalk.scrubbles.tech
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        4
        ·
        1 day ago

        yeah yeah I’ve heard this argument before. “What is learning if not like training.” I’m not going to define it here. It doesn’t “think”. It doesn’t have nuance. It is simply a prediction engine. A very good prediction engine, but that’s all it is. I spent several months of unemployment teaching myself the ins and outs, developing against llms, training a few of my own. I’m very aware that it is not intelligence. It is a very clever trick it pulls off, and easy to fool people that it is intelligence - but it’s not.

        • SorteKaninA
          link
          fedilink
          arrow-up
          2
          ·
          12 hours ago

          But how do you know that the human brain is not just a super sophisticated next-thing predictor that by being super sophisticated manages to incorporate nuance and all that stuff to actually be intelligent? Not saying it is but still.

          • Scrubbles@poptalk.scrubbles.tech
            link
            fedilink
            English
            arrow-up
            2
            ·
            12 hours ago

            Because we have reason, understanding. Take something as simple as the XY problem. Humans understand that there are nuances to prompts and questions. I like the XY because a human knows to step back and ask “what are you really trying to do?”. AI doesn’t have that capability, it doesn’t have reasoning to say “maybe your approach is wrong”.

            So, I’m not the one to define what it is or on what scale. But I can say that it’s not human intelligence.

    • webghost0101@sopuli.xyz
      link
      fedilink
      arrow-up
      7
      arrow-down
      10
      ·
      edit-2
      2 days ago

      This is true if you describe a pure llm, like gpt3

      However systems like claude, gpt4o and 1o are far from just a single llm, they are a blend of tailored llms, machine learning some old fashioned code to weave it all together.

      Op does ask “modern llm” so technically you are right but i believed they did mean the more advanced “products”

      Though i would not be able to actually answer ops questions, ai is hard to directly compare with a human.

      In most ways its embarrassingly stupid, in other it has already surpassed us.

      • fartsparkles@sh.itjust.works
        link
        fedilink
        arrow-up
        12
        arrow-down
        1
        ·
        1 day ago

        None of which are intelligence, and all of which are catered towards predicting the next token.

        All the models have a total reliance on data and structure for inference and prediction. They appear intelligent but they are not.

        • webghost0101@sopuli.xyz
          link
          fedilink
          arrow-up
          1
          arrow-down
          5
          ·
          edit-2
          1 day ago

          How is good old fashioned code comparing outputs to a database of factual knowledge “predicting the next token” to you. Or reinforcement relearning and token rewards baked into models.

          I can tell you have not actually tried to work with professional ai or looked at the research papers.

          Yes none of it is “intelligent” but i would counter that with neither are human beings, we dont even know how to define intelligence.

      • justOnePersistentKbinPlease@fedia.io
        link
        fedilink
        arrow-up
        1
        ·
        1 day ago

        No, unfortunately you are wrong.

        Gpt4 is a better version of gpt3.

        The brand new one that is allegedly “unhackable” just has a role hierarchy providing rules and that hasn’t been fulled tested in the wild yet.

        • webghost0101@sopuli.xyz
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          3 hours ago

          First, did you read even the research papers?

          Secondly, none are out that are actually immune to jailbreaking lol, Where did that claim come from?

          Gpt4 is just an llm. Indeed the better version of gpt3

          Gpt4o and 1o (claude-sonnet possibly also) rely on the generative capacities of the gpt4 model but there is allot more going under the hood that is not simply “generate the next token”

          We all agree that a pure text predictor are not at all intelligent.

          The discussion at hand is wether the current frontier of ai has moved the needle up. And i still would call it pretty dumb, but moving that needle, it did. Somewhere around (x2y0.5) if i have to use the meme. Stating its (0,0) just means people aren’t interested enough to pay attention, that these aren’t just llm anymore. That’s their right but i prefer people stopped joining the discussion so uninformed.

  • WatDabney@sopuli.xyz
    link
    fedilink
    arrow-up
    47
    arrow-down
    5
    ·
    1 day ago

    Intelligence is a measure of reasoning ability. LLMs do not reason at all, and therefore cannot be categorized in terms of intelligence at all.

    LLMs have been engineered such that they can generally produce content that bears a resemblance to products of reason, but the process by which that’s accomplished is a purely statistical one with zero awareness of the ideas communicated by the words they generate and therefore is not and cannot be reason. Reason is and will remain impossible at least until an AI possesses an understanding of the ideas represented by the words it generates.

  • mashbooq@lemmy.world
    link
    fedilink
    arrow-up
    24
    arrow-down
    1
    ·
    edit-2
    1 day ago

    There’s a preprint paper out that claims to prove that the technology used in LLMs will never be able to be extended to AGI, due to the exponentially increasing demand for resources they’d require. I don’t know enough formal CS to evaluate their methods, but to the extent I understand their argument, it is compelling.

  • Gamma@beehaw.org
    link
    fedilink
    English
    arrow-up
    40
    ·
    2 days ago

    They’re still word predictors. That is literally how the technology works

      • vrighter@discuss.tchncs.de
        link
        fedilink
        arrow-up
        3
        arrow-down
        3
        ·
        1 day ago

        no, they are not. try showing an ai a huge number of pictures of cars from the front. Then show them one car from the side, and ask them what it is.

        Show a human one picture of a car from the front, then the one from the side and ask them what it is.

        • novibe@lemmy.ml
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          1 day ago

          What if the human had never seen or heard of anything similar to cars?

          I bet it’d be confused as much as the llm.

          • vrighter@discuss.tchncs.de
            link
            fedilink
            arrow-up
            3
            arrow-down
            1
            ·
            1 day ago

            That’s why you show him one, before asking what that same car viewed from a different angle is.

            I had never seen a recumbent bike before. I only needed to see one to know and recognize one whenever I see one. Even one with a different color or make and model. The human brain definitely works differently.

            • novibe@lemmy.ml
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              1 day ago

              You know what bicycle are though. And you’re heard of recumbent bikes or things similar to it.

              If you had never heard of anything similar at all to bikes, and saw a picture of a recumbent bike from the front only, you’d probably think “ I have no fucking idea what that is”.

              Idk man, weird for you to think humans can kinda learn fully about something without all the required context.

              • vrighter@discuss.tchncs.de
                link
                fedilink
                arrow-up
                1
                arrow-down
                1
                ·
                1 day ago

                you keep missing the fact that I don’t know out of nowhere. You would have just shown me one and told me what it was. Yes of course I’d be able to tell you what it was. You just taught me. With one example.

                • novibe@lemmy.ml
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  edit-2
                  53 minutes ago

                  To understand a recumbent bicycle you have to understand bicycles. To understand bicycles you have to understand wheels. You have to understand humans, and human transportation. What IS transportation. What are roads. What is a pedal. What is steering. How physics works for objects in motion. Etc etc etc etc.

                  You truly underestimate the amount of context and previous knowledge you need to understand even the simplest things.

  • criitz@reddthat.com
    link
    fedilink
    arrow-up
    14
    ·
    edit-2
    1 day ago

    Shouldn’t those be opposite sides of the same axis, not two different axes? I’m not sure how this graph should work.

  • Nomecks@lemmy.ca
    link
    fedilink
    arrow-up
    13
    arrow-down
    2
    ·
    1 day ago

    I think the real differentiation is understanding. AI still has no understanding of the concepts it knows. If I show a human a few dogs they will likely be able to pick out any other dog with 100% accuracy after understanding what a dog is. With AI it’s still just stasticial models that can easily be fooled.

  • lunarul@lemmy.world
    link
    fedilink
    arrow-up
    14
    arrow-down
    1
    ·
    1 day ago

    Somewhere on the vertical axis. 0 on the horizontal. The AGI angle is just to attract more funding. We are nowhere close to figuring out the first steps towards strong AI. LLMs can do impressive things and have their uses, but they have nothing to do with AGI

    • Michal@programming.dev
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      19 hours ago

      AGI could be possible if a new breakthrough is made. Currently LLMs are just pretty good text predictor, and any intelligence exhibited by them is because they are trained on texts exhibiting intelligence (written by humans) . Make a large enough model, and it will seem like an intelligent being.

      • lunarul@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        18 hours ago

        Make a large enough model, and it will seem like an intelligent being.

        That was already true in previous paradigms. A non-fuzzy non-neural-network algorithm large and complex enough will seem like an intelligent being. But “large enough” is beyond our resources and processing time for each response would be too long.

        And then you get into the Chinese room problem. Is there a difference between seems intelligent and is intelligent?

        But the main difference between an actual intelligence and various algorithms, LLMs included, is that intelligence works on its own, it’s always thinking, it doesn’t only react to external prompts. You ask a question, you get an answer, but the question remains at the back of its mind, and it might come back to you 10min later and say you know, I’ve given it some more thought and I think it’s actually like this.

      • wewbull@feddit.uk
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 day ago

        A next word predictor algorithm is still a next word predictor algorithm even if you change it’s training algorithm. To think that a LLM will eventually lead to intelligence inherently asserts that intelligence comes from the ability to use language.

        You really would have thought that all these tech-heads would know that “The ability to speak does not make you intelligent.”

        We know, through studies on actual humans, that language filters, constrains and quantises our thoughts process, and that different languages do this in different ways. Language harms our ability to reason. We’ve internalised it to such a degree that it now forces our ideas to fit into what the language can express. However, the ability to share our thoughts with others and collaborate is a massive boon for us as a species.

        The whole this field is drawing pictures on the walls of Plato’s cave, trying to mimick the shadows being cast in from outside. Their drawings might look superficially similar to their inspiration, but they’re a poor imitation and that’s all they will ever be.

        • Communist@lemmy.frozeninferno.xyz
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          edit-2
          1 day ago

          Is it not the case that predicting the next word often requires reasoning about the next word in order to have any form of accuracy?

          And that if you select for better and better prediction, you have to also select for reasoning?

            • Communist@lemmy.frozeninferno.xyz
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 day ago

              Did you watch the video I linked?

              It seems to be essentially about a way to trick them into doing general reasoning, and a direct response to your comment.

              • It’s not a direct response.

                First off, the video is pure speculation, the author doesn’t really know how it works either (or at least doesn’t seem to claim to know). They have a reasonable grasp of how it works, but what they believe it implies may not be correct.

                Second, the way O1 seems to work is that it generates a ton of less-than-ideal answers and picks the best one. It might then rerun that step until it reaches a sufficient answer (as the video says).

                The problem with this is that you still have an LLM evaluating each answer based on essentially word prediction, and the entire “reasoning” process is happening outside any LLM; it’s thinking process is not learned, but “hardcoded”.

                We know that chaining LLMs like this can give better answers. But I’d argue this isn’t reasoning. Reasoning requires a direct understanding of the domain, which ChatGPT simply doesn’t have. This is explicitly evident by asking it questions using terminology that may appear in multiple domains; it has a tendency of mixing them up, which you wouldn’t do if you truly understood what the words mean. It is possible to get a semblance of understanding of a domain in an LLM, but not in a generalised way.

                It’s also evident from the fact that these AIs are apparently unable to come up with “new knowledge”. It’s not able to infer new patterns or theories, it can only “use” what is already given to it. An AI like this would never be able to come up with E=mc2 if it hasn’t been fed information about that formula before. It’s LLM evaluator would dismiss any of the “ideas” that might come close to it because it’s never seen this before; ergo it is unlikely to be true/correct.

                Don’t get me wrong, an AI like this may still be quite useful w.r.t. information it has been fed. I see the utility in this, and the tech is cool. But it’s still a very, very far cry from AGI.