Bill Gates feels Generative AI has plateaued, says GPT-5 will not be any better::The billionaire philanthropist in an interview with German newspaper Handelsblatt, shared his thoughts on Artificial general intelligence, climate change, and the scope of AI in the future.

  • ShittyBeatlesFCPres@lemmy.world
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    1
    ·
    10 months ago

    I’m not sure I’d say it’s plateaued today but I definitely think machine learning is going to hit a wall soon. Some tech keeps improving until physical limits stop progress but I see generative AI as being more like self-driving cars where the “easy” parts end up solved but the last 10% is insanely hard.

    There’s also the economic reality of scaling. Maybe the “hard” problems could, in theory, be easily solved with enough compute power. We’ll eventually solve those problems but it’s going to be on Nvidia’s timeline, not OpenAI’s.

    • erwan@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      10 months ago

      Yes, especially when you consider that the human brain runs on 15W of power!

  • scarabic@lemmy.world
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    1
    ·
    10 months ago

    Let me save you a click: he doesn’t say anything interesting about why he thinks this.

  • astronaut_sloth@mander.xyz
    link
    fedilink
    English
    arrow-up
    37
    arrow-down
    16
    ·
    10 months ago

    Cool, Bill Gates has opinions. I think he’s being hasty and speaking out of turn and only partially correct. From my understanding, the “big innovation” of GPT-4 was adding more parameters and scaling up compute. The core algorithms are generally agreed to be mostly the same from earlier versions (not that we know for sure since OpenAI has only released a technical report). Based on that, the real limit on this technology is compute and number of parameters (as boring as that is), and so he’s right that the algorithm design may have plateaued. However, we really don’t know what will happen if truly monster rigs with tens-of-trillions of parameters are used when trained on the entirety of human written knowledge (morality of that notwithstanding), and that’s where he’s wrong.

    • Vlyn@lemmy.zip
      link
      fedilink
      English
      arrow-up
      68
      arrow-down
      6
      ·
      10 months ago

      You got it the wrong way around. We already have a ton of compute and what this kind of AI can do is pretty cool.

      But adding more compute power and parameters won’t solve the inherent problems.

      No matter what you do, it’s still just a text generator guessing the next best word. It doesn’t do real math or logic, it gets basic things wrong and hallucinates new fake facts.

      Sure, it will get slightly better still, but not much. You can throw a million times the power at it and it will still fuck up in just the same ways.

      • scarabic@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        ·
        edit-2
        10 months ago

        it’s still just a text generator guessing the next best word. It doesn’t do real math or logic, it gets basic things wrong and hallucinates new fake facts.

        If humans are any kind of yardstick here, I’d say all this is true of us too on many levels. The brain is a shortcut engine, not a brute force computer. It’s not solving equations to help you predict where that tennis ball will bounce next. It’s making guesses based on its corpus of past experience. Good enough guesses are frankly our brains’ bread and butter and most of us get through most days on little more than this.

        It’s true that we can do more. Some of us, anyway. How many people actually exercise math and logic though? Sometimes it seems like… not a lot. And how many people hallucinate fake facts? A lot.

        It’s much like evaluating self-driving cars. We may be tempted to say they’re just bloody awful, but so are human drivers.

        • Vlyn@lemmy.zip
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          10 months ago

          I’d say the majority of humans know what 2 + 2 is. Chat GPT doesn’t. As it found the answer in some texts it will tell you 4, but all it takes is you telling it that’s wrong and suddenly it’s 5. So even for the most simple math problem it’s extremely easy to throw the whole thing off. Which also means for any prompt you put in it can go in wildly wrong directions at times.

          And this is all with good input data, there’s plenty of trolls online and the data will only get worse (it already did, the original data up to 2021 was okayish, in the last year tons of crap was put out on top, some of it by Chat GPT itself. So the new model might input the crap it produced before, getting worse over time). The problem on top of that is that you don’t know the sources it used. If you ask about a recent event you might receive an insane answer it picked up from a right wing conspiracy site, you simply don’t know. There is no fact checking in place.

          It’s a stunningly good text generator, but that’s all it is and it ever will be, at least until they do much more than just add more compute power to it.

          • scarabic@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            10 months ago

            Hehe. I’m imagining sitting 100 human test subjects down in a lab setting and asking them what 2+2 is, and then telling them they’re wrong when they answer 4. I don’t know how many of them would guess again but I know it’s not zero. Meanwhile, GPT can probably give a better answer to any advanced math or science query than the majority of humans.

            I’m a writer and a language nerd and I watch people all the time use words incorrectly because they think they know what they mean, but they really don’t. They’re just regurgitating them in what they think is the same situation they heard them. They don’t “understand” the word and are just guessing and churning out crap.

            I don’t have a dog in this race but I think it’s interesting how people judge artificial intelligence with too much credit given to what goes on with human intelligence. Most people who say it’s “just a next word predictor” read that phrase somewhere and are regurgitating it, not at all dissimilarly to what LLMs do. They use phrases like “it doesn’t actually understand” without being able to define, with any clarity or precision, and without resorting to examples, what would actually impress them as real intelligence.

            • Vlyn@lemmy.zip
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              2
              ·
              10 months ago

              GPT can probably give a better answer to any advanced math or science query than the majority of humans

              Only if that answer is already out there and in the model. So pretty much a Google search away.

              GPT isn’t coming up with new math or science facts (at least not real ones).

              It literally is a word predictor, an insanely complex one, it’s the best way to describe it. If you start with layers, parameters and so on most people lose interest. But there’s some really good explanations around.

              Generic AI (real AI) has internal logic, can learn and improve itself and can do self motivated actions. Chat GPT can tell you exactly how to create an account and order something from Amazon, but despite being able to put that text out it will never be able to actually follow them itself.

              • scarabic@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                10 months ago

                Only if that answer is already out there

                Again, pretty similar to the vast majority of humans. How many times in your science education did you learn ann equation that you’d already figured out on your own previously?

                And to be fair, GPT doesn’t have hands and the ability to conduct experiments. So we have to, in a sense, judge its success on an apples to apples basis of what it, and we, do with the corpus of written knowledge.

                In contrast to humans, GPT has at least read it all ;) (I say this in jest - I know it doesn’t have access to everything, but humans are too lazy to read, for the most part, even things they have access to).

              • Blue_Morpho@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                10 months ago

                Only if that answer is already out there and in the model.

                That’s not true. I wanted a vba script for Excel. I don’t know vba or excel so I spent hours searching Google for help. There were explanations of functions but no working code. I tried GPT for the fun of it and it spit back working code. Code that was nowhere on the Internet.

                It was able to put together functions into working code based on the definition of functions, not simply cutting and pasting what somebody else had already written.

              • guacupado@lemmy.world
                cake
                link
                fedilink
                English
                arrow-up
                1
                ·
                10 months ago

                Chat GPT can tell you exactly how to create an account and order something from Amazon, but despite being able to put that text out it will never be able to actually follow them itself.

                This is a really good ELI5 explanation of its limit.

                • scarabic@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  9 months ago

                  How is it a definitional limit on its intelligence that it can’t use interfaces designed for people with hands? You also cannot send an http request with your lips no matter how you try - that’s just not an interface made for you.

                  Bots can 100% operate websites and take online actions, conduct quality tests, write fake reviews. That doesn’t mean they are intelligent. I just can’t see how it has any bearing either way whether ChatGPT can place an Amazon order.

                • Vlyn@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  10 months ago

                  That would still give it too much credit in that case. It’s purely an input output system. You put text in (the prompt), you get text out (the result). If there is no input from you, there is no output. It doesn’t have any intrinsic functionality that runs on its own.

                  Maybe a bit too much for an ELI5.

      • astronaut_sloth@mander.xyz
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        11
        ·
        10 months ago

        I mean, that’s more-or-less what I said. We don’t know the theoretical limits of how good that text generation is when throwing more compute at it and adding parameters for the context window. Can it generate a whole book that is fairly convincing, write legal briefs off of the sum of human legal knowledge, etc.? Ultimately, the algorithm is the same, so like you said, the same problems persist, and the definition of “better” is wishy-washy.

        • Vlyn@lemmy.zip
          link
          fedilink
          English
          arrow-up
          11
          arrow-down
          2
          ·
          10 months ago

          It will obviously get even better, but you’ll never be able to rely on it. Sure, 99.9% of that generated legal document will look perfect, till you overlook one sentence where the AI hallucinated. There is no fact checking in there, that’s the issue.

    • lorkano@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      10 months ago

      The problem is that between gpt 3 and 4 there is massive increase in number of parameters, but not massive increase in its abilities

    • OldWoodFrame@lemm.ee
      link
      fedilink
      English
      arrow-up
      6
      ·
      edit-2
      10 months ago

      Yeah and I think he may be scaling to like true AGI. Very possible LLMs just don’t become AGI, you need some extra juice we haven’t come up with yet, in addition to computational power no one can afford yet.

      • astronaut_sloth@mander.xyz
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        3
        ·
        10 months ago

        Except that scaling alone won’t lead to AGI. It may generate better, more convincing text, but the core algorithm is the same. That “special juice” is almost certainly going to come from algorithmic development rather than just throwing more compute at the problem.

        • 0ops@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          10 months ago

          See my reply to the person you replied to. I think you’re right that there will need to be more algorithmic development (like some awareness of its own confidence so that the network can say IDK instead of hallucinating its best guess). Fundamentally though, llm’s don’t have the same dimensions of awareness that a person does, and I think that that’s the main bottleneck of human-like understanding.

      • 0ops@lemm.ee
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        2
        ·
        edit-2
        10 months ago

        My hypothesis is that that “extra juice” is going to be some kind of body. More senses than text-input, and more ways to manipulate itself and the environment than text-output. Basically, right now llm’s can kind of understand things in terms of text descriptions, but will never be able to understand it the way a human can until it has all of the senses (and arguably physical capabilities) that a human does. Thought experiment: Presumably you “understand” your dog - can you describe your dog without sensory details, directly or indirectly? Behavior had to be observed somehow. Time is a sense too. EDIT: before someone says it, as for feelings I’m not really sure, I’m not a biology guy. But my guess is we sense our own hormones as well

        • LinuxSBC@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          First, they do have senses. For example, many LLMs can “see” images. Second, they’re actually pretty good at describing things. What they’re really bad at is analysis and logic, which is not related to senses at all.

          • 0ops@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            10 months ago

            I’m not so convinced that logic is completely unrelated to the senses. How did you learn to count, add, and subtract mentally? You used your fingers. I don’t know about you, but even though I don’t count my fingers anymore I still tend to “visualize” math operations. Would I be capable of that if I were born blind? Maybe I’d figure out how to do the same thing in a different dimension of awareness, but I have no doubt that being able to conceptualize visually helps my own logic. As for more complicated math, I can’t do that mentally either, I need a calculator and/or scratch paper. Maybe analogues to those can be implemented into the model? Maybe someone should just train a model on khan academy videos, and it’ll pick this stuff up emergently? I’m not saying that the ability to visualize is the only roadblock though, I’m sure that improvements could be made to the models themselves, but I bet that it’ll be key to human-like reasoning

    • scarabic@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      10 months ago

      I’ll listen to his opinions more than some, but unfortunately this article doesn’t say anything interesting about why he has this opinion. I guess the author supposes we will simply regard him as an oracle on name recognition alone.

  • Pxtl@lemmy.ca
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    1
    ·
    10 months ago

    I hope so. Theyve already got scary implications for creative parts of the economy.

    That said, we’re in the Cambrian explosion of the tech. As it plateaus, the next step will be enhanced tooling and convenience around it. Better inputs than just text, better, more applications in new spaces, etc.

  • oldfart@lemm.ee
    link
    fedilink
    English
    arrow-up
    11
    arrow-down
    3
    ·
    10 months ago

    Not a single comment yet stating how Gates is a great human being because of his foundation, and how all you haters should fuck the fuck off? sigh, let me the first one.

    • oldfart@lemm.ee
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      4
      ·
      10 months ago

      Just to make things extremely clear, the above comment has been sarcastic. He’s an awful person.

    • banneryear1868@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      2
      ·
      10 months ago

      You mean his tax haven?

      I mean they’ve done some good things, but the capitalist system that gave him his wealth is the same one that causes poverty and his foundation isn’t working to change that.

    • Random_Character_A@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      10 months ago

      Now now. He only hired assholes and monsters to execute immoral MS mob style tactics, while he played the great innocent altruist.

  • The Menemen!@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    edit-2
    10 months ago

    Maybe, but I am sure the tools the AIs can use will improve making the AIs jobs easier and thus the AI more efficient. I hope he is right tbh.

    Eww, as a long time Linux user I need to take a shower now. I feel dirty.

  • Metal Zealot@lemmy.ml
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    edit-2
    10 months ago

    Bill, you’re not even AT Microsoft anymore.

    Edit: AND, you think a box of Pizza Rolls cost nearly 3 times as much as they actually do. I’m not trusting your opinion, man

    • LinuxSBC@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      10 months ago

      Why does him being rich mean that his opinion on LLMs is invalid? If anything, LLMs require such vast amounts of money that he has more experience in that area because of his riches, not less.

  • Immersive_Matthew@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    4
    ·
    10 months ago

    Since I started using AI in March, it has not stop getting better and better with it seemingly accelerating not decelerating. If we are heading to a plateau, there are no signs of it yet. I am sure there will be plateaus, and maybe we have even had one earlier this year, but with the speed of AI development, it may mean a plateau of months versus the normal years. We will see. Bill has been saying a lot of things these past years that make no sense, unless you look at it from a money/market manipulation perspective.