• Free_Opinions@feddit.uk
    link
    fedilink
    English
    arrow-up
    53
    arrow-down
    4
    ·
    15 days ago

    We’ve had definition for AGI for decades. It’s a system that can do any cognitive task as well as a human can or better. Humans are “Generally Intelligent” replicate the same thing artificially and you’ve got AGI.

    • LifeInMultipleChoice@lemmy.ml
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      2
      ·
      15 days ago

      So if you give a human and a system 10 tasks and the human completes 3 correctly, 5 incorrectly and 3 it failed to complete altogether… And then you give those 10 tasks to the software and it does 9 correctly and 1 it fails to complete, what does that mean. In general I’d say the tasks need to be defined, as I can give very many tasks to people right now that language models can solve that they can’t, but language models to me aren’t “AGI” in my opinion.

    • rational_lib@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      14 days ago

      So then how do we define natural general intelligence? I’d argue it’s when something can do better than chance at solving a task without prior training data particular to that task. Like if a person plays tetris for the first time, maybe they don’t do very well but they probably do better than a random set of button inputs.

      Likewise with AGI - say you feed an LLM text about the rules of tetris but no button presses/actual game data and then hook it up to play the game. Will it do significantly better than chance? My guess is no but it would be interesting to try.

      • Free_Opinions@feddit.uk
        link
        fedilink
        English
        arrow-up
        4
        ·
        14 days ago

        It should be able to perform any cognitive task a human can. We already have AI systems that are better at individual tasks.

    • ipkpjersi@lemmy.ml
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      4
      ·
      edit-2
      15 days ago

      That’s kind of too broad, though. It’s too generic of a description.

      • Entropywins@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        2
        ·
        15 days ago

        The key word here is general friend. We can’t define general anymore narrowly, or it would no longer be general.

      • CheeseNoodle@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        14 days ago

        That’s the idea, humans can adapt to a broad range of tasks, so should AGI. Proof of lack of specilization as it were.