Despite the rush to integrate powerful new models, about 5% of AI pilot programs achieve rapid revenue acceleration; the vast majority stall, delivering little to no measurable impact on P&L.

The research—based on 150 interviews with leaders, a survey of 350 employees, and an analysis of 300 public AI deployments—paints a clear divide between success stories and stalled projects.

  • Pennomi@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    4
    ·
    4 months ago

    To be fair, that also falls under the blanket of AI. It’s just not an LLM.

    • leisesprecher@feddit.org
      link
      fedilink
      arrow-up
      8
      arrow-down
      3
      ·
      4 months ago

      No, it does not.

      A deterministic, narrow algorithm that solves exactly one problem is not an AI. Otherwise Pythagoras would count as AI, or any other mathematical formula for that matter.

      Intelligence, even in terms of AI, means being able to solve new problems. An autopilot can’t do anything else than piloting a specific aircraft - and that’s a good thing.

      • wheezy@lemmy.ml
        link
        fedilink
        arrow-up
        6
        ·
        4 months ago

        Not sure why you’re getting downvoted. Well, I guess I do. AI marketing has ruined the meaning of the word to the extent that an if statement is “AI”.

        • leisesprecher@feddit.org
          link
          fedilink
          arrow-up
          4
          ·
          4 months ago

          To a certain extent, yes.

          ChatGPT was never explicitly trained to produce code or translate text, but it can do it. Not super good, but it manages some reasonable output most of the time.