You can take “justifiable” to mean whatever you feel it means in this context. e.g. Morally, artistically, environmentally, etc.

  • Canopyflyer@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 day ago

    LLM’s have their use, there is no doubt about that. I’m in the middle of creating a home brew campaign for my D&D group and unfortunately I’m a lousy artist and I wanted a few things visualized. Well, I used a photo generating AI to create something that had the visual I wanted. I’m going to use it for my campaign and it will probably just sit on my hard drive after I’m done.

    My employer is rolling out AI and is asking us to find places to insert it into our workflows. I am doing that with my team, but none of us are really sure if it will be of any benefit.

    The problem right now is we’re at the stage where idiots are convinced it is something that it is not and they have literally thrown 10’s of billions of dollars at it. Now… They are staring at the wide abyss that is the amount of money they invested vs the amount of money people are willing to pay for it.

    I’ve seen arguments for and against the presence of an AI bubble… Personally, I think it’s a bubble that’s so large that it will take down several long established computer industry manufacturers when if pops. Those that are arguing its absence probably have large investments that they do not want to see fail.

    • jabberwock@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      1
      ·
      1 day ago

      LLMs specifically are great for intermediate use cases. You had a campaign in mind, but needed help with visuals. I was designing a piece of jewelry and had a series of reference images. Fed all those into a VLM and got something closer to my imagination, but still worked with a jeweler to realize the final product.

      These tools are best when you have a foundation of knowledge and need a little extra guidance, but fall off when you get to deep expertise. I’ve used them to troubleshoot my server but I already had a basic understanding of how a config should look. I also wouldn’t trust an LLM to properly configure something like crypto for it.

      To me, the biggest ethical concerns surround the training and creation of LLMs - stealing artists’ work to train them, energy usage, etc. I suppose in using the models I’m creating ongoing demand for them, so I’m not sure the answer. The best I’ve seen so far is what Anthropic used to espouse, no new frontier models until we can guarantee safety. And I’d throw in “utility”. Train new models when people are actually using them and clamoring for new use cases, not because a bunch of private equity shows line go up.

      • Apytele@sh.itjust.works
        link
        fedilink
        arrow-up
        1
        ·
        17 hours ago

        Literally everything I’ve vibe coded the #1 security feature is local only storage. I trust it naught with security LOL.