• MystikIncarnate@lemmy.ca
    link
    fedilink
    English
    arrow-up
    13
    ·
    6 days ago

    A person who hasn’t debugged any code thinks programmers are done for because of “AI”.

    Oh no. Anyways.

  • gravitas_deficiency@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    24
    ·
    7 days ago

    Lmfao I love these threads. “I haven’t built anything myself with the thing I’m claiming makes you obsolete but trust me it makes you obsolete”

  • maplebar@lemmy.world
    link
    fedilink
    arrow-up
    17
    ·
    edit-2
    7 days ago

    AI isn’t ready to replace just about anybody’s job, and probably never will be technically, economically or legally viable.

    That said, the c-suit class are certainly going to try. Not only do they dream of optimizing all human workers out of every workforce, they also desperately need to recoup as much of the sunk cost that they’ve collectively dumped into the technology.

    Take OpenAI for example, they lost something like $5,000,000,000 last year and are probably going to lose even more this year. Their entire business plan relies on at least selling people on the idea that AI will be able to replace human workers. The minute people realize that OpenAI isn’t going to conquer the world, and instead end up as just one of many players in the slop space, the entire bottom will fall out of the company and the AI bubble will burst.

  • Wanpieserino@lemm.ee
    link
    fedilink
    arrow-up
    21
    arrow-down
    1
    ·
    edit-2
    7 days ago

    My mate is applying to Amazon as warehouse worker. He has an IT degree.

    My coworker in the bookkeeping department has two degrees. Accountancy and IT. She can’t find an IT job.

    At the other side though, my brother, an experienced software developer, is earning quite a lot of money now.

    Basically, the industry is not investing in new blood.

      • boonhet@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        7 days ago

        My company was desperate to find a brand new dev straight out of the oven we could still mold to our sensibilities late last year when everything seemed doomed. Yes, it was one hire out of like 10 interviewed candidates, but point is, there are companies still hiring. Our CTO straight up judges people who use an LLM and don’t know how the code actually works. Mr. “Just use an AI agent” would never get the job.

      • Wanpieserino@lemm.ee
        link
        fedilink
        arrow-up
        1
        ·
        7 days ago

        Don’t you worry, my job will be replaced by AI as well. By 2026 peppol invoices will be enforced in Belgium. Reducing bookkeepers their workload.

        ITers replacing my job: 😁😁😁

        ITers replacing their own jobs: 😧😧😧

    • fuck_u_spez_in_particular@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      7 days ago

      Basically, the industry is not investing in new blood.

      Yeah I think it makes sense out of an economic motivation. Often the code-quality of a junior is worse than that of an AI, and a senior has to review either, so they could just directly prompt the junior task into the AI.

      The experience and skill to quickly grasp code and intention (and having a good initial idea where it should be going architecturally) is what is asked, which is obviously something that seniors are good at.

      It’s kinda sad that our profession/art is slowly dying out because juniors are slowly replaced by AI.

      • Terrasque@infosec.pub
        link
        fedilink
        arrow-up
        2
        ·
        7 days ago

        Yeah, I’ve been seeing the same. Purely economically it doesn’t make sense with junior developers any more. AI is faster, cheaper and usually writes better code too.

        The problem is that you need junior developers working and getting experience, otherwise you won’t get senior developers. I really wonder how development as a profession will be in 10 years

  • digitalnuisance@infosec.pub
    link
    fedilink
    arrow-up
    15
    ·
    edit-2
    7 days ago

    I had a dude screaming pretty much the same thing at me yesterday on here (on a different account), despite the fact that I’m senior-level, near the top of my field and that all the objective data as well as anecdotal reports from tons of other people says otherwise. Like, okay buddy, sure. People seem to just like fighting things online to feel better about themselves, even if the thing they’re fighting doesn’t really exist.

    • Event_Horizon@lemmy.world
      link
      fedilink
      arrow-up
      6
      ·
      7 days ago

      I’m a senior BA working on a project to replace some outdated software with a new booking management and payment system. One of our minor stakeholders is an overly eager tech bro who insists on bringing up AI in every meeting, he’s gone as far as writing up and sending proposals to myself and project leads.

      We all just roll our eyes when a new email arrives. Especially when there’s almost no significant detail in these proposals, it’s all conjecture based of what he’s read online…on tech bro websites.

      Oh and the best part, this guy has no experience in system development or design or anything AI related. He doesn’t even work in IT. But he researchs AI in his spare time and uses it as a side hustle…

  • Aggravationstation@feddit.uk
    link
    fedilink
    arrow-up
    15
    ·
    edit-2
    7 days ago

    AI isn’t ready to replace programmers, engineers or IT admins yet. But let’s be honest if some project manager or CTO somewhere hasn’t already done it they’re at least planning it.

    Then eventually to save themselves or out of sheer ignorance they’ll blame the chaos that results on the few remaining people who know what they’re doing because they won’t be able to admit or understand the fact that the bold decision they took to “embrace” AI and increase the company’s bottom line which everyone else in their management bubble believes in has completely mangled whatever system their company builds or uses. More useful people will get fired and more actual work will get shifted to AI. But because that’ll still make the number go up the management types will look even better and the spread of AI will carry on. Eventually all systems will become an unwieldy mess nobody can even hope to repair.

    This is just IT, I’m pretty sure most other industries will eventually suffer the same fate. Global supply chains will collapse and we’ll all get sent back to the dark ages.

    TL,DR: The real problem with AI isn’t that it’ll become too powerful and choose to kill us, but that corporate morons will overestimate how powerful it already is and that will cause our eventual downfall.

    • Terrasque@infosec.pub
      link
      fedilink
      arrow-up
      3
      ·
      7 days ago

      AI isn’t ready to replace programmers, engineers or IT admins yet.

      On the other hand… it’s been about 2.5 years since chatgpt came out, and it’s gone from you being lucky it could write a few python lines without errors to being able to one shot a mobile phone level complexity game, even with self hosted models.

      Who knows where it’ll be in a few years

  • samus12345@lemm.ee
    link
    fedilink
    English
    arrow-up
    6
    ·
    6 days ago

    As an end user with little knowledge about programming, I’ve seen how hard it is for programmers to get things working well many times over the years. AI as a time saver for certain simple tasks, sure, but no way in hell they’ll be replacing humans in my lifetime.

  • arc@lemm.ee
    link
    fedilink
    arrow-up
    7
    ·
    6 days ago

    AI is certainly a very handy tool and has helped me out a lot but anybody who thinks “vibe programming” (i.e. programming from ignorance) is a good idea or will save money is woefully misinformed. Hire good programmers, let them use AI if they like, but trust the programmer’s judgement over some AI.

    That’s because you NEED that experience to notice the AI is outputting garbage. Otherwise it looks superficially okay but the code is terrible, or fragile, or not even doing what you asked it properly. e.g. if I asked Gemini to generate a web server with Jetty it might output something correct or an unholy mess of Jetty 8, 9, 10, 11, 12 with annotations and/or programmatic styles, or the correct / incorrect pom dependencies.

  • OmgItBurns@discuss.online
    link
    fedilink
    English
    arrow-up
    10
    ·
    7 days ago

    AI is a tool, Ashish is 100% correct in that it may do some things for developers but ultimately still needs to be reviewed by people who know what they’re doing. This is closer to the change from punch cards to writing code directly on a computer than making software developers obsolete.

  • lalala@lemmy.world
    link
    fedilink
    arrow-up
    9
    ·
    7 days ago

    English isn’t my first language, so I often use translation services. I feel like using them is a lot like vibe coding — very useful, but still something that needs to be checked by a human.

  • Lovable Sidekick@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    edit-2
    7 days ago

    I’ve always said as a software developer that our longterm job is to program ourselves out of a job. In fact, in the long term EVERYBODY is “cooked” as automation becomes more and more capable. The eventual outcome will be that nobody will have to work. AI in its present state isn’t ready at all to replace programmers, but it can be a very helpful assistant.

    • DarkenLM@kbin.earth
      link
      fedilink
      arrow-up
      6
      ·
      7 days ago

      Management can’t blame AI when shit hits the fan, though. We’ll be fine. Either that or everything just collapses back into dust, which doesn’t sound so bad in the current times.

      • Lovable Sidekick@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        7 days ago

        That’s the beauty of AI tho - AI shit rolls uphill, until it hits the manager who imposed the decision to use it (or their manager, or even their manager).

    • fuck_u_spez_in_particular@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      7 days ago

      but it can be a very helpful assistant.

      can, but usually when stuff gets slightly more complex, being a fast typewriter is usually more efficient and results in better code.

      I guess it really depends on the aspiration for code-quality, complexity (yes it’s good at generating boilerplate). If I don’t care about a one-time use script that is quickly written in a prompt I’ll use it.

      Working on a big codebase, I don’t even get the idea to ask an AI, you just can’t feed enough context to the AI that it’s really able to generate meaningful code…

      • Lovable Sidekick@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        7 days ago

        I actually don’t write code professionally anymore, I’m going on what my friend says - according to him he uses chatGPT every day to write code and it’s a big help. Once he told it to refactor some code and it used a really novel approach he wouldn’t have thought of. He showed it to another dev who said the same thing. It was like, huh, that’s a weird way to do it, but it worked. But in general you really can’t just tell an AI “Create an accounting system” or whatever and expect coherent working code without thoroughly vetting it.

        • fuck_u_spez_in_particular@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          6 days ago

          I’ll use it also often. But when the situation is complex and needs a lot of context/knowledge of the codebase (which at least for me is often the case) it seems to be still worse/slower than just coding it yourself (it doesn’t grasp details). Though I like how quick I can come up with quick and dirty scripts (in Rust for the Lulz and speed/power).

        • fuck_u_spez_in_particular@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          6 days ago

          Ughh I tried the gemini model and I’m not too happy with the code it came up with, there’s a lot of intrinsities and concepts that the model doesn’t grasp enough IMO. That said I’ll reevaluate this continuously converting large chunks of code often works ok…

          • Terrasque@infosec.pub
            link
            fedilink
            arrow-up
            1
            ·
            6 days ago

            Well, it wasn’t a comment on the quality of the model, just that the context limitation has already been largely overcome by one company, and others will probably follow (and improve on it further) over time. Especially as “AI Coding” gets more marketable.

            That said, was this the new gemini 2.5 pro you tried, or the old one? I haven’t tried the new model myself, but I’ve heard good things about it.

    • skuzz@discuss.tchncs.de
      link
      fedilink
      arrow-up
      3
      arrow-down
      1
      ·
      edit-2
      7 days ago

      I had an AI render a simple diagram for a presentation with explicit instructions. It rendered a Rube Goldberg nonsense graphic. I included it anyway for the lulz. Sure, they will get better, and maybe some day be almost as useful as the Enterprise computer. No way they’ll be Lt. Cmdr. Data this century.

  • Ronno@feddit.nl
    link
    fedilink
    English
    arrow-up
    4
    ·
    6 days ago

    The way I see it, there are two types of developers we should take into consideration for this discussion:

    • Software Engineers
    • Code editors

    Most “programmers” these days are really just code editors, they know how to search stack overflow for some useful pointers, copy that code and edit it to what they need. That is absolutely fine, this advances programming in so many ways. But the software engineers are the people that actually answer the stack overflow questions with detailed answers. These engineers have a more advanced skillset in problem solving for specific coding frameworks and languages.

    When people say: programmers are cooked, I keep thinking that they mean code editors, not software engineers. Which is a similar trend in basically all industries in relation with AI. Yes, AI has the potential to make some jobs in health care obsolete (e.g. radiologist), but that doesn’t mean we no longer need surgeons or domain expert doctors. Same thing applies to programming.

    So if you are a developer today, ask yourself the following: Do actually know my stuff well, am I an expert? If the answer is no, and you’re basically a code editor (which again, is fine), then you should seriously consider what AI means for your job.

    • stormeuh@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      6 days ago

      I agree with the overall sentiment, but I’d like to add two points:

      1. Everyone starts off as a code editor, and through a combination of (self-)education and experience can become a software engineer.

      2. To the point of code editors having to worry about LLM’s taking their job, I agree, but I don’t think it will be as over the top as people literally being replaced by “AI agents”. Rather I think it will be a combination of code editors becoming more productive through use of LLMs, decreasing the demand for code editors, and lay people (i.e. almost no code skills) being able to do more through LLMs applied in the right places, like some website builders are doing now.

    • Jankatarch@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      6 days ago

      If the “code editor” uses AI they will never become a software engineer.

      “Oh I will just learn by asking AI to explain” that’s not happening. You won’t learn how to come.up with a solution. Mathematiciams know better than anyone you can’t just memorize how the professor does stuff and call yourself a problem solver. Now go learn the heruistic method.

      As much as people hate it, stackoverflow people rarely give the answer directly. They usually tell you easier alternative methods or how to solve a similar problem with explanation.

      They way it will work is that every single college student that relies on AI and gets away with “academic dishonesty, the tool” will become terrible programmers that can’t think for themselves or read a single paragraph of documentation. Similar consequences for inexperienced developers.

  • Phoenixz@lemmy.ca
    link
    fedilink
    arrow-up
    3
    ·
    6 days ago

    The day that AI can program perfectly is the day it can improve the itself perfectly and it’s the day that we’ll all be fucked.

    I personally vote for some sort of direct brain interface (no Elmo, you’re not allowed to play) that DOES allow direct recall of queries but does NOT allow ads ffs) that allows us to grow with AI in intelligence. If you can’t beat em (we can’t), join em.

    • borth@sh.itjust.works
      link
      fedilink
      arrow-up
      1
      ·
      6 days ago

      I highly doubt some of these rich fucks would pass up an opportunity to put ads straight into people’s brains.