I personally think it’s a lot of bs
As a software engineer/data scientist who has spent quite a while to find some good AI work, this sounds like absolute bullshit. Most companies don’t need AI. Prompt engineer seems like a niche thing, I can’t imagine that most companies really need someone who does that. It really frustrates me that these bullshit articles keep coming out without any sense or reason. AI is cool technology (imo), but currently it’s just the latest bait for CEOs, managers, etc. Somehow these kinda people are just so vulnerable for hype words without ever thinking more than as second about how to use it or whether it’s even useful.
deleted by creator
By using big data on the IOT of course!
To add to this, even when they need AI work, they don’t look for actual AI specialists. What they do is take one person who’s done some random ML work and a bunch of random junior devs and that’s the “AI team”.
That’s very closely related to something I’ve come to think about tech: nerd equivalency. If there is a computer involved, then a nerd is required and they are all interchangeable.
Basically, someone says “we’re not moving fast enough, hire another nerd!” and nobody in the chain of command or in the hiring process has a clue which particular skills are required, assuming that everyone can do everything.
That’s why so many corporate projects have what amounts to random people doing randomly assigned work producing insecure, unreliable products with obscure and even hostile UIs.
Not only that, but these “bootcamps” aren’t exactly going to be churning out the highly skilled people needed to really make good use of AI systems.
AI is cool technology (imo), but currently it’s just the latest bait for CEOs, managers, etc. Somehow these kinda people are just so vulnerable for hype words without ever thinking more than as second about how to use it or whether it’s even useful.
I think that’s a general problem with most technology that is fundamentally about computing.
People outside any field have only the barest grasp of that field, but the problems are so much worse as soon as computers are involved. They are so ubiquitous and so useful to so many people with little or no training or understanding that everyone just succumbs to a form of magical thinking.
Not needing isn’t the same as won’t hire and allocate vast amounts of resources.
deleted by creator
ML =/= AI. There are legit uses for ML that don’t have anything to do with LLMs and the cloud. I worked on an ML project 3 or 4 years ago to listen for fan noise that might indicate that it was about to fail soon. We trained a tiny GAN on good and bad noises. It runs on a tiny CPU, locally. Highly specialized work, and I have to imagine there are and will continue to be lots of similar opportunities to bring efficiencies by getting computers to make good observations and decisions - even if only about “simple” things like “does this thing seem like it’s about to break?”
ML literally is a subfield of AI
Fair enough. ML ⊆ AI then. But these days when everyone talking breathlessly about AI taking away jobs they’re almost always taking about LLMs. This article is about ML in particular which is a different discipline with different applications.
It’s not a different discipline, an LLM is an example of a machine learning model.
LLMs are neural networks which is literally ML.
The LLM designation refers to what they are trained to do.
GANs are so much fun and so tedious. I trained one on dungeon generation in college. It sucked but it worked in the end mostly. I dont know exactly how an LLM works but GANs are way different.
The most important line on that page:
“FORTUNE may receive compensation for some links to products and services on this website. Offers may be subject to change without notice.”
deleted by creator
It depends on what you’re calling AI. The LLM hype may die down, but Ml/AI in general has been continuing to grow and expand for well over a decade. It’s just unlikely that all the things being prophesied right now will come to fruition.
deleted by creator
deleted by creator
Not with the stuff we currently have.
Image generation is neat. But now there are so many badly generated images out there that any new model likely feeds on shitty data. Also the whole copyright debate for the images used.
LLM are neat. But there is no point in widespread adoption. Any model that is built to generate correct sentences but not correct content kind of wastes a lot of time for the user and nothing else. They are inpressive, they can be fun and sometimes they are really helpful - but you never know whether or not they hallucinate any given information.
Voice and Avatar generation is neat. Like genuinely neat. But you rgey are so easy to use already, you don’t nees that many specialists.
Here’s the article’s source: https://www3.weforum.org/docs/WEF_Future_of_Jobs_2023.pdf
That report’s data is a survey they sent out to companies. Quantising “so… what do you think is gonna happen?” seems… shonky?
This is the best summary I could come up with:
Many businesses across a variety of industries are spending more on AI—from Papa John’s to Canva—thus translating to a need for workers to have relevant skillsets such as knowing natural language processing, prompt engineering, and Python.
One broad example is online learning platform Udemy, which hosts dozens of offerings across a wide spectrum of experience level, length, and price.
Kara Sasse, chief product officer at Springboard, says the bootcamps are catered to fit the needs of those working professionals who are eager to upskill and succeed in increasingly AI-focused job environments.
“As ML and AI continue to transform every aspect of our lives, forward-thinking organizations must actively take inventory of potential skills gaps and look for professionals with the tools to succeed in this evolving landscape,” Sasse tells Fortune.
Fullstack Academy similarly offers an AI and ML bootcamp that covers fundamentals as well as emphasizes practical application, according to the company’s CEO, Nelis Parts.
Topic examples: Deep Learning with Keras and TensorFlow; Applied Data Science with Python; Essentials of Generative AI, Prompt Engineering, and ChatGPT
The original article contains 773 words, the summary contains 175 words. Saved 77%. I’m a bot and I’m open source!