I understand that people in this “field” regularly use pseudo-scientific language (I actually deleted that part of my comment).
But the terminology has never been suitable so it shouldn’t be used in the first place. It pre-supposes the hypothesis that they’re supposedly “disproving”. They’re feeding into the grift because that’s what the field is. That’s how they all get paid the big bucks.
Yep. I’m retired now, but before retirement a month or so ago, I was working on a project that relied on several hundred people back in 2020. “Why can’t AI do it?”
The people I worked with are continuing the research and putting it up against the human coders, but…there was definitely an element of “AI can do that, we won’t need people” next time. I sincerely hope management listens to reason. Our decisions would lead to potentially firing people, so I think we were able to push back on the “AI can make all of these decisions”…for now.
The AI people were all in, they were ready to build an interface that told the human what the AI would recommend for each item. Errrm, no, that’s not how an independent test works. We had to reel them back in.
Why would they “prove” something that’s completely obvious?
The burden of proof is on the grifters who have overwhelmingly been making false claims and distorting language for decades.
They’re just using the terminology that’s widespread in the field. In a sense, the paper’s purpose is to prove that this terminology is unsuitable.
I understand that people in this “field” regularly use pseudo-scientific language (I actually deleted that part of my comment).
But the terminology has never been suitable so it shouldn’t be used in the first place. It pre-supposes the hypothesis that they’re supposedly “disproving”. They’re feeding into the grift because that’s what the field is. That’s how they all get paid the big bucks.
That’s called science
Not when large swaths of people are being told to use it everyday. Upper management has bought in on it.
Yep. I’m retired now, but before retirement a month or so ago, I was working on a project that relied on several hundred people back in 2020. “Why can’t AI do it?”
The people I worked with are continuing the research and putting it up against the human coders, but…there was definitely an element of “AI can do that, we won’t need people” next time. I sincerely hope management listens to reason. Our decisions would lead to potentially firing people, so I think we were able to push back on the “AI can make all of these decisions”…for now.
The AI people were all in, they were ready to build an interface that told the human what the AI would recommend for each item. Errrm, no, that’s not how an independent test works. We had to reel them back in.