OpenAI collapses media reality with Sora AI video generator | If trusting video from anonymous sources on social media was a bad idea before, it’s an even worse idea now::Hello, cultural singularity—soon, every video you see online could be completely fake.
I looked at these videos with very mixed emotions. On the one hand, I marveled at how far we’ve gotten. In a few years we went from generating sort of okay images in a very confined domain and essentially uncontrollable, to generating high resolution video that on first glance looks real.
But then the sadness struck me. I think we’re entering the post-truth era, where the truth is harder and harder to find because all the fake stuff looks so real. We can generate text, images, sound, and now also video of whatever we want in the blink of an eye. Combine this with the tendency of people to accept any “information” that fits their view, and the filter bubbles that already exist, and we can see that humanity will start living in separate bubbles. Every bubble will have their own truth, and even if someone proves that a video or image is fake, that information will probably not even reach them because the truth doesn’t generate enough clicks.
I want to stay optimistic, we’ve overcome so much stuff as a species, maybe we’ll right the ship at some point. But with all the shit that is already going on in the world, the last thing we need is the ability to fake videos like this in no time at all. At some point the separate filter bubbles will tear our stable western world as we knew it apart, and we’ll see shit like WW II again. The situation is already heating up.
It’s funny that in the human history there will be a gap of around 100 years where photos and video were considered to be solid proof and evidence that could determine the outcome of somebody’s future
we’re back at square one I guess
Naah, that was never a thing. EG: In 1917, two young girls created some photographs of fairies, the Cottingley Fairies. Arthur Conan Doyle, the inventor of Sherlock Holmes, endorsed them as real. When you have eliminated the impossible, whatever remains, however improbable, must be the truth. That quote is terrible advice.
The last 1 or 2 decades were really the golden age of credible evidence. Everyone has a video camera and can upload these videos almost immediately (proving that the videos were not edited later). Yet, at the same time, misinformation has become this huge topic.
We’re not back to square 1, either. You can still immediately upload a video (or a hash, or get it certified in some way). Say, you do this with dashcam footage after a collision, ASAP. That makes it almost unassailable as evidence, because you can’t have had time to forge it; certainly not in a way that is congruent with independent evidence and testimony.
If several people, upload videos of the same event at about the same time, then they either are all in it together and carefully prepared the videos beforehand, or not.
Damn, the AI that wrote this is really good!
Haha thanks
I’m actually glad that AI is making people realize that what they see is likely not real. For the history of media, the default has been for the written word or images or video to be taken as 100% truth, when in reality, it has always been very easy to deceive and manipulate. Now that we will suspect everything, maybe there will finally be critical thinking.
Honestly I think we’ve been there for a while. The only difference now is that it’s very easy for anyone to fake something, which might actually force us to face it? Or not who knows.
Sounds like what the board of openai thought when they attempted to fire him
This kind of AI stuff bums me out. You get people legitimately sharing AI images (and potentially videos in the future) and saying “look what I made!”. It’s totally inauthentic.
My boss loves this shit, on the other hand. Looking forward to the day she can automate our jobs away, I assume.
Good luck when an ai can lay bricks and mortar.
Why are they working so hard on making humanity worse?
I really have to say it?
Because the rich profit the most when everyone else is in fear and confusion.
So they generate it as much as possible and reap the rewards of ignorance and knee-jerk policies.
Because we’re all born selfish assholes*, and some people never learn to not be so.
*We’re all born as selfish idiots, how can we be otherwise? We’re helpless at birth, thrust from perfect comfort and safety into discomfort, utterly ignorant and wholly dependent, with no knowledge there are others, who are just as dependent and helpless when they’re born. Learning about others, and how to get along is part of maturing.
Sorry no, your perceptions are skewed by how well society rewards selfish assholes.
Most humans are inherently empathic and compassionate, just the tiny handful of sociopaths that run everything are projecting.
Every day were getting closer to “The Running Man.”
More dancing women in tv???
FANTASTIC reference! This movie is so funny and awesome, and it seems to have completely disappeared from pop culture. I never understood why Conan looms so large in our collective memory, but this movie totally vanished.
I’d buy that for a monthly subscription!
OpenAI singlehandedly breaking the internet. Props, tbh.
They really are. I don’t know about Google but with DDG when searching for information I feel like most of the top results are articles written by AI. Luckily it’s still somewhat easy to recognize but that’s not going to be the case for long. It’s inevitable though so I don’t really blame them. If not OpenAI then it would have just been someone else. I’m just worried about where this is going. I can think of more ways this could go wrong than right.
Genuine question: why do we need this type of thing?
Especially in view of the harm it can cause, what’s the point of creating this aside from generating shareholder value?
Sure, creating a video out of text quickly is cool, but is there an actual need for this?
I mean, the ability to generate whatever video you want without having to pay the costs normally associated with filming, location, actors etc is going to be very appealing to people like advertisers. This way you can have a few seconds of a beach for your travel company advert, for example, without having to pay for the stock footage or film it yourself. In fact I can see this transforming stock footage in general. Why bother to pay someone to make a generic video of ‘people having a meeting’ when an AI can do it for free in half the time. Doesn’t even need to be that good if you’re only using it briefly in a presentation. Not saying any of this is a good thing, but here we are…
I get what you’re saying, but there isn’t really any NEED for that
Thats kind of like saying we never needed cars because we had horses.
I mean, yes in so far that it opens those options up to people who may not have been able to afford it before. Whether that’s a ‘need’ or not depends on your opinion of the company I guess.
Also there are other applications beyond this, of course. Easily made videos could help reduce the costs associated with treating some mental health issues for example.
The you have the ability for novice film makers to make content, a bit like how engines like Unity have made it easier for people to make games. Sure there’ll be shit-tier tat, but there’ll also be content made by creators that may never have had the chance otherwise.
It may help to make synthetic training data for other models / simulators
Which then just feeds to the system. But as this is a globally impactful thing, is there any real world need that outweighs the harm?
Cost savings are a need. It frees resources for everyone. Sure the vast majority of the profit goes to the shareholders but that’s true of every labor saving device.
Do we NEED computers? You can hire people to do calculations by hand. The word Computer used to mean a job title, not a device.
OpenAI’s take is someone will create this technology - it might as well be them since their motivation is relatively pure. OpenAI is a non profit and they do work hard to minimise the damage their tech can cause. Which is why this video generation feature has not been launched yet.
OpenAI is no longer “pure.” They are not open. They do not publish the details of any of the discoveries they’ve made (which used to be standard practice, even in the private sector). Their leadership is now in the “effective accelerationism” camp that worships capitalism, and sees developing AGI as their moral obligation, regardless of what harm it may cause to society. (They are also delusional, because it’s very unlikely AGI will be developed anytime soon).
OpenAI is only technically non-profit. They’re a proxy for Microsoft in all but name. They started out mostly pure, but their dickhead CEO has worked hard to undo all of that nonsense and has created parallel companies for OpenAI that can absolutely make profit while the main company gets to keep its nonprofit status. That was literally the entire basis for the board firing him (the CEO) a few months back.
This way you can have a few seconds of a beach for your travel company advert, for example, without having to pay for the stock footage or film it yourself.
Advertising holidays at places that do not exist! Exactly what we needed!
So no different than currently? Pictures and videos in ads all have heavy editing and post processing to make them look better.
It‘s highly debatable how much editing is ok or when it get‘s deceiving. Video can only show you so much of an actual experience, senses like smells, warmth and time need some editing to be somewhat communicated. That‘s just film making.
For an AI generated clip it‘s not even a debate. It doesn‘t exist and doesn‘t try to advertise anything. And since it‘s so much easier and cheaper than doing all of the above… Well let‘s just say the incentive to deceive was just pushed up to 11.
Uhh, ads already do that.
Usually their computer animations are a pretty obvious style choice. This is completely different than that and even more different than photos.
OK, poor example, but you get my point.
Art. Presentations. Visualizations. Porn.
Weird music videos. Just this week Youtube had pushed me music videos all done in the weird warpy AI style.
It’s kind of cool and simultaneously already feels like a fad that tired with.
Who said anything about need?
It was created because someone thought of it. How it’s used is a measure of the person using it.
People will find ways to utilize whatever someone creates. And usually in ways the creator never envisioned.
“Needing” something comes after a tool becomes ubiquitous. Imagine trying to screw in a Phillips screw with a slot screw driver - you’d need a Phillips driver because those screws are now ubiquitous (and I can’t wait for them to go away. All hail our stripped screw saviour Torx!)
Genuine question: When have we ever needed the things corporations push as disruptive technologies?
This isn’t being done because we asked for it, it’s being done by the owner class as a method of control and cover.
It isn’t ‘we’, and never has been.
Isn’t this a bit over dramatic, seeing as we had deepfake tech for a while now?
I don’t think so, as deep fake stuff was about switching faces and voices. You needed actual footage to train this on.
So if you wanted to stage something, it would take considerable effort, money, time, and manpower.
Now anyone will be able to just type in a prompt and have a video generated.
We saw the Joe Biden deepfake that made calls to tell people not to vote for him. Now just wait until we are having videos of him saying it sent out in mass.
Imagine generating 5,000 videos of different people (likenesses pulled from Facebook) reacting to a fake calamity staged in a certain city.
Imagine seeing it every day to a point you can’t see the real calamity coming because you stopped believing in them entirely.
Oh man reality is going to be so strange to you when you get old enough to understand it.
https://aftermath.site/openai-sora-scam-sillicon-valley
It was a scam peoples
Targeted stupids investors were not supposed to share it
The link does not say in what way people were not supposed to share.
The link is the same kind of self-delusion people show around all of these generative tools: “look the faces are weird, the bird has wrong feathers, the cat has only 2 legs, nothing to worry about” while forgetting that most everything else in a clip works well and that it is the first-of-the-first releases which will get gradually better.
This shit is sold as the new Pixar
Yeah in a few months it will be a tool to help animators with shadows and lightings In some years it could produce a decent GIF of Pepe having sex with Sonic
At least it is less a scam than an NFT, still a scam
Still a waste of money and energetic ressources
Invest in it if you can, sure you can have a nice taxe refund in 3-5 years
Well I get that AI companies over promise and stuff, but that opinion piece really just confirmes what we’re already able to see in said clips. Sure, many animals look eerie es hell and that monobloc excavation video is one hell of an acid trip, but there’s already a lot there. More than I’m comfortable with.
I don’t really see the big problem yet. There’s still a hint of uncanny valley in that video.
This is only the beginning. It’s only going to get harder and harder to know what is and isn’t real online.
Sure, you and I are aware of this and have an idea of what to look out for. But do my older parents or grandparents know about this stuff and what to look for? I seriously doubt it.With the amount of people who are either lying or genuinely can’t tell when images are made by AI… I’m scared
Show it to your parents and ask what they think. Guaranteed they can’t tell it’s fake.
My parents are dead.
My mother is about the only person of that vintage that would probably say “this doesn’t look right”
Even my older siblings (boomers) would probably fall for it.
Lets revisit this comment in 3 years.
One year. The will smith spaghetti video came out last year. It’s progressing at an impossible pace already.
9 months