A nightmare scenario previously only imagined by AI researchers, where AI image generators accidentally spit out non-consensual pornography of real people, is now reality.
A nightmare scenario previously only imagined by AI researchers, where AI image generators accidentally spit out non-consensual pornography of real people, is now reality.
My complaint actually hinges on models not emitting an exact replica. That would be obvious infringement. In cases like DreamGF, they would be wide-open to lawsuits from very wealthy people whose primary asset is their right of publicity.
What these ML companies are doing is: They are identifying where the line of definite infringement lies, and aiming their business as close to that line as possible.
But not over that line. If you draw lines and then these companies stay within them, what’s the problem?
People don’t have unique appearances.
Well, setting aside that there’s no law against being a total asshole, so like… We don’t have to make a bad behavior illegal in order to complain about it…
There’s the letter of the law, and there’s the intent. We start with a shared cultural attitude of how we should treat each other, and then we turn that into a quantifiable, objective rule that we can enforce through law.
We can try to make the law match our cultural attitudes as closely as possible, but there will always be gaps.
Now, I’ve got my own beef with how our IP and publicity laws work, and I’d like them to be more permissive in many ways. Much of IP law is exploitative, takes advantage of creators more than it protects them, and has lagged way behind where our social norms are these days.
But these ML companies aren’t interested in abiding by any social norms at all. Only paying lip service to current laws, which were written in a time before these “AI” services were even a possibility – skating by on technicalities, like a little brother poking the air 2 inches from your face and taunting “I’m not touching you! I’m not touching you!”