787
OpenAI now tries to hide that ChatGPT was trained on copyrighted books, including J.K. Rowling's Harry Potter series
(www.businessinsider.com)
This is a most excellent place for technology news and articles.
You joke but AI advocates seem to forget that people have fundamentally different rights than tools and objects. A photocopier doesn't get the right to "memorize" and "learn" from a text that a human being does. As much as people may argue that AIs work different, AIs are still not people.
And if they ever become people, the situation will be much more complicated than whether they can imitate some writer. But we aren't there yet, even their advocates just uses them as tools.
You should read this article by Kit Walsh, who’s a senior staff attorney at the EFF too. The EFF is a digital rights group who most recently won a historic case: border guards now need a warrant to search your phone.
But this falls exactly under what I just said. To say that using Machine Learning to imitate an artist without permission is fine, because humans are allowed to learn to each other, is making the mistake of assigning personhood to the system, that it ought to have the same rights that human beings do. There is a distinction between the rights of humans as opposed to tools, so to say that an AI can't be trained on someone's works to replicate their style doesn't need to apply to people.
Even if you support that reasoning, that still doesn't help the writers and artists whose job is threatened by AI models based on their work. That it isn't an exact reproduction doesn't change that it relied on using their works to begin with, and it doesn't change that it serves as a way to undercut them, providing a cheaper replacement for their work. Copyright law as it was, wasn't envisioned for a world where Machine Learning exists. It doesn't really solve the problem to say that technically it's not supposed to cover ideas and styles. The creators will be struggling just the same.
Either the law will need to emphasize the value of human autorship first, or we will need to go through drastic socioeconomic changes to ensure that these creators will be able to keep creating despite losing market to AI. Otherwise, to simply say that AI gets to do this and change nothing else, will cause enormous damage to all sort of creative careers and wider culture. Even AI will become more limited with less fresh new creators to learn elements from.
The system doesn't get personhood, it is your tool, and as said in the article:
It is your right, not the system's you're upholding.
There is a difference between "analyzing" and derivating. The authorship of AI-created works is also not the user's, it takes more than a prompt for that, and that seems to be the conclusion courts are leaning towards.
Still, even if that turns out to be technically correct, it still doesn't help the creators getting undercut who might be driven out of their careers by AI.
It was just ruled AI can't be authors or hold copyright. AI itself can’t be authors or hold a copyright, but humans using them can still be copyright holders of any qualifying works.
They do specify that the human's involvement needs to be more extensive than prompting for a certain image or text. The output itself is not copyrightable. If we are speaking about the process of "analysis" that the ML model does, then the user does not get the rights over it.
This discussion is becoming increasingly overly specific and getting away from my point. My sole concern in all this is what happens to the artists who'll have to compete with AI?
It says :
And you do get rights to your own original analysis of data. That isn't even in question.
I guess all I have to say here is that generative models are a free and open source tool anyone can use. It took us 100,000 years to get from cave drawings to Leonard Da Vinci. This is just another step, like Camera Obscura.
When you call the output itself "analysis", that's not what they say.
This is in your own link. Simply prompting Midjourney doesn't get the user copyright.
That is not something many of those people whose work is being used to enable it even want to use. Not to mention, if AI art were to be the "next evolution" in media, which it isn't since it output the same medium, there wouldn't be a need for as many AI prompters as there are artists right now. This glosses over the issue entirely.
There's more to generative models than just prompting. In that specific case, the images were generated with just a prompt because Midjourney doesn't have the tools to let you do anything else.
It might not be for everyone, but there are already plenty of artists leveraging these new techniques.
That might be the case, but people often say that most art is only appeals to a few people, so just like Source Film Maker allowed more animations to be made, I expect the same kind of widening of scope for projects solo artists can make. You can already see this happening already. Not everyone has the time or motivation to do something as simple as make their own sandwich, I don't think they're going to want to sit down and hammer out a picture or whatever themselves.
We are talking about something that is threatening artists who are already hammering it out themselves. What then?
I don't understand. Can you expand on this?
Here is an alternative Piped link(s): https://piped.video/watch?v=HtbEuERXSqk
https://piped.video/watch?v=tWZOEFvczzA
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I'm open-source, check me out at GitHub.
How do you see that as a difference? Tools are extensions of ourselves.
Restricting the use of LLMs is only restricting people.
When we get to the realm of automation and AI, calling tools just an "extension of ourselves" doesn't make sense.
Especially not when the people being "extended" by Machine Learning models did not want to be "extended" to begin with.