this post was submitted on 16 Sep 2023
65 points (88.2% liked)
Games
16803 readers
496 users here now
Video game news oriented community. No NanoUFO is not a bot :)
Posts.
- News oriented content (general reviews, previews or retrospectives allowed).
- Broad discussion posts (preferably not only about a specific game).
- No humor/memes etc..
- No affiliate links
- No advertising.
- No clickbait, editorialized, sensational titles. State the game in question in the title. No all caps.
- No self promotion.
- No duplicate posts, newer post will be deleted unless there is more discussion in one of the posts.
- No politics.
Comments.
- No personal attacks.
- Obey instance rules.
- No low effort comments(one or two words, emoji etc..)
- Please use spoiler tags for spoilers.
My goal is just to have a community where people can go and see what new game news is out for the day and comment on it.
Other communities:
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Artists who look at art are processing it in a relatable, human way. An AI doesnt look at art. A human tells the AI to find art and plug it in, knowing that work is copyrighted and not available for someone else's commercial project to develop an AI.
That's not how AI art works. You can't tell it to find art and plug it in. It doesn't have the capability to store or copy existing artworks. It only contains the matrix of vectors which contain concepts. Concepts cannot be copyrighted.
Kind of. The AI doesn't go out and find/do anything, people include images in its training data though. So it's the human that's finding the art and plugging it in — most likely through automated processes that just scrape massive amounts of images and add them to the corpus used for training.
Sorry, this is wrong. You definitely can train AI to produce works that are very nearly a direct copy. How "original" works created by the AI are is going to depend on the size of the corpus it got trained on. If you train the AI (or put a lot of weight on) training for just a couple works from one specific artist or something like that it's going to output stuff that's very similar. If you train the AI on 1,000,000 images from all different artists, the output isn't really going to resemble any specific artist's style or work.
That's why the company emphasized they weren't training the AI to replicate a specific artist's (or design company, etc) works.
As a general statement: No, I am not. You're making an over specific scenario to make it true. Sure, if I take 1 image and train a model just on that one image, it'll make that exact same image. But that's no different than me just pressing copy and paste on a single image file. The latter does the job whole lot better too. This entire counter argument is nothing more than being pedantic.
Furthermore, if I'm making such specific instructions to the AI, then I am the one who's replicating the art. It doesn't matter if I use a pencil to trace out the existing art, using photoshop, or creating a specific AI model. I am the one who's doing that.
You didn't qualify what you said originally. It either has the capability or not: you said it didn't, it actually does.
Not really. It isn't that far-fetched that a company would see an artist they'd like to use but also not want to pay that artist's fees so they train an AI on the artist's portfolio and can churn out very similar artwork. Training it on one or two images is obviously contrived, but a situation like what I just mentioned is very plausible.
So this isn't true. What you said isn't accurate with the literal interpretation and it doesn't work with the more general interpretation either. The person higher in the thread called it stealing: in that case it wasn't, but AI models do have the capability to do what most people would probably call "stealing" or infringing on the artist's rights. I think recognizing that distinction is important.
Yes, that's kind of the point. A lot of people (me included) would be comfortable calling doing that sort of thing stealing or plagiarism. That's why the company in OP took pains to say they weren't doing that.
Yeah, sure. But there's nothing that says "it's not stealing if you do it in a relatable, human way". Stealing doesn't have anything to do with that.
And it is available for someone else's commercial project to develop a human artist? Basically, the "an AI" part is still irrelevant to. If the works are out there where it's possible to view them, then it's possible for both humans and AIs to acquire them and use them for training. I don't think "theft" is a good argument against it.
But there are probably others. I can think of a few.
I just want fucking humans paid for their work, why do you tech nerds have to innovate new ways to lick the boots of capital every few years? Let the capitalists make aeguments why AI should own all of our work, for free, rights be damned, and then profit off of it, and sell that back to us as a product. Let them do that. They don't need your help.
That's a problem whether or not we're talking about AI.
That's really not how it works. "Tech nerds" aren't licking the boots of capitalists, capitalists just try to exploit any tech for maximum advantage. What are the tech nerds supposed to do, just stop all scientific and technological progress?
AI doesn't "own your work" any more than a human artist who learned from it does. You don't like the end result, but you also don't seem to know how to come up with a coherent argument against the process of getting there. Like I mentioned, there are better arguments against it than "it's stealing", "it's violating our rights" because those have some serious issues.