Doubled down on the “yea were not gonna credit artist’s our AI stole from”. What a supreme douche
I don’t think it’s as simple as all that. Artists look at other artists’ work when they’re learning, for ideas, for methods of doing stuff, etc. Good artists probably have looked at a ton of other artwork, they don’t just form their skills in a vacuum. Do they need to credit all the artists they “stole from”?
In the article, the company made a point about not using AI models specifically trained on a smaller set of works (or some artist’s individual works). Doing something like that would be a lot easier to argue that it’s stealing: but the same would be true if a human artist carefully studied another person’s work and tried to emulate their style/ideas. I think there’s a difference between that an “learning” (or learning) for a large body of work and not emulating any specific artist, company, individual works, etc.
Obviously it’s something that needs to be handled fairly carefully, but that can be true with human artists too.
Artists who look at art are processing it in a relatable, human way. An AI doesnt look at art. A human tells the AI to find art and plug it in, knowing that work is copyrighted and not available for someone else’s commercial project to develop an AI.
Artists who look at art are processing it in a relatable, human way.
Yeah, sure. But there’s nothing that says “it’s not stealing if you do it in a relatable, human way”. Stealing doesn’t have anything to do with that.
knowing that work is copyrighted and not available for someone else’s commercial project to develop an AI.
And it is available for someone else’s commercial project to develop a human artist? Basically, the “an AI” part is still irrelevant to. If the works are out there where it’s possible to view them, then it’s possible for both humans and AIs to acquire them and use them for training. I don’t think “theft” is a good argument against it.
But there are probably others. I can think of a few.
That’s not how AI art works. You can’t tell it to find art and plug it in. It doesn’t have the capability to store or copy existing artworks. It only contains the matrix of vectors which contain concepts. Concepts cannot be copyrighted.
I wouldn’t call myself a “good artist” at all, and I’ve never released anything, I just make music for myself. Most of the music I make starts with my shamelessly lifting a melody, chord progression, rhythm, sound, or something else, from some song I’ve heard. Then I’ll modify it slightly, add my own elements elsewhere, modify the thing I “stole” again, etc, and by the time I’ve finished, you probably wouldn’t even be able to tell where I “stole” from because I’ve iterated on it so much.
AI models are exactly the same. And, personally, I’m pretty good at separating the creative process from the end result when it comes to consuming/appreciating art. There are songs, paintings, films, etc, where the creative process is fascinating to me but I don’t enjoy the art itself. There are pieces of art made by sex offenders, criminals and generally terrible people - people who I refuse to support financially in any way - but that doesn’t mean my appreciation for the art is lessened. I’ll lose respect for an artist as a person if I find out their work is ghostwritten, but I won’t lose my appreciation for the work. So if AI can create art I find evocative, I’ll appreciate that, too.
But ultimately, I don’t expect to see much art created solely by AI that I enjoy. AI is a fantastic tool, and it can lead to some amazing results when someone gives it the right prompts and edits/curates its output in the right way. And it can be used for inspiration, and to create a foundation that artists can jump off, much like I do with my “stealing” when I’m writing music. But if someone gives an AI a simple prompt, they tend to get a fairly derivative result - one that’ll feel especially derivative as we see “raw output” from AIs more often and become more accustomed to their artistic voice. I’m not concerned at all about people telling an AI to “write me a song about love” replacing the complex prog musicians I enjoy, and I’m not worried about crappy AI-generated games replacing the lovingly crafted experiences I enjoy either.
I swear I’m old enough to remember this exact same fucking debate when digital tools started becoming popular.
It is, simply put, a new tool.
It’s also not the one and done magic button people who’ve never used shit think it is.
The knee-jerk reaction of hating on every art made with AI, is dangerous.
You’re free to like it or not, but it’s already out of the hat.
Big companies will have the ressources to train their own model.
I for one would rather have it in the public domain rather than only available to big corps.
How would they credit the artists? Generative AI is trained on thousands and millions of images and data points from equally numerous artists. He might as well say, “I give credit to humanity.”
Generative AI is trained on thousands and millions of images and data points from equally numerous artists.
Congrats on pinpointing the problem.
I didn’t see the sub at first and thought it was a kickstarter for a real-life mars terraformation project
A small team of 7 was able to create something of this magnitude , all thanks to the various tools of today like Generative AI.
We talk about the bad stuff of AI. But here’s the good… small mom and pop shops being able to release top tier products like the big companies.
If AI art is stolen data, then every artists on earth are thieves too.
Do you think artists just spontaneously conjure up art? No. Through their entire life of looking at other people’s works, they learned how to do stuff, they emulate and they improve. That’s how human artists come to be. Do you think artists go around asking permission from millions of past artists if they can learn from their art? Do artists track down whoever made the fediverse logo if I want to make a similar shaped art with it? Hell no. Consent in general is impossible too because whole lot of them are likely too dead to give consent be honest. Its the exact same way AI is made.
Your argument holds no consistent logic.
Furthermore, you likely have a misunderstanding of how AI is trained and works. AI models do not store nor copy art that it’s trained on. It studies shapes, concepts, styles, etc. It puts these concepts into matrix of vectors. Billions of images and words are turned into mere 2 gigabytes in something like SD fp16. 2GB is virtually nothing. There’s no compression capable of anywhere near that. So unless you actually took very few images and made a 2GB model, it has no capability to store or copy another person’s art. It has no knowledge of any existing copyrighted work anymore. It only knows the concepts and these concepts like a circle, square, etc. are not copyrightable.
If you think I’m just being pro-AI for the sake of it. Well, it doesn’t matter. Because copyright offices all over the world have started releasing their views on AI art. And it’s unanimously in agreement that it’s not stolen. Furthermore, resulting AI artworks can be copyrighted (lot more complexity there, but that’s for another day).
L take, AI is not a person and doesn’t have the right to learn like a person. It is a tool and it can be used to replicate others art.
It’s arguably not good that we’re normalizing people being able to use this while its training relied on other creators who were not compensated.
Oonga boonga wants his royalty checks for having first drawn a circle 25,000 years ago.
Then we shouldn’t have artists because they looked at other art without paying.
Devil’s advocate. It means that only large companies will have AI, as they would be the only ones capable of paying such a large number of people. AI is going to come anyway except now the playing field is even more unfair since you’ve removed the ability for an individual to use the technology.
Instituting these laws would just be the equivalent of companies pulling the ladder up behind them after taking the average artist’s work to use as training data.
How would you even go about determining what percentage belongs to the AI vs the training data? You could argue all of the royalties should go to the creators of the training data, meaning no one could afford to do it.
My programming training relied on other creators who were not compensated.
Were they in public forums and sites like stack overflow and GitHub where they wanted people to use and share their code?
I imagine creators who… released their work for free, and/or open source?
Humans using past work to improve, iterate, and further contribute themselves is not the same as a program throwing any and all art into the machine learning blender to regurgitate “art” whenever its button is pushed. Not only does it not add anything to the progress of art, it erases the identity of the past it consumed, all for the blind pursuit of profit.
Awesome, I didn’t know they had a kickstarter going. No such thing as bad press I guess.
Good interview. They didn’t let them off the hook, but weren’t pushing an agenda either.
This is going to be a moving target that someone is going to pay big bucks to figure out in court. International laws are not up to speed on what is or isn’t ok here, and the ethical discussion is interesting to watch unfold.