top | item 39771245

(no title)

five_lights | 1 year ago

>The entire AI industry is powered by piracy at a massive scale.

ARRRRR..

This is a grey area still for me. It's a neural network. It works similar to our brains work, but more consistent. It's doesn't seem like piracy to me. If an artist was really into Salvidor Dali, and happened to imitate his surrealist style, it would not be considered piracy. In fact, this is how art has evolved over the centuries. Each relevant artist in the past has incrementally contributed to what we call art today.

I feel like the people unwilling to accept that AI may impact their career are more worried about putting food on the table than anything else, which is very understandable, but it's just the cost of progress.

The bigger problem we need to deal with is how to retrain and provide job placement who are affected by disruptive technologies. We've really failed the public on this in the past and I don't think it's worth nerfing emerging tech just to keep people employed. This is not the first or last time this has happened, and it's going to be more frequent as technology advances.

discuss

order

__loam|1 year ago

> It's a neural network. It works similar to our brains work, but more consistent.

Irrelevant and incorrect.

> It's doesn't seem like piracy to me.

It's pretty indisputably piracy, whether or not it's legal/fair use/whatever. Many of the training sets included material like the books3 corpus which was downloaded to a server somewhere. That is simply piracy, doesn't matter why they downloaded it.

I believe many artists rightly refuse to accept this threat to their livelihoods because it was built on their labor. It's so fucking rich to see people patronizingly suggest that this is just an economic problem and those artists better just figure out a new profession.

You built a commercial product on unlicensed data. Do you actually think the law is going to agree that that's fair use?

ben_w|1 year ago

> It's pretty indisputably piracy, whether or not it's legal/fair use/whatever.

Ah, this is obviously some strange usage of the word 'indisputably' that I wasn't previously aware of.

> I believe many artists rightly refuse to accept this threat to their livelihoods because it was built on their labor.

This model is trained from scratch using only public domain/CC0 and copyright images with specific permission for use: https://huggingface.co/Mitsua/mitsua-diffusion-one

Does it change anything?

If all the other models were deleted, and this was the only one left, and all future models also had to be similarly licensed, would it change even one single point?

Even if it was the only remaining model and this kind of licensing a requirement for all future work, artists would still be automated out of their highly skilled yet poorly paid profession. It still sucks. There's still no nice way to convey that.

> You built a commercial product on unlicensed data. Do you actually think the law is going to agree that that's fair use?

What do you think the Google search engine is, if not a commercial product built on unlicensed data?

The courts go both ways on this specific question with Google depending on the exact details, because nothing in law is as easy or simple as the clear-cut, goodies-vs.-baddies, black-and-white morality play you want this to be.

The fact that Stability AI have not yet been sued out of existence in a simple open-and-shut court case about copyright infringement ought to have demonstrated both this point, and also that the question "is this piracy?" is, in fact, disputable.