knotthatone

joined 1 year ago
[–] [email protected] 2 points 2 months ago

This toggle allows you to opt out of having profiling used for future decisions that produce legal or similarly significant effects about you.

The what now?

This sounds strangely ominous.

[–] [email protected] 46 points 10 months ago (4 children)

I only flip if the bottom bun is losing structural integrity.

[–] [email protected] 3 points 10 months ago (1 children)

My bet is: it’s going to depend on a case by case basis.

Almost certainly. Getty images has several exhibits in its suit against Stable Diffusion showing the Getty watermark popping up in its output as well as several images that are substantially the same as their sources. Other generative models don't produce anything all that similar to the source material, so we're probably going to wind up with lots of completely different and likely contradictory rulings on the matter before this gets anywhere near being sorted out legally.

Copyright laws are not necessarily wrong; just remove the “until author’s death plus 70 years” coverage, go back to a more reasonable “4 years since publication”, and they make much more sense.

The trouble with that line of thinking is that the laws are under no obligation to make sense. And the people who write and litigate those laws benefit from making them as complicated and irrational as they can get away with.

[–] [email protected] 3 points 10 months ago* (last edited 10 months ago)

Not a single original sentence of the original work is retained in the model.

Which is why I find it interesting that none of the court cases (as far as I'm aware) are challenging whether an LLM is copying anything in the first place. Granted, that's the plaintiff's job to prove, but there's no need to raise a fair use defense at all if no copying occurred.

[–] [email protected] 13 points 10 months ago (4 children)

Clearly transformative only applies to the work a human has put in to the process. It isn't at all clear that an LLM would pass muster for a fair use defense, but there are court cases in progress that may try to answer that question. Ultimately, I think what it's going to come down to is whether the training process itself and the human effort involved in training the model on copyrighted data is considered transformative enough to be fair use, or doesn't constitute copying at all. As far as I know, none of the big cases are trying the "not a copy" defense, so we'll have to see how this all plays out.

In any event, copyright laws are horrifically behind the times and it's going to take new legislation sooner or later.

[–] [email protected] 1 points 10 months ago (1 children)

This thread is about ChatGPT, an LLM. It is not a general purpose AI.

[–] [email protected] 3 points 10 months ago (3 children)

So if someone builds an atom-perfect artificial brain from scratch, sticks it in a body, and shows it around the world, should we expect the creator to pay licensing fees to the owners of everything it looks at?

That's unrelated to an LLM. An LLM is not a synthetic human brain. It's a computer program and sets of statistical data points from large amounts of training data to generate outputs from prompts.

If we get real general-purpose AI some day in the future, then we'll need to answer those sorts of questions. But that's not what we have today.