this post was submitted on 25 Aug 2023
85 points (100.0% liked)
Technology
37360 readers
230 users here now
Rumors, happenings, and innovations in the technology sphere. If it's technological news or discussion of technology, it probably belongs here.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Only very rarely, under extreme cases of overfitting. Overfitting is a failure state that LLM trainers want to avoid anyway, for reasons unrelated to copyright.
There simply isn't enough space in a LLM's neural network to be storing actual copies of the training data. It's impossible, from a data compression perspective, to fit it in there.