LLMs abstract information collected from the content through an algorithm (what they store is the result of a series of tests/analysis, not the content itself, but a set of characteristics/ideas). If that makes it derivative, then all abstractions are derivative. It’s not possible to make abstractions without collecting data derived from a source you are observing.
If derivative abstractions were already something that copyright can protect then litigants wouldn’t resort to patents, etc.
LLMs don’t have anything to do with abstract ideas, they quite literally produce derivative content based on their training data & prompt.
LLMs abstract information collected from the content through an algorithm (what they store is the result of a series of tests/analysis, not the content itself, but a set of characteristics/ideas). If that makes it derivative, then all abstractions are derivative. It’s not possible to make abstractions without collecting data derived from a source you are observing.
If derivative abstractions were already something that copyright can protect then litigants wouldn’t resort to patents, etc.