i’d guess they could hyper optimise for “perceived difference” rather than data loss specifically… they do a pretty good job of generating something from nothing, so i’d say with enough data they’d probably generate a pretty reasonable facsimile of “standard” stuff
There have been a lot of studies done (and published) on what humans can and can't perceive. I wouldn't have much trouble believing that the LLM has access to them and can pattern match on the variables involved.