As much as people would like there to be, there really does not seem to be anything here. The original author doesn't think so, either (would need to refind the tweet).
I took the previous user's question as asking more like: Say there's an image in the training set with the description "Salvador Dali sitting at desk with pen and paper, 1957", and you put in the prompt "Salvador Dali sitting at desk with pen and paper, 1957", how close would the result be to that original training image?
It doesn’t have to be the same to achieve low loss. The date is not relevant, for example, for the image content. It can memorize a few images, but you can’t “compress” all of the internet’s images into text (if you can, make a startup with the most efficient compression algorithm ever made).
As much as people would like there to be, there really does not seem to be anything here. The original author doesn't think so, either (would need to refind the tweet).