Discussion about this post

User's avatar
Lois Thomson Bowersock's avatar

Wow Wyrd! The world needs people like you because there are people like me. Bentley looks like a true K9 angel! Keep up the good work and I'll do likewise. Happy Fourth to you and your family.

Expand full comment
Joseph Rahi's avatar

This is fantastic! I really like the idea of doing it from "blanked" pixels/characters. That was kind of part of my initial thought process, thinking about seeing complexity as how much the picture "makes sense" (is compressible) as we uncover the full picture. It's about how "integral" it is.

>On the left, the file is just zeros or spaces (or any single character). On the right, each byte is entirely random. Neither extreme has any meaning.

I'm not sure this is correct. The non-randomised text should be compressible to something which would appear much closer to randomness, but would still contain the same meaning (once decompressed anyway), so I think the degree of randomness quite possibly has no connection at all to how much meaning there might be within it. I'm not sure, but I think under an ideal compression system, all messages would appear as if they were random. What do you think?

Something interesting I noticed with the pictures is that it's much easier to see the image as it emerges from randomness than it is with it emerging from all black. Which you might not expect, since both have the same % of the original image visible. If anything, I'd have guessed that emerging from all black would be easier, since your brain doesn't have to pick out signal from random noise. Perhaps the randomness gives the brain's pattern recognition system more to latch onto?

You might be interested to hear that I've since experimented a little with LLM-based text compression, which integrates some degree of the semantic structure into the compression system. Hypothetically, if the LLM were a perfect model of the language, it should be very close to perfect compression for real-world, meaningful texts. Weirdly, the final steps of removing the random noise actually make much less difference to its compressibility for the LLM-based compression, whereas with the normal compression method the last steps make the biggest difference. I suppose the LLM is picking up the error-correcting structures within language? I should probably try to write something up about it.

Expand full comment
4 more comments...

No posts