top | item 41475797

(no title)

csjh | 1 year ago

I think the most surprising part here is the gzipped-base64'd-compressed data almost entirely removes the base64 overhead.

discuss

order

zamadatix|1 year ago

It feels somewhat intuitive since (as the article notes) the Huffman encoding stage effectively "reverses" the original base64 overhead issue that an 8 bit (256 choices) index is used for 6 bits (64 choices) of actual characters. A useful compression algorithm which _didn't_ do this sort of thing would be very surprising as it would mean it doesn't notice simple patterns in the data but somehow compresses things anyways.

ranger_danger|1 year ago

how does it affect error correction though?