/acropalypse_matching_sha256.py Secret
Last active
March 24, 2025 14:19
Note: This is not the latest version of the code! It's been improved since, but this version matches the hash in this tweet: https://twitter.com/David3141593/status/1610030175688736768
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import zlib | |
import sys | |
import io | |
if len(sys.argv) != 5: | |
print(f"USAGE: {sys.argv[0]} orig_width orig_height cropped.png reconstructed.png") | |
exit() | |
PNG_MAGIC = b"\x89PNG\r\n\x1a\n" | |
def parse_png_chunk(stream): | |
size = int.from_bytes(stream.read(4), "big") | |
ctype = stream.read(4) | |
body = stream.read(size) | |
csum = int.from_bytes(stream.read(4), "big") | |
assert(zlib.crc32(ctype + body) == csum) | |
return ctype, body | |
def pack_png_chunk(stream, name, body): | |
stream.write(len(body).to_bytes(4, "big")) | |
stream.write(name) | |
stream.write(body) | |
crc = zlib.crc32(body, zlib.crc32(name)) | |
stream.write(crc.to_bytes(4, "big")) | |
orig_width = int(sys.argv[1]) | |
orig_height = int(sys.argv[2]) | |
f_in = open(sys.argv[3], "rb") | |
magic = f_in.read(len(PNG_MAGIC)) | |
assert(magic == PNG_MAGIC) | |
# find end of cropped PNG | |
while True: | |
ctype, body = parse_png_chunk(f_in) | |
if ctype == b"IEND": | |
break | |
# grab the trailing data | |
trailer = f_in.read() | |
print(f"Found {len(trailer)} trailing bytes!") | |
# find the start of the nex idat chunk | |
try: | |
next_idat = trailer.index(b"IDAT", 12) | |
except ValueError: | |
print("No trailing IDATs found :(") | |
exit() | |
# skip first 12 bytes in case they were part of a chunk boundary | |
idat = trailer[12:next_idat-8] # last 8 bytes are crc32, next chunk len | |
stream = io.BytesIO(trailer[next_idat-4:]) | |
while True: | |
ctype, body = parse_png_chunk(stream) | |
if ctype == b"IDAT": | |
idat += body | |
elif ctype == b"IEND": | |
break | |
else: | |
raise Exception("Unexpected chunk type: " + repr(ctype)) | |
idat = idat[:-4] # slice off the adler32 | |
print(f"Extracted {len(idat)} bytes of idat!") | |
print("building bitstream...") | |
bitstream = [] | |
for byte in idat: | |
for bit in range(8): | |
bitstream.append((byte >> bit) & 1) | |
# add some padding so we don't lose any bits | |
for _ in range(7): | |
bitstream.append(0) | |
print("reconstructing bit-shifted bytestreams...") | |
byte_offsets = [] | |
for i in range(8): | |
shifted_bytestream = [] | |
for j in range(i, len(bitstream)-7, 8): | |
val = 0 | |
for k in range(8): | |
val |= bitstream[j+k] << k | |
shifted_bytestream.append(val) | |
byte_offsets.append(bytes(shifted_bytestream)) | |
# bit wrangling sanity checks | |
assert(byte_offsets[0] == idat) | |
assert(byte_offsets[1] != idat) | |
print("Scanning for viable parses...") | |
# prefix the stream with 32k of "X" so backrefs can work | |
prefix = b"\x00" + (0x8000).to_bytes(2, "little") + (0x8000 ^ 0xffff).to_bytes(2, "little") + b"X" * 0x8000 | |
for i in range(len(idat)): | |
truncated = byte_offsets[i%8][i//8:] | |
# only bother looking if it's (maybe) the start of a non-final adaptive huffman coded block | |
if truncated[0]&7 != 0b100: | |
continue | |
d = zlib.decompressobj(wbits=-15) | |
try: | |
decompressed = d.decompress(prefix+truncated) + d.flush(zlib.Z_FINISH) | |
decompressed = decompressed[0x8000:] # remove leading padding | |
if d.eof and d.unused_data in [b"", b"\x00"]: # there might be a null byte if we added too many padding bits | |
print(f"Found viable parse at bit offset {i}!") | |
# XXX: maybe there could be false positives and we should keep looking? | |
break | |
else: | |
print(f"Parsed until the end of a zlib stream, but there was still {len(d.unused_data)} byte of remaining data. Skipping.") | |
except zlib.error as e: # this will happen almost every time | |
#print(e) | |
pass | |
else: | |
print("Failed to find viable parse :(") | |
exit() | |
print("Generating output PNG...") | |
out = open(sys.argv[4], "wb") | |
out.write(PNG_MAGIC) | |
ihdr = b"" | |
ihdr += orig_width.to_bytes(4, "big") | |
ihdr += orig_height.to_bytes(4, "big") | |
ihdr += (8).to_bytes(1, "big") # bitdepth | |
ihdr += (2).to_bytes(1, "big") # true colour | |
ihdr += (0).to_bytes(1, "big") # compression method | |
ihdr += (0).to_bytes(1, "big") # filter method | |
ihdr += (0).to_bytes(1, "big") # interlace method | |
pack_png_chunk(out, b"IHDR", ihdr) | |
# fill missing data with solid magenta | |
reconstructed_idat = bytearray((b"\x00" + b"\xff\x00\xff" * orig_width) * orig_height) | |
# paste in the data we decompressed | |
reconstructed_idat[-len(decompressed):] = decompressed | |
# one last thing: any bytes defining filter mode may | |
# have been replaced with a backref to our "X" padding | |
# we should fine those and replace them with a valid filter mode (0) | |
print("Fixing filters...") | |
for i in range(0, len(reconstructed_idat), orig_width*3+1): | |
if reconstructed_idat[i] == ord("X"): | |
#print(f"Fixup'd filter byte at idat byte offset {i}") | |
reconstructed_idat[i] = 0 | |
pack_png_chunk(out, b"IDAT", zlib.compress(reconstructed_idat)) | |
pack_png_chunk(out, b"IEND", b"") | |
print("Done!") |
Excellent work on this @DavidBuchanan314.
FYI: when I tried using the above script, I got this error when trying to open the resulting PNG:
Fatal error reading PNG image file: bad adaptive filter value
https://github.com/frankthetank-music/Acropalypse-Multi-Tool is a repo based on your original code (they do attribute your work in the README). I've submitted this pull to provide a standalone CLI version: frankthetank-music/Acropalypse-Multi-Tool#2
Thought I'd share in case someone else hits it:)
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Thanks for your fast response David! 😄