Comment on Recreating uncensored Epstein PDFs from raw encoded attachments
hodgepodgin@lemmy.zip 23 hours ago
I tried to leave a comment, but it doesn’t seem to be showing up there.
I’ll just leave it here:
too tired to look into this, one suggestion though - since the hangup seems to be comparing an L and a 1, maybe you need to get into per-pixel measurements. This might be necessary if the effectiveness of ML or OCR models isn’t at least 99.5% for a document containing thousands of ambiguous L’s. Any inaccuracies from an ML or OCR model will leave you guessing 2^N candidates which becomes infeasible quickly. Maybe reverse engineering the font rendering by creating an exact replica of the source image? I trust some talented hacker will nail this in no time.
RIotingPacifist@lemmy.world 22 hours ago
How big is N though?
Qwaffle_waffle@sh.itjust.works 16 hours ago
64
hodgepodgin@lemmy.zip 6 hours ago
Since there’s 78 pages, I’m guessing at least 1 ambiguity per page? Anyways, it’s dreadfully big.
RIotingPacifist@lemmy.world 6 hours ago
2^78 is large but computers can do an awful lot per second, so if only about some the pages contain attachments 2^40-55 is something you could bruteforce in weeks if you can do millions of attempts a second
mEEGal@lemmy.world 19 hours ago
Asking the real questions