Hacker News new | past | comments | ask | show | jobs | submit login
PhotoDNA (wikipedia.org)
38 points by sandwichsphinx 9 hours ago | hide | past | favorite | 26 comments





https://anishathalye.com/inverting-photodna/

You could almost certainly produce nearly photo realistic PhotoDNA inversions with a finetuned diffusion model now. Is it possible to create a perceptual hashing algorithm where this isn't possible?


You can certainly not produce inversions. The data that is left in the hash is not enough to produce anything vaguely photorealistic.

However, you can fill the gaps and generate photorealistic photos that fit to the extremely reduced information you get from the hash. You are generating believable (as defined by the training data) photos that fit the hash.

That’s a huge difference.

Statements like yours are extremely dangerous. Without proper understanding of what GenAI can and can not do, people start relying on things that are not there.

Imagine your photorealistic inversion AI putting a mole or a wrinkle in the face of somebody without any foundation in the actual hash. Just because it fits better to the trained data. Explain that to the judge, when the person with just the right facial features sits in front of them.


This modern narrative of people posting their opinions or assumptions somewhere being "dangerous" because someone could just believe it is much more dangerous because it can be applied to any opinion anywhere that was ever published.

No judge will ever rule on something based on a comment they read in the Internet.


Secretive, unaccountable, uncontrollably expanding, driven by shady “independent” NGO which is in fact completely in bed with certain branches of government. Russian DPI censorship system is really like that... Oh, I'm sorry, we're discussing something that is a decade older, and belongs to a “free world”, which is a completely different thing.

These things are simply selling their services to the highest bidder. It's a business model on a power connections market. They are made to be offered to, and controlled by, entities that enjoy having such tools. Sometimes they are also offered to smaller fish, like media corporations, to hurt competitors (pirates and foreign services). Also, social media corporations can proudly state that they themselves “censor nothing”, because it's outsourced.

There's a great portrayal of people who run such services: https://www.newyorker.com/magazine/2019/11/04/a-cybersecurit...


What's novel about this? iirc, Apple withdrew its plan to hash photos client-side a couple years ago after an outcry. Dropbox has been hashing every file forever to save storage space. Store your shit with a cloud provider, expect it to get scanned, right?

Also, there are a million cute methods to make two different photos produce the same hash; that was actually what the outcry about Apple's version was about. The more the hash algorithm tried to produce the same hash for different variants of a photo, the more likely it was that someone could get their hands on a flagged hash and theoretically send you an innocuous looking photo that registered as CSAM. Pretty sure that's why Apple pulled it.


Technical details are actually irrelevant, people just like to show sophistication by discussing patterns on emperor's new clothes. It was, most likely, a second ad-hoc solution after first ad-hoc solution (most likely, having something to do with md5()) stopped working well enough.

What is important in this article is what isn't written there, and has to be deduced. How exactly “terrorist content” was included. What else has been discussed behind the closed doors. Who actually decides how the thing works.


> that was actually what the outcry about Apple's version was about. The more the hash algorithm tried to produce the same hash for different variants of a photo, the more likely it was that someone could get their hands on a flagged hash and theoretically send you an innocuous looking photo that registered as CSAM.

That was totally infeasible. There were two separate hashes, a public one and a private one, and there needed to be multiple false positives for the system to trigger. So not only would you need to generate collisions for two separate hashes simultaneously, including one for which there are no public details, you would need to do it for several images.

People made a lot of assumptions about how it would work without actually reading the papers Apple published on how it would work. So there’s this caricature of the system in people’s minds that is a lot simpler and easier to fool than the reality. That’s what Apple was forced to react to.


>That’s what Apple was forced to react to.

Perhaps apple could have made a better explanation available? Unless, there wasn't one?


> What's novel about this?

Literally in the first paragraph you can see that it’s not novel.

> PhotoDNA was developed by Microsoft Research and Hany Farid, professor at Dartmouth College, beginning in 2009


Also this isn't novel.ycombinator.com

> theoretically send you an innocuous looking photo that registered as CSAM. Pretty sure that's why Apple pulled it.

Apple also had human reviewers in the mix. The only reason they pulled it was PR/optics.


I would assert the only reason they pursued it in the first place is PR/optics, since the "optics" of not being able to proactively police what users do using E2EE services you provide is somewhat a problem. That said, I think the concept of having your own computer covertly report you to the authorities is a level too dystopian to accept even from Apple.

I agree the reason they pulled it was probably PR/optics. But given the problems with human reviews of apps on the app store, I wouldn't be confident that an underpaid employee somewhere wouldn't blindly agree with the algorithm.

Wouldn't that require exfiltrating the original photo? I remember them swearing that wasn't part of the deal.

Going from memory here but IIRC the deal was that on device they'd produce a hash using a known pHash, and if that was positive, they'd send the photo to check it against a second pHash that wasn't publicly-disclosed (to try to mitigate the problem of intentional collisions) and then if both of them were positive matches, they would have human reviewers in the loop.

It was a lot more advanced and abuse-resistant than people assumed. I really wish people had read how it worked instead of guessing it was something a lot simpler. There were two different perceptual hashes. If both matched, and the number of positive matches was high enough, a thumbnail would be able to be decrypted by Apple. Neither the device nor the server were able to independently check for a match, so the device wasn’t able to just scan all your files and flag the matches. It was tied into the iCloud upload process.

While this is understandable, the unfortunate issue was that Apple could be coerced into adding images certain authoritarian governments didn’t like to the list. Though imo it’s all moot if iCloud Photos aren’t end to end encrypted anyway.

“Coerced”? Check some recent news to see that for Apple rainbow-washing stops at the very moment they are held responsible for providing basic censorship circumvention tools.

I am amazed how people still cling to the hope that one day a corporation will do something nice for them without any hidden motive.


The fact that it is CSAM makes it an even harder problem to solve. With e.g. copyright infringement, you could keep some kind of records why a particular file is in the system, potentially even letting trusted and vetted organizations audit the files, but doing that for CSAM would be highly illegal and defeat the purpose of the system.

Personally I think they had a lot of pressure from naive external and internal sources and had to take the PR hit on this one.

If Apple were to do something like that with any level of real intent, the first thing you’d find out about it was when it went live.


Is there any good article/paper that describes how it actually works or is implemented not just in high-level and hand-waving terms?


Yeah Hacker Factor's multi-post critiques are where I first saw it analyzed. For reference they run the popular fotoforensics.com image analysis site.

They also have scathing critique (eg [1]) about the Adobe-led C2PA digital provenance signing, having themselves been part of various groups that seek solutions to the provenance problem.

[1] https://www.hackerfactor.com/blog/index.php?/archives/1013-C...


thanks!

There tends to be more information under the search term "perceptual hashing"

The secrecy of the inner tech is intentional.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: