Detecting Child Sexual Abuse Material¶
Using perception and a subscription to Thorn’s Safer service, you can easily check for child sexual abuse material against a database of known bad content without having to send any images to a third party. You do this by sending compact, irreversible image hashes to get matches with a high degree of precision. We support matching using 16x16 PHash hashes and md5 hashes.
See usage example below. Please contact email@example.com to discuss Thorn’s Safer service and subscription options and visit getsafer.io to learn more.
from perception import tools matcher = tools.SaferMatcher( api_key='YOUR_API_KEY', url='MATCHING_SERVICE_URL' ) matches = matcher.match(['myfile.jpg'])
In some cases, you may have a username/password instead of an API key, in which case you can pass those instead (see API documentation for details).