In addition to the CAID Training Data, there is:- Training data categorized by CSA Investigators.
– Fine Grained Categorization – not just detection.
– Extensive performance assessment (larger dataset means better training and better evaluation).
– The fact it is built by a team that includes computer vision experts and CSA Investigation expertise (we know the problem and the solution).
– Video classifier to go along with the image classifier.
No – the classifier is not trained on your data and no CSAM or even performance data is shared back to Krunam. Instead, we prefer to have regular performance check-ins where you share information that you are comfortable sharing with Krunam so that we can increase performance.
This is a giant leap forward from perceptual hashing – moving from only known images to finding the entire class of CSAM through computer vision. For the developing AI solutions, that’s where our dataset, training procedure and CV/AI expertise set us apart.
Our classifier uses a computer vision algorithm that was trained using state of the art AI and deep learning techniques. The technology identifies “hallmarks” of CSAM from being trained on the CAID dataset, the pre-eminent collection of privacy safe, legally obtained CSAM.
Yes, fully. We never actually house the CSAM data – and all of our non-CSAM data is legally obtained. We have no individual identification ability built into the system, as it looks at classes of behavior and conduct.
Once you are under an NDA, we can share our internal testing, which is rigorous and multi-layered. We have held back a sizable portion of our CSAM dataset for testing, and built a robust proprietary hard negative dataset.
Performance is high enough that our law enforcement partners are using it in real operational work to speed up detection and categorization, and they also don’t want knowledge of the specific results out in the public domain. Please contact us to discuss this further.