Generative Al Principles to Prevent Child Sexual Abuse
in partnership with
We have joined Thorn, All Tech is Human, and other leading companies in the effort to prevent the misuse of generative AI technologies to perpetrate, proliferate, and further sexual harms against children.
We are committed to implementing preventative and proactive principles into our generative AI technologies and products.
We are also committed to publishing transparency reports annually as of 2025, documenting and sharing our progress on these principles.
As part of this Safety by Design effort, Stability AI is committed to the following principles:
-
Develop:
Develop, build and train generative AI models that proactively address child safety risks
-
Deploy:
Release and distribute generative AI models after they have been trained and evaluated for child safety, providing protections throughout the process.
-
Maintain:
Maintain model and platform safety by continuing to actively understand and respond to child safety risks.
To ensure tangible action, Stability AI is also committing to the following mitigations, stemming from the above principles. We will:
Develop
Responsibly source our training data: avoid data that have a known risk of containing CSAM and CSEM.
Detect, remove, and report CSAM and CSEM from our training data.
Separate depictions/representations of children from adult sexual content in our training datasets for our video, image or audio generation models.
Conduct red teaming, incorporating structured, scalable, and consistent stress testing of our models.
Include content provenance by default in any image or video that our models output.
Define specific training data and model development policies.
Prohibit customer use of our models to further sexual harms against children.
Deploy
Detect abusive content (CSAM, AIG-CSAM, and CSEM) in inputs and outputs.
Include user reporting, feedback, or flagging options.
Include an enforcement mechanism.
Assess models before hosting on our platforms, collaborating with LE agencies globally where possible, to legally and safely test our models for their potential to generate AIG-CSAM and CSEM.
Include prevention messaging for CSAM solicitation.
Incorporate phased deployment when risks are unknown, monitoring for abuse in early stages before launching broadly.
Incorporate a child safety section into our model cards.
Maintain
Remove services for “nudifying” images of children from search results.
When reporting to NCMEC, use the Generative AI File Annotation.
Detect, report, remove, and prevent CSAM, AIG-CSAM and CSEM on our platforms.
Maintain the quality of our mitigations.
Disallow the use of generative AI to deceive others for the purpose of sexually harming children. Explicitly ban AIG-CSAM from our platforms.
Leverage Open Source Intelligence (OSINT) capabilities to understand how our platforms, products and models are potentially being abused by bad actors.
The detailed nature of these specific mitigations and others to enact these principles are further recommended and defined in the associated whitepaper: Safety by Design for Generative AI: Preventing Child Sexual Abuse. More detail about the principles and the whitepaper can be found at https://teamthorn.co/gen-ai