Stability AI: Stability AI uses AI-powered content filters on its platforms and API to detect and block policy-violating content, including prompts or images that may produce NSFW material, child sexual abuse material (CSAM), or other prohibited content. This system operates automatically on both user inputs and generated outputs. | AI Trace
Content ModerationVerified
Stability AI uses AI-powered content filters on its platforms and API to detect and block policy-violating content, including prompts or images that may produce NSFW material, child sexual abuse material (CSAM), or other prohibited content. This system operates automatically on both user inputs and generated outputs.
Details
According to Stability AI's published integrity transparency report, the company operates multiple layers of content moderation: in-house text prompt filters that block generation requests violating the Acceptable Use Policy; in-house NSFW image classifiers that flag uploaded images and video; CSAM hashing systems using industry hash lists from Thorn's Safer and the Internet Watch Foundation (IWF); and a combination of automated and human review by an internal Integrity team. The company also uses in-house NSFW classifiers and open-source classifiers to filter training data. Confirmed CSAM is reported to NCMEC.
Products affected
Stability AI Developer Platform APIStable AssistantStable Diffusion (hosted versions)