Details
Roblox's moderation infrastructure uses large transformer-based machine learning models optimized through distillation and quantization to classify text, images, voice, and 3D assets for policy violations in milliseconds. In March 2026, Roblox launched a new real-time multimodal moderation system that evaluates entire in-game scenes—including avatars, text, and 3D objects together—to catch content that passes individual item checks but is problematic in combination (e.g., an offensive drawing combined with an avatar). As of early 2026, the system is shutting down approximately 5,000 servers per day for violations and is targeting 100% playtime monitoring. Roblox states that AI is deployed for moderation only when it outperforms humans in both precision and recall at scale, with human moderators retained for nuanced cases, complex investigations, and appeals.
Have evidence about Roblox Corporation's AI practices? Submit a report.
Report a Sighting →