Content Moderation
Using AI to automatically detect and filter inappropriate, harmful, or policy-violating content. Moderation systems classify text, images, and video for toxicity, violence, and hate speech. Used by platforms to enforce community standards at scale.