Toxicity Detection
AI systems that identify harmful, offensive, or abusive language in text. Detectors classify content for hate speech, threats, insults, and profanity. They serve as guardrails for AI outputs and moderation tools for online platforms.