Moderate Harmful or Sensitive Content
Clavata is often used to detect and flag content that violates platform policies or poses harm to users. Common categories include:- Harmful or illegal content like CSAM (Child Sexual Abuse Material), NCII (Non-Consensual Intimate Images), Bestiality, Incest or Rape
- Community guideline violations, like Nudity, Violence or Hate speech
Label Content or Build Taxonomies
In addition to harmful content, Clavata can be used to apply non-violative, descriptive labels to help structure your content, build robust taxonomies and improve recommendations, search, or tagging workflows. Some examples include:- Demographic or identity-based labels e.g., Straight, LGBTQ+, Fictional, Human, Mixed
- Content maturity e.g., PG, PG-13, R, X
- Language labels e.g., English, Spanish, Formal, Informal
- Fine-grained features in text or image e.g., Cat, Dog, Eyeliner, Blue eyes, Brown eyes
We’d Love to Hear From You
Whether you have a suggestion, feedback, or a bug to report, here are the best ways to get in touch:- In the App: Use the Feedback button for direct suggestions.
- On Slack: Reach out to the team in your shared channel.
- With your AM: Talk to your dedicated account manager.
- Via Email: Send a message to [email protected].