Sexism Detection flags whether a response contains sexist content. Output is a binary classification of whether a response is sexist or not.
Calculation method
Sexism detection is computed through a specialized process:Model Architecture
The detection system is built on a Small Language Model (SLM) that combines training from both open-source datasets and carefully curated internal datasets to identify various forms of sexist content.
Optimizing your AI system
Addressing Sexism in Your System
Implement guardrails: Flag responses before being served to prevent future occurrences.
Fine-tune models: Adjust model behavior to reduce sexist outputs.
Identify responses that contain sexist comments and take preventive measures to ensure fair and unbiased AI interactions.