This Plugion provides a set of tools to enhance the safety of generative AI applications with advanced guardrails for responsible AI. Azure AI Content Safety is an AI service that detects harmful user-generated and AI-generated content in applications and services.
Scans text for sexual content, violence, hate, and self harm with multi-severity levels.
Learn more abount the text moderation categories here.
Scans text for sexual content, violence, hate, and self harm with multi-severity levels.
Learn more abount the text moderation categories here.
Prompt Shields analyzes LLM input and detects adversarial user input attacks.
Learn more abount the prompt shields here.
See Plugin README for configuration and usage details.
See the official documentation for the supported languages.
This plugin is open-source and contributions are welcome. Please visit the GitHub repository to contribute.