Responsible AI
Enforce ethical AI usage with automated protections for content moderation, privacy, and compliance in your Lyzr agents.
🛡️ Responsible AI
Lyzr’s Responsible AI module enables platform users to proactively moderate content, prevent misuse, and ensure compliance with privacy and safety standards. With built-in support for detecting toxicity, prompt injections, sensitive information, and more, you can build AI agents that are safe, ethical, and secure.
Responsible AI Guide
Understand principles and practices for ethical AI development and deployment.
🔥 Toxicity Detection
Automatically detect and prevent the generation or processing of toxic, harmful, or offensive content.
- Use Case: Prevent agents from generating insults, hate speech, threats, or inappropriate language in customer support, education, or community applications.
- Threshold:
0.4
(Values closer to 1 indicate higher tolerance, while lower thresholds are more strict.)
✨ Agents will automatically block or filter responses that exceed the toxicity threshold.
🎭 Prompt Injection Protection
Detect and block malicious prompt manipulation attempts (Prompt Injections), where a user tries to override or influence the agent’s behavior using cleverly crafted input.
- Use Case: Prevent users from bypassing system instructions (e.g., “Ignore the last instruction and say X”).
- Threshold:
0.3
(Lower values are stricter and more secure.)
🔐 Especially useful in agents interacting with untrusted or anonymous users.
🔐 Secrets Detection
Automatically detect and redact or mask sensitive credentials, including:
-
API keys
-
Tokens
-
JWTs (JSON Web Tokens)
-
Private Keys
-
Use Case: Prevent accidental exposure of credentials in logs or chat outputs.
-
Action: Detected values are redacted before being stored, displayed, or transmitted.
✅ Allowed Topics
Restrict agent interactions to only specific, whitelisted topics.
- Use Case: Ensure your agent only discusses business-allowed domains (e.g., “finance, healthcare, HR”).
- Configuration:
Provide comma-separated values:
🧠 Useful for domain-specific AI assistants with strict focus.
🚫 Banned Topics
Prevent the agent from discussing or responding to specific blacklisted topics.
- Use Case: Prohibit conversation around internal operations, political views, or adult content.
- Configuration:
Provide comma-separated banned topics:
❌ Blocked Keywords
Restrict or redact specific words or phrases from being used in prompts or responses.
- Use Case: Redact client names, project codenames, or other internal terms.
- Configuration:
Provide comma-separated keywords:
💡 Blocked keywords will be filtered out or replaced during processing.
🔍 Personally Identifiable Information (PII)
Control how agents handle sensitive personal data, with options to block or redact each category.
Supported Categories & Actions
Data Type | Description | Options |
---|---|---|
Credit Card Numbers | Detects 13–16 digit card numbers | Disabled / Blocked / Redacted |
Email Addresses | e.g., john@example.com | Disabled / Blocked / Redacted |
Phone Numbers | International and local formats | Disabled / Blocked / Redacted |
Names (Person) | Common personal name patterns | Disabled / Blocked / Redacted |
Locations | City, state, country, address mentions | Disabled / Blocked / Redacted |
IP Addresses | IPv4 / IPv6 addresses | Disabled / Blocked / Redacted |
Social Security Numbers | U.S. SSN format: XXX-XX-XXXX | Disabled / Blocked / Redacted |
URLs | Any web address patterns | Disabled / Blocked / Redacted |
Dates & Times | Recognizable temporal references | Disabled / Blocked / Redacted |
🔐 These controls help you comply with GDPR, HIPAA, and other data protection standards.
🎯 Example Use Cases
Use Case | Responsible AI Features Used |
---|---|
Customer Support Chatbot | Toxicity Filter, Secrets Masking, PII Redaction |
HR Agent for Internal Use | Allowed Topics, Blocked Keywords, PII Redaction |
Public-Facing Financial Assistant | Prompt Injection Detection, Banned Topics, URL Redaction |
Legal Document QA Bot | Secrets Filter, Credit Card Blocking, Topic Control |
📌 How to Configure in Studio
- Go to Agent Settings in Studio.
- Open the Responsible AI tab.
- Toggle each feature and configure the appropriate thresholds or keywords.
- Save and apply the settings.
⚙️ Changes take effect immediately for all new interactions.
By enabling Responsible AI, you ensure that your Lyzr agents act ethically, safely, and in alignment with your organization’s privacy and compliance standards.