Guardrails are now a core Nexus feature! For the complete documentation on using guardrails, see Guardrails in Nexus.
Overview
Guardrails are security policies that control how AI assistants interact with tools and data. Unlike traditional access controls that simply grant or deny access, guardrails provide nuanced, context-aware rules that make AI systems safer and more predictable.Why guardrails matter
AI assistants are increasingly powerful, with access to email, code repositories, databases, and more. This power comes with unique risks:- The AI is not a human: Cannot distinguish between legitimate requests and prompt injection attacks
- Scale amplifies risk: What takes a human hours, an AI can do in seconds
- Context windows have limits: Large responses can overwhelm the AI
Learn more
For complete documentation on guardrails, including:- How to manage guardrails through natural conversation
- Built-in PII protection and prompt injection detection
- Request and response guardrail types
- The guardrail hierarchy (account, toolkit, user)
- Common examples and best practices

