Protect against adversarial attacks like jailbreaks and prompt injections as well as unauthorized tool calls
Implement adversarial testing program to validate system resilience against adversarial inputs and prompt injection attempts in line with adversarial threat taxonomy
Implement monitoring capabilities to detect and respond to adversarial inputs and prompt injection attempts
Implement controls to prevent over-disclosure of technical information about AI systems and organizational details that could enable adversarial targeting
Implement safeguards to prevent probing or scraping of external AI endpoints
Implement real-time input filtering using automated moderation tools
Implement safeguards to limit AI agent system access based on context and declared objectives
Establish and maintain access controls and admin privileges for AI systems in line with policy
Implement security measures for AI model deployment environments including encryption, access controls and authorization
Implement output limitations and obfuscation techniques to reduce information leakage
"We need a SOC 2 for AI agents— a familiar, actionable standard for security and trust."
"Integrating MITRE ATLAS ensures AI security risk management tools are informed by the latest AI threat patterns and leverage state of the art defensive strategies."
"Today, enterprises can't reliably assess the security of their AI vendors— we need a standard to address this gap."
"Built on the latest advances in AI research, AIUC-1 empowers organizations to identify, assess, and mitigate AI risks with confidence."
"AIUC-1 standardizes how AI is adopted. That's powerful."
"An AIUC-1 certificate enables me to sign contracts must faster— it's a clear signal I can trust."