The Architecture of Trust: Securing AI in Enterprise Environments

Organizations moving AI into production are discovering a security problem that was not salient during experimentation: the question of what the AI system is allowed to access, what it is allowed to do, and how the organization knows those boundaries are being respected.

Part of the Phase II — Understanding series

By Michael E. Ruiz

Organizations that have moved beyond AI experimentation and into production deployment are discovering a category of security problem that was not salient when AI was a proof-of-concept. The problem is trust: specifically, the question of what the AI system is allowed to access, what it is allowed to do with that access, and how the organization knows, in real time, that those boundaries are being respected. Traditional security models were built around human users and defined application behaviors. AI systems are neither.

The trust problem in enterprise AI has several distinct dimensions. The first is data access. Language models and AI assistants connected to enterprise data sources, including documents, emails, and databases, can access and synthesize information across scope boundaries that would be immediately obvious to a human. A human analyst who stumbles across a document they should not have read generates an access log and, if the access is flagged, a conversation. An AI system that traverses the same boundary in the process of completing a task generates a response. The traversal may not be visible unless the access logging is specifically configured to capture AI-initiated requests alongside human-initiated ones.

The second dimension is action authority. AI agents, which are systems that can take actions on behalf of users rather than simply generating text, operate with a combination of the permissions of the user who invoked them and the architectural constraints of the system design. If those constraints are not explicit and enforced at the system level, the agent can take actions that the user did not intend and would not have authorized if presented with them directly. This is not a theoretical concern. It is a design problem that emerges when AI systems are connected to action-capable APIs without a principled framework for what actions require explicit user confirmation versus what can be executed autonomously.

The third dimension is prompt security. AI systems exposed to external content, including documents, web pages, and user-submitted data, are subject to prompt injection: adversarially constructed inputs designed to redirect the AI's behavior. An AI assistant that reads emails on behalf of a user can be instructed, by a malicious email, to exfiltrate data, modify calendar entries, or send messages the user did not author. The security model for this threat is not the same as the security model for traditional injection attacks, because the vector is natural language rather than structured code, and the behavior being exploited is the model's tendency to follow instructions embedded in the content it processes.

Addressing these dimensions requires treating AI systems as a new category of principal in the enterprise security architecture: not a user, not an application, but something with characteristics of both and the security model of neither.

The practical requirements follow from this framing. AI systems need defined permission scopes that reflect what they actually need to do their job, enforced at the access control layer. They need audit trails that capture their actions at the same level of detail captured for human users. They need architectural boundaries that limit their exposure to adversarial content without eliminating their ability to be useful. And they need governance frameworks that define who is accountable for their behavior and what the remediation process is when they behave unexpectedly.

The organizations building this architecture now are doing work that will look prescient in three years. The ones treating AI security as a problem for later, after deployment and after the organization gets comfortable with the technology, are accumulating technical and governance debt that will be harder to address once AI systems are embedded in production workflows. Trust boundaries are easier to design in than to retrofit.

These ideas are available as keynote presentations and executive briefings. Explore speaking topics →