The privacy architecture of a personal intelligence is the set of structural commitments that make deep access to a user's data safe. The architecture has six components. Encryption in transit and at rest. Prohibition on training. Data sovereignty. Sandboxed action. User confirmation. Auditability. Each addresses a specific class of risk, and the components together break what the security researcher Simon Willison has called the lethal trifecta of agent vulnerabilities. moccet is being built with these components from the beginning rather than retrofitted as compliance.
This essay explains what each component does, why the older privacy framework is no longer sufficient for systems that act on the user's behalf, and what to look for when evaluating a personal intelligence.
What is the lethal trifecta and why does it change AI privacy?
In June 2025, a security researcher demonstrated something that should have changed how the AI industry talks about agent privacy. The researcher sent a single crafted email to a corporate inbox running Microsoft 365 Copilot. The email contained no attachment, no link, and no instruction visible to the human recipient. What the email did contain was a hidden prompt, a passage of text that Copilot, during its routine summarisation of the user's inbox, would parse as an instruction. Within seconds of the email being summarised, the agent had pulled sensitive data from the user's OneDrive, SharePoint, and Microsoft Teams accounts and exfiltrated it through a domain Microsoft itself trusted. The user clicked nothing. The user opened nothing. The agent did the harm on its own, having been silently instructed by an attacker the user had never met.
The vulnerability, catalogued as CVE-2025-32711, earned a CVSS score of 9.3 out of 10. Microsoft patched the specific issue. The deeper architectural problem the attack exposed remained, and was repeated, by January 2026, in production exploits against four major AI agent products in five days, using variations of the same attack pattern.
The pattern has a name. Simon Willison, the British programmer who first popularised the term prompt injection in 2022, has called it the lethal trifecta. An AI agent is exploitable by design when the system has three properties simultaneously. Access to private data. Exposure to untrusted content. The ability to externally communicate. Most consumer AI agents in 2026 have all three. The architectural fact is that large language models cannot reliably distinguish between instructions and data. Models treat everything in their input as text that may contain instructions, and any document, email, or web page they ingest can therefore become a vector for taking control of the agent.
The conversation about privacy in personal AI tends to focus on the wrong layer. The traditional privacy questions still matter. Does the company sell my data? Do they use it for training? Who has access? The deeper concern in agent systems is that the data does not need to leave through traditional exfiltration channels for it to be compromised. The agent itself can be turned into the exfiltration channel by an external attacker with no access to the user's account and no knowledge of the user's password.
The architectural answer to this is not a privacy policy. The architectural answer is a system design that breaks the trifecta.
What are the six components of a privacy architecture?
The privacy architecture of a personal intelligence has six components. Each addresses a different class of risk. The first three are the layers that have always mattered for handling personal data. The last three are specific to systems that act on the user's behalf.
The first component is encryption in transit and at rest. The user's data is encrypted on the way to the system and encrypted while it sits in storage. Encryption does not protect against an insider with access to keys. Encryption protects against external actors who do not have them, against accidental exposure of storage volumes, and against the routine forms of breach that account for most security incidents in commercial software. Every credible product in the category has it. A product that does not is not viable.
The second component is the prohibition on training. The user's personal data is not used to train models. The reason this matters more in this category than in older ones is that AI training is irreversible. Once a model has learned from a piece of data, the data is part of the model in a way that survives the user's account being deleted. If a model has been trained on a user's emails and the user later cancels their service, the emails continue to influence the model's outputs to other users in ways that cannot be undone.
The third component is data sovereignty. The user's right to revoke access to any source at any time, the right to delete all their data, the right to export it, the right to know what the system has stored. These rights are increasingly mandated by regulation. GDPR in Europe, CCPA in California, similar frameworks in other jurisdictions. The rights also convert the user's relationship with the system from captivity to delegation. A user who cannot leave is not a user. A system whose retention is built on lock-in rather than usefulness is a system whose privacy guarantees are not worth much.
The fourth component is sandboxed action. When the system takes an action on the user's behalf, the action runs in a constrained environment. The system cannot do arbitrary things. The system can only do things from a defined list of allowed actions. The list is visible to the user. New actions are added carefully and explicitly, with documentation about what each action allows and what it does not. The sandbox is the layer that prevents an action loop, especially one influenced by a prompt-injection attack, from doing damage at scale.
The fifth component is user confirmation. Every action the system takes is confirmed by the user before it executes. The system drafts. The user approves. The action runs. Confirmation is structural, not optional, and it is the design feature that addresses Willison's third leg of the lethal trifecta, the ability to externally communicate. An agent that cannot send an email, transfer a file, or call an API without an explicit user confirmation has had the exfiltration leg of the trifecta surgically removed. Even if the agent is successfully prompt-injected by malicious content, the attack cannot complete without the user noticing. A fuller account of the engineering of confirmation is in the essay on what it means for an AI to take action.
The sixth component is auditability. The system maintains a log of every action it has taken, every piece of data it has read in deciding to take the action, every confirmation the user has given. The log is visible to the user in a form they can actually parse, not buried in a developer console. The audit trail is the receipt for the relationship.
