Oblivious is not always Private

As AI systems process more and more sensitive personal information, companies have started developing "private" AI offerings of various sorts. In this blog post, we unpack what privacy means in the context of AI applications, and why OHTTP is often not the right tool for AI privacy.
As demand for privacy in AI applications continues to grow, especially in enterprise contexts, we see different definitions of privacy emerge from AI providers.
For some AI providers, "privacy" merely means vague promises like "we don't train on your data"—while saying nothing about other ways they might use your data.
Other companies claim privacy by using Oblivious HTTP (OHTTP), a protocol designed to decouple user identity from web requests. OHTTP is an excellent tool for identity protection and a critical component in many web-based systems. However, it's important to understand its specific design purpose, design context, and limitations when applied to AI systems.
In particular, OHTTP was designed for identity anonymization (hiding IP addresses) when performing web requests. It was not designed for applications where data privacy is needed. As such, it fails to provide meaningful privacy when it comes to AI applications, as we explain next.
Understanding OHTTP: Protecting Who, Not What
OHTTP decouples user identity from web request content. It works by routing encrypted Internet requests and responses through proxies that sit between the client and its destination, separating who initiated the request from what was sent in the request. This provides unlinkability (i.e., anonymity) by hiding the user's IP address from the service provider.
How OHTTP Works
OHTTP functions like a VPN (e.g., Proton or NordVPN) by hiding the user's IP address from the destination server.
However, unlike VPNs, where the VPN service sees both the client and destination IP, OHTTP proxies the connection through two non-colluding relays.
This ensures that the first proxy – which sees the client IP address – does not see the final destination of the request.
Similarly, the second proxy – which knows the destination IP address – does not see the source of the connection.
Put together, OHTTP guarantees that neither proxy sees both the client IP address and the destination IP address, which grants more anonymity to end users compared to a VPN.
OHTTP provides identity unlinkability, not data privacy
While OHTTP prevents linking the user identity to requests, it is important to understand that it only protects who sent the data, not the data itself from the destination server. This works great for fixed transactional requests that reveal information over time through metadata, such as DNS queries or web browsing. In particular, OHTTP is designed for either accessing publicly available content (such as a video on YouTube) or sending generic data such as a crash report that does not contain personally-identifiable information (PII). However, when personal data is present in the request, all bets are off.
OHTTP + AI ≠ Privacy
For AI applications processing emails, documents, or other content containing sensitive data or PII, this creates a big problem: Once decrypted at the destination, the actual content (including any PII) is fully exposed to the receiving server. If your data contains identifying information (like your name), OHTTP does not provide any meaningful privacy guarantees, since the PII is linked to your identity! It's analogous to wearing a hood and mask to protect your identity and then pinning your ID to your forehead.
So is OHTTP Useful?
Yes! OHTTP is a critical component of a comprehensive Internet privacy solution. In many situations, the unlikability property achieved by OHTTP is vital in preventing ISPs and other network observers from potentially learning sensitive information about you. For example, using OHTTP prevents ISPs from learning your browsing history or Apple from learning which applications you're running through crash reports and other telemetry sent from your devices. However, neither of these contexts involve sending over personal information that can be linked back to you.
For AI applications, we cannot assume that conversations and other queries are free of PII. Copy-pasting an email from your boss into ChatGPT? That contains PII and even potentially proprietary company information. In this case, ChatGPT can link the request back to you, regardless of whether you anonymized the connection through OHTTP. To have privacy in such cases, we need a stronger property: data privacy.
Tinfoil Privacy: Protecting the Data Itself
Tinfoil provides actual data privacy through secure enclaves. Tinfoil protects code and data inside the enclave with confidentiality and integrity. Data confidentiality guarantees that only the inference logic ever has access to the data, and that all entities (including Tinfoil) outside the secure enclave environment never see your data.
How Tinfoil protects your privacy
Secure enclaves provide an isolated, protected environment within the processor for sensitive operations. This isolation is enforced at the hardware level, ensuring neither the operating system nor other applications can access data being processed within the enclave. The secure enclave uses hardware-based memory encryption to isolate specific application code and data in memory.
For AI applications processing personal data, secure enclaves protect the data during use. This means personal data remains encrypted even in memory during processing, allowing Tinfoil to run AI models in the cloud while maintaining full confidentiality of the data.
Conclusion
As AI increasingly processes sensitive personal information, understanding the distinction between privacy technologies is crucial. OHTTP (and even VPNs) can create the illusion of privacy by hiding your IP address, but they are not designed to protect you from your data. It is easy to shoot yourself in the foot by assuming that network-level anonymity also means your data stays private from the processor – don't fall for that mistake.
For users trusting AI systems with sensitive information—whether it's proprietary code, work emails, or personal documents—confidential computing or fully on-prem deployments are the only solutions that guarantee privacy regardless of the data you feed into the AI system. You can always use OHTTP in conjunction with Tinfoil to hide your IP address from us, but regardless of whether or not you use OHTTP, we never see your data.
As AI systems become more integrated into our professional and personal lives, the question isn't just "Does anyone know I used this service?" but rather "Can anyone see what I shared with this service?" For true privacy in AI applications deployed on the cloud, confidential computing needs to be part of the picture.
Subscribe for Updates
Stay up to date with our latest blog posts and announcements.