Got Privacy?

We're seeing AI impact everything from creative writing to coding. Most of us feel the tectonic shift that AI has introduced both in our personal and professional lives, and with that, some of us also realize the magnitude of the data we send to these AI-powered applications. It's no secret that many companies are eager to use this data for profit. Aravind Srinivas, the CEO of Perplexity, openly talks about this goal for his company.1
But what if we want to use AI while keeping our data private? In this blog post, we explore the options we have today for private AI and see secure enclaves emerge as the desirable option when it comes to speed, convenience, and privacy. This is why at Tinfoil, we choose secure enclaves as the primitive to build private AI.
The Privacy Challenge in AI
AI systems provide us with the most value when we share sensitive information. We need to trust them as much as our work and life partners (if not more). This creates an inherent privacy challenge: how can users benefit from AI assistance without compromising the confidentiality of their data? The key privacy risks in AI inference include:
- Data exposure - Sensitive information might be accessed by unauthorized parties during processing
- Data persistence - Our conversations and queries may be stored or logged for longer than promised or desired, making it impossible to know where our data ends up
- Secondary use - Information shared during inference might be used for purposes beyond generating the immediate response (e.g., profiling and advertising)
Current Options for Private AI
1. Policy-based ("pinky-promise") Privacy
Major AI providers like OpenAI and Anthropic offer privacy commitments through their terms of service, often including statements about data usage limitations.
How it works: These companies establish contractual agreements with specific privacy commitments about how they handle user inputs during chat and inference.
Strengths:
- Contractual commitments regarding specific data usage limitations
- Typically include defined data retention periods
- Enterprise offerings often feature enhanced privacy controls and guarantees
Limitations:
- Data typically processed in plaintext on provider infrastructure leaving it exposed to hackers, vulnerable to data leaks, and subject to future policy changes
- Limited technical mechanisms for users to verify data handling practices, we need to fully trust that these providers adhere to their promises
As such, policy-based approaches represent an important step in the privacy direction. However, they primarily rely on organizational controls and contractual agreements rather than technical mechanisms that can mathematically or physically prevent unauthorized data access.
Enforcing these policies often proves to be an impossible task for large organizations. For example, a leaked 2021 internal Meta document states:
We do not have an adequate level of control and explainability over how our systems use data, and thus we can't confidently make controlled policy changes or external commitments such as 'we will not use X data for Y purpose
which aptly highlighting the extent to which we cannot rely on privacy policies to keep our data secure.
2. Oblivious HTTP (OHTTP)
Services like Duck.ai and Brave Leo use the Oblivious HTTP protocol to provide anonymity when interacting with AI systems.
How it works: OHTTP routes requests through relay servers that separate the end user's identity from the content of their requests. It essentially hides who initiated the request from the service provider but not the data itself.
Limitations: Importantly, OHTTP is not a data privacy mechanism, which has limitations when applied to AI systems. We go in-depth into why OHTTP is not well-suited for AI services in our other blog post. In particular, your data is still visible even if your identity is anonymized, and so solutions relying on OHTTP effectively resort to the policy-based "pinky-promise" privacy.
3. Apple's Private Cloud Compute
Apple recently announced Private Cloud Compute (PCC) for its Apple Intelligence features, implementing a hardware-based approach to privacy.
How it works: PCC extends Apple's device security model to the cloud, using custom Apple silicon servers with secure enclaves and a hardened operating system.
Strengths:
- Hardware-level security protections
- End-to-end encryption from device to PCC nodes
- Data is only processed within secure enclaves and invisible to Apple
Limitations:
- Closed-source ecosystem with limited transparency
- Users must trust Apple's implementation
- Not available as a general-purpose solution for other AI applications
Apple's approach represents an important step in cloud privacy, leveraging their control over the entire hardware-software stack. However, its proprietary nature means users must still place significant trust in Apple's claims without the ability to fully verify the implementation.
4. PII Redaction
Companies like Private AI and Redactor.ai offer solutions to identify and remove personally identifiable information before it reaches AI systems.
How it works: Automated systems scan text, documents, images, or audio to detect and redact sensitive information (names, addresses, financial details, etc.) before sending to AI services.
Strengths:
- Prevents sensitive identifiers from reaching AI providers
- Can be added as an additional preprocessing layer to all existing AI systems
Limitations:
- Depends on accurate identification of PII (risks missing some PII or contextual identifiers)
- Does not protect the overall content from the AI provider
- May reduce the quality and accuracy of the responses
PII redaction is valuable as a component of a privacy strategy but doesn't solve the fundamental issue of sharing data with third-party AI providers.
5. On-Prem Deployments
Running AI models locally, either at home or on their own infrastructure can represent the strongest form of privacy, limiting explicit exposure to any third parties.
How it works: Organizations deploy AI models on their own servers or infrastructure, eliminating the need to share data with external providers.
Strengths:
- Data never leaves organizational boundaries
- Complete control over data storage and retention
- No dependency on third-party privacy practices
Limitations:
- Generally limited to smaller, less capable models due to computational resource constraints (GPUs are expensive)
- Requires technical expertise to deploy, secure, and maintain
- Significant infrastructure investments for larger models
- Limited access to the latest AI capabilities
While local hosting offers strong privacy protections, it comes with significant tradeoffs in model capabilities, costs, and maintenance complexity. It can even introduce security vulnerabilities by relying on bespoke setups that may be more susceptible to hacking compared to standard cloud deployment options.
6. FHE-Based Solutions (Fully Homomorphic Encryption)
Startups like Lattica.ai are developing systems that use Fully Homomorphic Encryption to enable computation on encrypted data in the cloud.
How it works: FHE allows AI models to process data while it remains encrypted, providing the ability to use AI without ever exposing the raw data to the provider in any capacity.
Strengths:
- Data remains encrypted throughout processing
- Neither the cloud provider nor the model provider can see the unencrypted data
- Offers mathematical privacy guarantees rather than policy-based promises
Limitations:
- Significant computational overhead making it suitable only for tiny models
- Still in early commercial stages with very limited real-world adoption
- Cannot be used with large models due to massive performance constraints
- Vulnerable to attacks by malicious inference servers,2 which requires significant additional computational overheads to mitigate effectively
FHE represents a promising direction but has traditionally faced practical challenges in performance that have limited widespread adoption. While some companies are working to make FHE more practical by optimizing for specific AI workloads, significant theoretical and architectural breakthroughs are still needed before FHE becomes a sufficiently practical tool for real-world applications.
7. The Secure Enclave Approach
Solutions like Tinfoil and services built on NVIDIA's confidential computing capabilities offer hardware-enforced isolation for AI workloads, similarly to Apple's Private Cloud Compute solution.
How it works: These systems leverage hardware-based trusted execution environments (TEEs) that isolate memory and processing to prevent unauthorized access, even from administrators of the host machine.
Privacy strengths:
- Hardware-enforced isolation of computing resources
- Attestation mechanisms provide verification of the security environment
- Compatible with high-performance AI hardware (GPUs)
Privacy limitations:
- Requires trust in the hardware manufacturer (e.g., NVIDIA and AMD)
- Vulnerable to side-channel attacks, which we cover in-depth in our other blog post about side-channels and our mitigation strategies
NVIDIA's Hopper and Blackwell GPUs support confidential computing, enabling data privacy for AI workloads without performance overheads. This represents a state-of-the-art approach to practical AI with strong data confidentiality guarantees.
The benefits of Tinfoil
Among these approaches, Tinfoil's use of secure enclave technology offers a notable solution for private AI inference. Here are the key aspects of why Tinfoil outperforms other solutions (summarized in Table 1):
- Hardware-Based Security Foundation: Tinfoil leverages hardware-based TEEs (AMD SEV-SNP or Intel TDX) coupled with NVIDIA GPUs that support confidential computing, creating a secure processing environment for AI workloads.
- Comprehensive Memory Protection: Tinfoil's approach includes memory encryption throughout the processing pipeline, providing protection against various types of access attempts during data processing.
- Verification Mechanisms: The architecture allows for verification of security properties through attestation, helping organizations confirm the integrity of the secure environment where their data is processed.
- AI-Specific Design: The platform is specifically engineered around AI applications, with an abstraction layer designed for the unique requirements of private inference use cases.
- Transparency Through Open Source: Users can review open-source components to better understand the security implementation, enhancing confidence in the protection mechanisms.
- No Performance tradeoffs: The approach is designed to maintain strong performance while adding security protections, making it suitable for production AI inference workloads.
Conclusion
The landscape of private AI solutions encompasses diverse approaches, each with distinct privacy properties suited to different organizational needs and risk profiles. While commercial providers' policy-based approaches offer convenience, they lack concrete guarantees. Data transformation provides flexible obfuscation, and hardware-based solutions like secure enclaves deliver robust technical guarantees. Though FHE shows promise as a cryptographic approach, its performance and scalability remain constrained.
Organizations handling sensitive data with AI systems must choose privacy technology based on four key factors: data sensitivity, performance requirements, implementation complexity, and verification capabilities. Hardware-based approaches using secure enclaves — like Tinfoil's implementation — strike an optimal balance between strong technical protections and real-world performance for production AI workloads.
The future of private AI lies in combining these complementary approaches — merging identity protection, hardware security, local processing, and cryptographic advances — to establish comprehensive privacy for next-generation AI applications.
Footnotes
Subscribe for Updates
Stay up to date with our latest blog posts and announcements.