Stars
Stars
Back to Blog
SecurityJan 10, 20268 min read

How to Build AI Solutions Without Giving Up Privacy

Andrew Jiang
Andrew JiangCo-founder and CEO

AI is becoming the interface to everything. It drafts, summarizes, searches, and decides. But the trade is rarely stated plainly: to be useful, AI needs your private context - messages, contacts, history, intent.

In crypto, that context isn't just personal. It's economically sensitive. Deal flow. Alpha. Founder chats. Allocation discussions. Timing. Relationships. If it leaks, it doesn't just embarrass you - it can be exploited.

Most AI products still follow the same pattern: ingest user data, decrypt it server-side, run a model, store outputs (and often inputs), and pray your access controls stay perfect forever.

They won't.

"Encryption-at-rest" is not a privacy strategy. It protects data while it's sitting still. The moment you use that data - embedding it, classifying it, routing it to an LLM - it becomes plaintext somewhere in memory. And plaintext is where real failures happen: a debug log, an over-broad role, a compromised host, a vendor retention setting you didn't intend.

If you're building AI on sensitive data, the goal isn't better privacy promises. It's building systems where even the operator can't see the data.

At LockIn, we're building sales intelligence off Telegram and X networks - surfacing warm paths and intent signals. The raw content is deeply sensitive, so we designed around a stronger requirement: privacy that can be verified, not merely asserted.

The privacy gap in AI products

AI systems don't just store data - they transform it. And transformation creates copies. A single message becomes chunks. Chunks become embeddings. Embeddings become retrieved context. Context becomes prompts. Prompts become outputs. Every step is another surface where sensitive content can accidentally end up.

That's why AI privacy failures rarely look like a dramatic breach. They look like normal engineering:

  • A request body captured by observability
  • A queue that retries and persists payloads
  • A "temporary" debug flag that ships to prod
  • A contractor pulling samples to evaluate quality
  • A vendor system retaining more than you assumed

The default architecture makes plaintext easy to touch.

So the real question isn't "do we encrypt?" It's: Where is plaintext allowed to exist - and who can access that place?

TEEs and attestation

If you want "even the operator can't see it," you need two building blocks:

  1. A place where plaintext is protected from the infrastructure running it
  2. A way to prove what code is running in that place

That's what Trusted Execution Environments (TEEs) and attestation give you.

What is a TEE?

A TEE is hardware-backed isolated compute. Think of it as a sealed room: data can enter, results can leave, but the host machine can't look inside while computation happens. No casual admin access. No "just SSH in." No scraping memory from the outside.

This changes the rules. Plaintext is no longer "server default." It becomes a tightly controlled exception.

"Prove it" with attestation

A sealed room is great - unless nobody can verify what's inside.

Attestation is the cryptographic proof a TEE can produce that says: this is the software I'm running. If the code changes, the measurement changes. If the measurement changes, verification fails. If verification fails, sensitive workflows don't run.

That's how you replace "trust us" with "verify us."

The privacy lifecycle

Once you enforce the boundary, the lifecycle becomes simple and auditable:

Encrypted → Enters enclave → Decrypted (inside) → Processed → Re-encrypted → Leaves enclave

Plaintext exists in exactly one place: inside hardware-protected memory running verified code.

This is how we built LockIn. Your messages never leave a secure enclave. Even we can't see them. And you don't have to take our word for it - you can verify it.

Stars

Ready to Lock In More Deals?

Join top Web3 teams turning Telegram noise into qualified pipeline
logoLead intelligence for Web3
© 2025 LockIn. All rights reserved.