With Privatemode, you get on-prem like privacy combined
with the convenience and flexibility of the cloud.
Trusted by security‑critical teams at enterprises and public institutions
Product
Overview
Privatemode gives you access to leading open-weight models through one secure endpoint, so you can choose the best model for each workload without exposing sensitive data.
The inference API runs on modern GPU infrastructure to deliver low‑latency responses and high token throughput, even under heavy load. You keep real‑time performance for agents and user‑facing apps without managing your own GPU clusters.
Privatemode layers end‑to‑end encryption with confidential computing and verifiable attestation, so you can verify where and how your data is processed. Explore the security architecture to see how key management and attestation defend against both external attacks and insider access.

The Privatemode Encryption Proxy is fully compatible with the OpenAI and Anthropic API, including existing SDKs and client libraries. Point your apps to the proxy URL and reuse your code while all traffic is transparently encrypted end‑to‑end.
Predictable usage-based pricing with no hidden fees, from free tier to production. Enterprise plans with custom rate limits available on request.
Introduction
Confidential inference API
Run the Privatemode Encryption Proxy and keep your existing OpenAI clients, SDKs, and request formats—just point them to the new endpoint. Your apps talk to a familiar API while the proxy transparently encrypts prompts and decrypts responses for you.
Prompts and responses are encrypted before they leave your infrastructure and stay encrypted in transit and at rest. Only the AI inside Privatemode’s confidential‑computing environment can access plaintext, never operators or cloud providers.
Privatemode uses hardware‑based confidential computing to keep data encrypted even while it is processed in main memory. Remote attestation verifies the runtime and models before any request is decrypted, so only approved code ever sees plaintext.
How to get started
In 3 simple steps
docker run -p 8080:8080 ghcr.io/edgelesssys/privatemode/privatemode-proxy:latest --apiKey <your-api-key>The proxy verifies the integrity of the Privatemode service using confindential computing-based remote attestation. The proxy also encrypts all data before sending and decrypts data it receives.
#!/usr/bin/env bash
curl localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-oss-120b",
"messages": [
{
"role": "user",
"content": "Hello Privatemode!"
}
]
}'You now have programmatic access to an end-to-end secure AI. The proxy is compatible with the OpenAI Chat API. Process your own sensitive data or provide trustworthy services to others – the possibilities are endless.
Sign up for free and get your API key right away. Works in most tools that work with OpenAI or Anthropic.
Comparison
Cloud-grade AI with end-to-end encryption and confidential computing. Built for industries where compliance isn't optional.
Powerful AI, but prompts and data are processed unencrypted on OpenAI's servers, with no verifiable privacy guarantees
Full data control, but requires dedicated hardware, in-house ML ops, and ongoing maintenance.

Case Studies

Case Study
The German Federal Employment Agency


Case Study
Kauz.ai

Case Study
HÄVG
FAQ
Technical Details
Install Docker, run the Privatemode proxy with your API key, and start sending requests to localhost:8080. The proxy handles all encryption and attestation automatically. Full instructions are in the quickstart guide — setup takes around 10 minutes.
We're happy to show you around and give an overview of what's possible.