OpenAI-compatible API with Intel TDX hardware encryption. Drop-in replacement: change one URL, get hardware-level data protection. 8 pre-built agent templates + connect any agent. Python, Node.js, LangChain, CrewAI, OpenClaw compatible. From $0.008/1K tokens.
OpenAI-compatible. Change one URL, every request runs inside Intel TDX hardware enclaves. Works with Python, Node.js, LangChain, CrewAI, or any OpenAI SDK.
Create an account, then go to Dashboard → API Keys → Create Key. Your key starts with vgpu_.
Replace your OpenAI base URL with:
https://api.voltagegpu.com/v1/confidentialUse any agent slug as the model ID (see table below), or Qwen/Qwen2.5-7B-Instruct for general use.
curl https://api.voltagegpu.com/v1/confidential/chat/completions \
-H "Authorization: Bearer vgpu_YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "contract-analyst",
"messages": [
{"role": "user", "content": "Review this NDA clause: The Receiving Party shall not disclose any Confidential Information for 5 years..."}
],
"max_tokens": 2048,
"stream": true
}'from openai import OpenAI
# One line to change — same SDK, same code, hardware-encrypted
client = OpenAI(
base_url="https://api.voltagegpu.com/v1/confidential",
api_key="vgpu_YOUR_API_KEY",
)
# Use any of the 8 agents as model ID
response = client.chat.completions.create(
model="contract-analyst", # or: financial-analyst, compliance-officer, etc.
messages=[
{"role": "user", "content": "Review this NDA and flag non-standard terms:\n\n" + nda_text}
],
max_tokens=2048,
)
print(response.choices[0].message.content)import OpenAI from 'openai';
const client = new OpenAI({
baseURL: 'https://api.voltagegpu.com/v1/confidential',
apiKey: 'vgpu_YOUR_API_KEY',
});
const response = await client.chat.completions.create({
model: 'financial-analyst',
messages: [
{ role: 'user', content: 'Analyze this P&L for red flags:\n\n' + financialData }
],
stream: true,
});
for await (const chunk of response) {
process.stdout.write(chunk.choices[0]?.delta?.content || '');
}from langchain_openai import ChatOpenAI
llm = ChatOpenAI(
base_url="https://api.voltagegpu.com/v1/confidential",
api_key="vgpu_YOUR_API_KEY",
model="compliance-officer",
)
response = llm.invoke("Assess GDPR compliance gaps in our AI usage policy")
print(response.content)Use the agent slug as the model parameter. Each agent has a specialized system prompt that activates automatically — you just send your document.
| Model ID | Agent | Industry | Best for |
|---|---|---|---|
contract-analyst | Contract Analyst | Legal | NDA review, clause risk, liability analysis |
financial-analyst | Financial Analyst | Finance | P&L analysis, fraud detection, audit findings |
compliance-officer | Compliance Officer | GRC | GDPR gaps, policy review, regulatory risk |
medical-analyst | Medical Records Analyst | Healthcare | Patient records, drug interactions, clinical trials |
due-diligence | Due Diligence Analyst | M&A | Target assessment, concentration risk, valuation |
cybersecurity-analyst | Cybersecurity Analyst | Security | Incident triage, threat analysis, response plans |
hr-analyst | HR & Workplace Analyst | HR | Investigation analysis, compliance, policy review |
tax-analyst | Tax & Transfer Pricing | Tax | Transfer pricing review, tax exposure, structure analysis |
You can also list agents programmatically: GET /v1/confidential/models
/v1/confidential/chat/completionsSend messages and get a response from a confidential agent. Supports streaming.
| Parameter | Type | Required | Description |
|---|---|---|---|
model | string | Yes | Agent slug (e.g. contract-analyst) or model ID |
messages | array | Yes | Array of {role, content} objects. Max 100 messages, 200K chars total. |
max_tokens | integer | No | Max response tokens. Default: agent-specific (typically 4096). |
temperature | float | No | 0.0-1.0. Default: agent-specific (0.08-0.12 for precision). |
stream | boolean | No | If true, response is streamed as SSE. Default: false. |
| Header | Description |
|---|---|
X-Confidential | true if processed in TDX enclave |
X-Agent | Agent slug used |
X-Provider | Infrastructure provider (targon) |
/v1/confidential/modelsList all available confidential agents and models. No authentication required.
All requests require an API key in the Authorization header:
Authorization: Bearer vgpu_YOUR_API_KEY
vgpu_| Code | Meaning | What to do |
|---|---|---|
401 | Invalid or missing API key | Check your Authorization header |
402 | Insufficient balance | Top up at voltagegpu.com/billing |
429 | Rate limit exceeded | Wait or upgrade plan (Developer: 60/min, Team: 300/min) |
503 | TDX enclave starting up | Retry after 30-60s. We never fall back to non-encrypted infrastructure. |
1 API key, 60 req/min
10 API keys, 300 req/min, audit log
Unlimited keys, dedicated enclave, SLA
Every request runs in an Intel TDX enclave — no premium tier required for confidential compute. Billing is per-token, debited from your account balance.
TLS 1.3 for all API calls. API keys hashed with SHA-256 before storage.
Intel TDX enclaves on NVIDIA H200 GPUs. Data encrypted in CPU memory during inference. Operated by Manifold Labs (Bittensor Subnet 4).
Zero retention. Documents are destroyed after analysis. Encrypted RAM is flushed. We store usage metadata (tokens, cost) but never your content.
EU-based company (France). GDPR Art. 25 (data protection by design). DPA available on request.
Create an account, get your API key, change one URL. Every AI request runs in a hardware enclave from that moment on.