Dashboard
Credits remaining, monthly consumption, total requests at a glance.
OpenAI-compatible as a drop-in replacement. With anonymization per parameter.
The anymize developer console is your control center for API integrations: one API key for every frontier model (GPT, Claude, Gemini, Mistral, Perplexity, Kimi) plus our own models (Waterfall, Fountain). Supports the OpenAI schema – swap the base URL and your existing code keeps running. With anonymization when you need it. With fallback models when a provider fails. With Zero Data Retention when your compliance team demands it.
What you get
The developer console bundles every management function you need for productive API use.
Credits remaining, monthly consumption, total requests at a glance.
Create, name, rotate and revoke multiple keys in parallel.
Every available model with price per 1M input/output tokens.
Test models live with system prompt, temperature, max tokens, top p.
Fallback models, Zero Data Retention – one toggle per feature.
Cost tracking, rate limits, complete logs of every request.
All reachable via Back to chat – the console is not a separate product but the switch next to your normal anymize work.
OpenAI-compatible
The anymize API implements the OpenAI schema for chat completions. Any code that speaks to OpenAI today – the official Python SDK, the Node SDK, LangChain, LlamaIndex, your own backend or a no-code tool – works against anymize the moment you swap two values.
1from openai import OpenAI23client = OpenAI(4api_key="sk-proj-...",Changed5base_url="https://api.openai.com/v1"Changed6)
1from openai import OpenAI23client = OpenAI(4api_key="anymize_...",Changed5base_url="https://app.anymize.ai/api/v1/llm"Changed6)
No rewrite, no new SDK dependency, no parser overhaul. The chat/completions endpoint schema is identical – same roles, same messages, same parameters (temperature, max_tokens, top_p, stop, etc.). Streaming is supported.
What differs: the model names. Instead of "model": "gpt-5" at OpenAI you address anymize identifiers ("openai/gpt-5", "anthropic/claude-opus-4.7", "google/gemini-2.5-flash", "waterfall-1.0", "fountain-1.0"). The current list you can query at any time via GET /models.
Anonymization per parameter
The exciting trait of the anymize API: you enable anonymization via a parameter in the request. For every single call you decide whether the content runs through our anonymization pipeline before it reaches the chosen model.
You want to use a frontier model (GPT, Claude, Gemini) but your prompt contains personal data. If you enable anonymization, anymize replaces names, addresses, IBANs, case numbers and 40+ more categories before sending. The model's answer is automatically retranslated to your original data before it is returned. From your application's point of view the API behaves like a regular OpenAI API – only without any third-country transfer of personal data.
For content without personal reference – public studies, internal handbooks, structured data – you skip the anonymization step. Direct route, lower latency.
If you use our own models Waterfall or Fountain, you don't need anonymization in the first place: the models run with us in the EU, the data never leaves the EU. Ideal for code, Excel and JSON work, where anonymization would destroy semantic value.
The exact parameter syntax is in the API documentation – either via the dedicated endpoint for anonymized calls or as a body parameter on the standard endpoint, depending on your integration preference.
API keys
A single key is rarely enough. Production apps, staging environments, internal tools, ad-hoc scripts – for every setup a dedicated key, nameable, deletable, rotatable.
A compromised key is revoked without paralyzing the rest.
Per-key logs show which integration is making which requests.
Regular rotation is a standard recommendation in audits (GDPR, ISO 27001).
Models, prices, credits
The models overview in the console shows three groups, each model with input/output price per 1 million tokens – toggleable between EUR and credits.
European inference. No external providers, no additional contracts.
Variants hosted in Europe – for applications where the EU hosting aspect matters.
Claude, GPT, Gemini, Perplexity, Kimi at full strength. For sensitive data: enable anonymization.
The specialty: credits are universal. From the same wallet you pay for:
No separate API subscriptions on top of chat subscriptions. Frontier models run at provider price plus a thin anymize margin.
Chat playground
The playground is a reduced chat surface built for developer workflows – not for end-user content. You test exactly what your API calls would produce. Configuration goes one-to-one into the API body of your application.
Tune system prompts iteratively with the real parameters that later run in production.
Evaluate a new model candidate before it goes into production.
Reproduce an incident from a production API call – with the same parameters.
Fallback & Zero Data Retention
Under configuration sit two enterprise features that many API providers do not offer at all.
If your primary model does not answer in time (timeout, provider outage, rate limit), a fallback model kicks in automatically. You configure two in series – the second is often an in-house model, independent of external providers.
A toggle decides how anymize treats your request data. Metadata (tokens, model, timestamp) remain in both modes – content only when required.
Documents and content are stored in your account for later inspection (logs, reuse).
Request content is cleaned up after processing. Ideal for legal, healthcare, finance.
For industries with strict compliance requirements, ZDR is often set – the platform then behaves largely like a pass-through proxy with anonymization and fallback logic.
Usage tracking
Under usage you find three sub-areas – for budget, scaling and forensics.
Monthly and daily breakdown in EUR or credits, filterable by key, model and time range. You see immediately: which key burns the budget, which model is the cost driver, which workflow is inefficient.
Your current limits (depending on your plan) at a glance – plus history of how often you actually hit them. Anyone scaling productively sees in time whether an upgrade is due.
Every API request lands in the log: timestamp, API key, model, input/output tokens, latency, status. For debugging, audit, compliance evidence – and for the question “did our software really send it like that?”.
With Zero Data Retention enabled, metadata remains in the log, content is removed – billing is still complete.
Use cases
Six typical integration scenarios.
Base URL swapped, rest of the codebase unchanged
GDPR-compliant via anonymization, every model available, one contract
No-code workflows with AI steps
Standard LLM requests from your application
Fallback models, ZDR, central credit wallet
Tool calling, RAG, agents with multiple LLMs
Multiple models with one key, cost tracking per tool call
Automated summaries, draft generation
Anonymization protects client/customer data from US providers
Code review, refactoring, documentation generation
Waterfall/Fountain for code without anonymization – European stack
Frequently asked questions
Yes. The chat completions endpoint implements the OpenAI schema. You swap base_url and api_key in your existing code – that's it. The official OpenAI Python SDK, the Node SDK, LangChain, LlamaIndex, Claude Code, Aider, Cursor and anything else that supports the OpenAI format works immediately. Only the model names change to the anymize identifiers (e.g. openai/gpt-5, waterfall-1.0).
We stand behind anymize. And we know – when an AI tool touches client, patient or employee data, a demo video isn't enough. That's why we give you 14 days of full access – all models, all features, no credit card. Enough time to be certain, before you trust us.
Your AI workplace awaits.