AI knowledge bases.
Your company memory.

Retrieval, not copy-paste. Your documents as a second source for every AI answer.

Contracts, policies, case files, product documentation, research notes – put them into a knowledge base once, and the AI pulls from them in every conversation. Intelligently, targeted, only the passages that are relevant right now. No manual uploads per chat. No forgotten context. All GDPR-compliant, shareable across your team, usable with our own European models too.

What is a knowledge base?

Retrieval-Augmented Generation,
explained plainly.

A knowledge base in anymize is a store for your documents, prepared for AI use. Technically it's built on Retrieval-Augmented Generation (RAG) – an architecture that has become the standard in the AI world for handling large internal bodies of knowledge.

The three core building blocks

01

Chunks

Your documents are split into semantically meaningful sections (typically a few hundred to a few thousand characters each). That might be single paragraphs, sections of a contract or entries in a protocol.

02

Embeddings

Each chunk is translated into a mathematical representation (a vector) that captures its meaning. Chunks with similar meaning sit close to each other in vector space.

03

Vector database

Stores all chunks and their embeddings. When you later ask a question, your question is turned into a vector too, and the database returns the chunks closest to it in meaning.

The result: The AI only receives the passages that match your question. Not everything at once. Not irrelevant pages. Targeted.

In practice you don't see any of this. You upload documents, switch on the knowledge base in chat, ask questions. anymize does the rest.

RAG vs. context window

Why not just
„dump everything into the chat“?

Frontier models have impressively large context windows – a hundred thousand tokens, sometimes over a million. In theory you could stuff an entire company archive into a single prompt. In practice, this runs into three hard problems:

Cost
Everything in context
Token costs scale linearly with context length. A 500-page archive costs several times more per request than a targeted retrieval.
Knowledge base (RAG)
Only the relevant chunks go to the model – typically 2–10 short passages. Cost stays stable no matter how large the database grows.
Quality
Everything in context
The longer the context, the stronger the „lost in the middle“ effect: models overlook information in the middle of long texts.
Knowledge base (RAG)
The model sees focused, preselected content. No dilution effect.
Scalability
Everything in context
Every new document has to be uploaded again. No reuse across sessions.
Knowledge base (RAG)
Indexed once, available in every chat. Shareable, updatable, searchable.
Rule of thumb

Individual documents you open for a single analysis – upload them directly into the chat. Anything you need again and again – put it into a knowledge base.

How you use them

Three routes.
One body of knowledge.

01

Switch it on directly in chat

Every knowledge base appears as a toggle in the chat interface. One click – and from now on the AI pulls fitting passages from your database for every answer. Several databases can be active in parallel: e.g. „Client XY“ + „Case-law archive“ + „Firm standards“.

02

Anchor them in projects

Working regularly on a specific case, client or topic? Attach the fitting knowledge bases to your project. New chats inside that project have them active automatically. No switching on, nothing to forget.

03

Query them via API

For automated workflows: the anymize API offers retrieval straight from a knowledge base – including source metadata, chunk scores and an optional AI answer in a single request. Integrates with your own apps, CRM systems, agent workflows (n8n, Make.com, Zapier, Flowise, MCP servers).

Sharing & collaboration

One team,
one knowledge base.

Knowledge only gains its value when several people can use it. anymize therefore treats knowledge bases like documents in a modern office stack: private by default, shareable at the press of a button.

Three sharing modes

Private

Only you. Default on creation.

Team pool

All members of your Business or Enterprise account. Ideal for firm wikis, policies, client archives inside a law firm.

Person-to-person
Google-Docs principle

Invite individual anymize users, even from other organizations. Ideal for joint ventures, external experts, collaboration between firm and client.

Roles per knowledge base

Viewer

Can switch the database on in chat and generate answers from it, but cannot change anything.

Editor

Can add and remove documents and trigger re-indexing.

Requirement: an anymize account

For person-to-person sharing, both sides must be anymize users. That is deliberate: the contracts (DPA, BRAO, TOM) apply only to accounts in our system. External access without a contract would contradict the compliance foundation our customers expect from us.

Anonymization in knowledge bases

You decide
per upload.

When uploading each document, you choose how it should be stored:

Default for sensitive data

Upload anonymized

The document is automatically anonymized before indexing – over 40 categories of personal and business-sensitive data are replaced with placeholders (names, addresses, IBANs, case numbers, etc.). Chunks and embeddings in the database contain only placeholder versions. When the AI later cites this database, you get the answer back with your original data thanks to bidirectional anonymization.

Consequence: Even when an international frontier model (GPT, Claude, Gemini) is used for the answer, it only ever sees placeholders. No personal data leaves the anymize platform.

For uncritical content

Upload non-anonymized

Internal handbooks without any personal reference, public studies, product documentation, company policies – for content of this kind, you can skip anonymization. That saves processing time and avoids unnecessary placeholders in contexts where they would distort meaning.

You decide per upload, not per database. Inside the same database, some documents may be anonymized while others are not – depending on content.

With our own models

The full
EU-only workflow.

Knowledge bases work with every model in anymize – international frontier models (GPT, Claude, Gemini, Mistral, Perplexity, Kimi) as well as our own models anymize Waterfall and anymize Fountain.

For the most sensitive scenarios, this produces a setup that has been hard to find in Europe until now:

01

Data in the EU.

02

Retrieval in the EU.

03

Model in the EU.

04

Answer in the EU.

When you combine your knowledge base with Waterfall or Fountain, no byte of your data leaves the EU. No anonymization needed, because the models run with us anyway. No third-country transfer, no additional DPA, no compliance grey zone. For professional-secrecy holders, sensitive industries and high-security compliance, this is the hardest standard available.

Use cases

Where knowledge bases
make the difference.

Six prototypical use cases – drawn from the real working contexts of our customers:

Law firm
Content of the database

Client files, contract templates, case law

What you do with it

„How did we argue this the last time …?“ — instantly with citations from your own briefs.

Strategy consulting
Content of the database

Earlier due-diligence reports, market studies, interview transcripts

What you do with it

Pattern analyses across multiple client projects without rereading every single report.

Clinic / medical centre
Content of the database

Treatment guidelines, internal standards, specialist publications

What you do with it

„What is our standard protocol for …?“ — answers with a pointer to the internal SOP.

Financial services
Content of the database

Internal research reports, compliance rules, regulatory updates

What you do with it

Same-day assessments that draw on your entire internal knowledge.

HR department
Content of the database

Employment contracts, works agreements, policies

What you do with it

„What does our works agreement say about home office?“ — answer with the exact clause.

Engineering team
Content of the database

API documentation, internal standards, post-mortems

What you do with it

Code reviews grounded in your own conventions; debugging informed by past failure patterns.

The pattern: Everywhere the model's general knowledge isn't enough, because the answer is your specific context – client history, company standards, internal processes. That is exactly what knowledge bases deliver.

What you should know about knowledge bases.

Frequently asked questions

A collection of your documents, prepared for AI use. Technically it is based on Retrieval-Augmented Generation (RAG): documents are split into chunks, stored as embeddings in a vector database and retrieved selectively when you ask a question. The AI receives only the relevant passages as context – not the entire archive. This saves cost, improves answer quality and turns your company knowledge into a second source for every conversation.

Start now.
14 days free trial.

All models. All features. No credit card.

We stand behind anymize. And we know – when an AI tool touches client, patient or employee data, a demo video isn't enough. That's why we give you 14 days of full access – all models, all features, no credit card. Enough time to be certain, before you trust us.

Your AI workplace awaits.