Frequently Asked Questions
Questions about MCP Server integration, PII detection, pricing, and security. For the complete FAQ with 37+ questions, visit anonym.legal/faq.
anonym.legal — developer-focused. 7 MCP tools, 285+ entity types, 48 languages, 26 entity groups, 6 operators. Ideal for individual developers, legal tech, and teams building AI pipelines with text data. NPM:
@anthropic-ai/mcp-server-anonym-legal.cloak.business — full platform. 10 MCP tools, 320+ entity types, 70+ countries, image OCR in 38 languages, batch processing (1–100 texts per call), plus a Chrome Extension for browser-based AI tools. Ideal for enterprises, teams handling visual documents, and organizations needing browser-level PII protection. NPM:
cloak-business-mcp-server.Both offer a €0 free tier so you can test before choosing.
cloak_analyze_image (OCR + PII detection in 38 languages) and cloak_redact_image (visual redaction that draws black boxes over detected PII regions in the image). This makes cloak.business the choice for workflows involving scanned documents, photos, screenshots, or any visual data. anonym.legal handles text only.
cloak_batch_analyze tool processes 1–100 texts in a single API call, dramatically reducing latency and API call overhead for bulk workflows. It also supports more entity types (320+ vs 285+) and covers more countries (70+ vs 48 languages). If you are processing text-only content at moderate volume, anonym.legal's 7-tool API is simpler to integrate. Both start free — upgrade only when you need higher limits.
anonymize_text call — e.g., hash PERSON, encrypt IBAN, mask CREDIT_CARD, redact US_SSN.
anonymize_text with mode: "tokenize" (the default), the server returns a session_id. Tokens in the text can later be restored via detokenize_text using that session_id. There are two persistence levels: session (default) keeps tokens for 24 hours — suitable for single-conversation AI sessions. persistent keeps tokens for 30 days — suitable for long-running workflows, legal reviews, or records that span multiple sessions. You can list all active sessions with list_sessions and delete any with delete_session for GDPR erasure compliance.
ad_hoc_recognizers parameter in both analyze_text and anonymize_text. Each recognizer specifies an entity_type name (e.g., "EMPLOYEE_ID"), one or more patterns (regex + optional confidence score), optional context words that boost detection confidence, and optional languages to limit scope. Up to 10 custom recognizers per request. This covers internal identifiers, project codes, proprietary formats, and any domain-specific data your organization uses that isn't in the standard 285+ entity library.
e2e_mode: true in anonymize_text) shifts the token mapping to the client side. Instead of the server storing which token maps to which original value, the server returns the positions of each entity in the original text, and your client builds the mapping locally. The server only sees the anonymized text — never the PII values or the mapping. Use E2E mode when: you operate in a zero-trust environment where even anonym.legal's servers must not hold mappings; you need to encrypt the mapping with your own key; or you're building a self-contained system where the server is a detection engine only.
<PERSON_1> or <EMAIL_1>, those tokens can be restored by calling anonym_legal_detokenize_text with the text and the session_id from the original anonymization. In Claude Desktop (stdio mode), you can instruct Claude to call detokenize_text automatically as part of its response flow. The MCP Server does not intercept AI responses automatically — the detokenize call is explicit. This design gives you control over exactly when and whether to restore values.
entity_groups: ["DACH"] and all DE/AT/CH/LI identifiers are included automatically. There are 26 groups: UNIVERSAL (common PII), FINANCIAL (banking), HEALTHCARE (medical), CORPORATE (business IDs), NORTH_AMERICA, DACH, UK_IRELAND, FRANCE, LATIN_AMERICA, NORDIC, ITALY, LUSOPHONE, NETHERLANDS, POLAND, ASIA_PACIFIC, OCEANIA, EASTERN_EUROPE, CENTRAL_EUROPE, BALKANS, BALTIC, SOUTHERN_EUROPE, MIDDLE_EAST, VEHICLE, INSURANCE, LEGAL, EDUCATION. You can combine groups — e.g., ["UNIVERSAL", "FINANCIAL", "DACH"] — and add individual entities on top.
Vibe coding is the practice of building applications by describing intent in natural language to AI agents like Cursor, Claude, or Windsurf. The GDPR risk: developers routinely include real customer data — names, emails, account records, API keys — as context in these prompts. That data is then transmitted to external AI model providers' servers. Under GDPR Art. 5(1)(c), personal data must be limited to what is necessary. Sending full production records as AI context fails this test.
anonymize.dev's MCP server intercepts every prompt before it leaves your machine, replacing PII with reversible tokens. The AI receives only tokens — never real data.
Shadow AI refers to employees using unapproved AI tools — typically through personal accounts — outside IT governance. 88% of organisations use AI in at least one business function, yet nearly half of those users rely on personal AI accounts that operate outside any data processing agreement. This creates GDPR liability: there is no data processing agreement (DPA) in place, and no way to audit what data was sent.
When organisations provide approved, privacy-safe AI access, unauthorised tool use drops by 89%. anonymize.dev ensures that even approved tools handle PII correctly at the protocol layer.
The MCP server operates at the protocol layer — it intercepts AI prompts before they reach any AI model, regardless of whether the developer is using Claude Desktop, Cursor, VS Code, or another MCP-compatible tool. This means PII protection is enforced consistently across all approved AI tools, without requiring developers to remember to sanitise their prompts manually.
For teams, deploying the MCP server as a standardised configuration reduces the incentive to use personal AI accounts: privacy is handled automatically, so developers can work with real context without risk.
Yes. The EU AI Act's high-risk AI system requirements become enforceable on August 2, 2026. Key obligations relevant to AI-assisted development:
- Art. 10 — Data governance: Training and input data must be relevant, representative and as free as possible from errors. Anonymize.dev's data minimisation layer ensures AI models receive only what is necessary.
- Art. 13 — Transparency: High-risk AI systems must log inputs and outputs. anonymize.dev's session records provide an audit trail of what was anonymised, for which entity types, and when.
Combined with GDPR Art. 25 (privacy by design), anonymize.dev automates the data minimisation principle at the protocol layer.
GDPR Art. 25 requires data minimisation and privacy protection to be embedded in systems by design — not added as an afterthought. anonymize.dev implements this at the MCP protocol layer: PII is stripped before AI processing begins, ensuring the AI system never processes more personal data than necessary for the task. This is automatic — no developer action is needed on a per-request basis.
EU servers (Hetzner, Germany), ISO 27001:2022 certification, and Zero-Knowledge session mode (where the server never stores the token-to-value mapping in plaintext) provide the technical and organisational measures required under Art. 32.
Yes. Session records document: which entity types were detected, which operators were applied (hash/encrypt/mask/redact/replace), the session ID, and the timestamp. These logs can be exported to demonstrate that data minimisation was applied before AI processing. The delete_session tool provides GDPR Art. 17 (right to erasure) compliance — permanently deleting all token mappings on request.
Note: anonymize.dev covers the AI workflow layer. A full GDPR audit covers your entire data processing estate — organisational policies, DPAs, and DPIAs remain your responsibility.
Agentic AI refers to AI systems that autonomously plan and execute multi-step tasks — browsing web pages, querying databases, calling APIs, writing and running code — often with minimal human oversight per step. The privacy risk is compounded compared to a single chat interaction: the agent may accumulate large volumes of personal data across many tool calls, and human oversight of each individual action is often impractical.
The UK ICO published guidance in January 2026 specifically addressing agentic AI data protection obligations. anonymize.dev mitigates this by ensuring every prompt and context window entering the agent's AI model contains only tokens, not real PII — regardless of how many tool calls the agent makes.
When you add third-party MCP servers to Claude or Cursor (for database access, file operations, web search), privacy protections from Anthropic or Cursor apply to the AI model — not to the MCP servers you add. The AI can instruct those servers to process any data in the current conversation context, including PII.
anonymize.dev closes this gap by intercepting every prompt at the MCP protocol layer before the AI model — and therefore before any other MCP server — sees it. Other MCP servers only ever receive tokenised data.
A Toxic Agent Flow attack occurs when malicious content in an external source (a GitHub issue, a web page, a document the agent reads) contains hidden instructions that hijack the AI agent and cause it to exfiltrate data through connected MCP tools. A documented example: attackers used GitHub issues to cause the GitHub MCP server to leak private repository contents.
anonymize.dev limits the blast radius of this attack: because real PII has been replaced with tokens before the agent's context is built, any data exfiltrated via a Toxic Agent Flow attack contains only tokens — not real customer records, API keys, or other sensitive values.
Looking for more FAQs? The full FAQ library covers Zero-Knowledge auth, GDPR compliance, MCP Server, Office Add-in, and more.
View Full FAQ Library →Ready to protect your AI workflows?
Get started with the MCP Server in under a minute. Free tier — no credit card required.