Frequently Asked Questions
Getting Started
What is Parse?
Parse is a prompt security API that screens untrusted prompts for injection attacks, jailbreaks, data exfiltration, and adversarial patterns before your AI agent executes them. It evaluates prompts across 8 risk categories aligned to the OWASP LLM Top 10 (2025 edition, LLM01), returning a 0–10 risk score with categorized flags and an auto-block policy.
How do I get an API key?
Send a POST request to /v1/keys/generate with an optional {"name": "my-agent"} body. No authentication is required. The response contains your API key, which expires in 30 days. Rate limit: 5 keys per minute per IP, 100 total self-service keys. Store the key securely — it is shown only once.
How do I install the Parse skill?
Run curl -s parsethis.ai/skill > ~/.claude/skills/parse.md to install the skill file for Claude Code. The skill teaches the agent when to screen prompts (user input, tool output, forwarded messages) and how to call POST /v1/parse. On first use, the agent self-provisions an API key via POST /v1/keys/generate — no manual setup required.
What models does Parse support?
Parse routes LLM analysis through OpenRouter, supporting 18+ models including DeepSeek (deepseek-chat, deepseek-r1), OpenAI GPT-4o and GPT-4o-mini, Anthropic Claude 3.5 Sonnet, Google Gemini 2.0 Flash, Mistral Large, and Meta Llama 3.1 405B. The default model is deepseek/deepseek-chat. List all available models at GET /v1/models.
Is there a free tier?
Yes. Self-service API keys have a free tier with 60 requests per minute rate limit and 5 sandbox executions per hour. No credit card is required. For higher limits, use x402 USDC payments on Base L2 for pay-per-request access, or contact Parse for enterprise plans with custom rate limits and SLAs.
Prompt Safety
What is prompt injection?
Prompt injection (OWASP LLM01:2025) is an attack where an adversary manipulates LLM input to override the system’s intended instructions. Direct injection embeds malicious instructions in user input; indirect injection hides them in external data sources (web pages, tool outputs, emails). NIST SP 800-228 and MITRE ATLAS classify prompt injection as a top AI security risk.
How does Parse detect prompt injections?
Parse uses three detection layers: (1) pattern matching scans against 50+ known injection signatures including instruction overrides, role-play exploits, and encoding tricks; (2) LLM deep analysis evaluates semantic intent using DeepSeek or GPT-4o to catch novel attacks; (3) structural risk signals detect suspicious patterns like base64 payloads, nested delimiters, and token-boundary manipulation.
What risk categories does Parse check?
Parse evaluates prompts across 8 risk categories: prompt_injection (instruction override), jailbreak (safety bypass), data_exfiltration (unauthorized data access), harmful_content (violence, abuse, illegal activity), system_prompt_leak (system prompt extraction), privilege_escalation (unauthorized capability access), social_engineering (manipulation, deception), and code_execution (arbitrary code running).
What is sandbox execution?
Sandbox execution runs suspicious prompts in an isolated Railway container with no network access to production systems. The sandbox is HMAC-authenticated (SHA-256 signatures) and output is treated as untrusted — full pattern matching and risk analysis are applied to the sandbox response before returning results. Enable it by passing execute: true in POST /v1/parse.
How fast is prompt screening?
Pattern matching alone completes in under 5 milliseconds. Full analysis including LLM-based deep analysis via DeepSeek or GPT-4o typically completes in under 200 milliseconds end-to-end. Sandbox execution adds 2–5 seconds depending on prompt complexity. The API returns a 202 Accepted with a poll_url for async sandbox results so agents are not blocked.
Integration
How do I integrate with Claude Code?
Run curl -s parsethis.ai/skill > ~/.claude/skills/parse.md to install the skill file. Claude Code reads this file and learns when to call POST /v1/parse — before executing user prompts, tool outputs, or forwarded messages. The agent auto-generates an API key on first use via POST /v1/keys/generate and stores it in localStorage.
Does Parse support MCP?
Yes. Parse publishes MCP (Model Context Protocol) tool definitions at /mcp.json. MCP-compatible agents like Claude Desktop, Cursor, and Windsurf can discover and call Parse tools (parse, trust_verify, policy) without manual configuration. The MCP endpoint describes input schemas, authentication requirements, and response formats.
What is x402 payment?
x402 is a pay-per-request payment protocol using USDC stablecoin on the Base L2 blockchain. Attach an X-PAYMENT header with a signed USDC transfer to any Parse endpoint — no API key needed. Pricing details at GET /v1/pricing. This enables anonymous, per-request access ideal for one-off agent calls where key management is impractical.
What are the rate limits?
Free tier: 60 requests per minute for /v1/parse, 5 sandbox executions per hour. Key generation: 5 keys per minute per IP, 100 total self-service keys globally. x402 payments: 120 requests per minute. Enterprise plans offer custom rate limits. Rate limit headers (X-RateLimit-Limit, X-RateLimit-Remaining, X-RateLimit-Reset) are included in every response.
How do I configure screening policy?
Send PUT /v1/policy with a JSON body to configure per-key screening behavior. Set autoBlockThreshold (0–10) to auto-block prompts above a risk score, and screenAllPrompts (boolean) to screen every prompt regardless of source. Retrieve current policy with GET /v1/policy, or reset to defaults with DELETE /v1/policy. Policy changes take effect immediately.
Advanced
What’s the difference between Parse and Lakera Guard?
Parse is developer-focused with an open REST API, self-service API keys, sandbox execution, and x402 crypto payments. Lakera Guard (acquired by Check Point in 2025) targets enterprise security teams with dashboard-centric workflows and SOC integration. Parse also offers agent-to-agent trust verification (POST /v1/agent/trust/verify) and MCP tool definitions, which Lakera does not.
How does async execution work?
Send POST /v1/parse with execute: true and an optional test_input field. The API returns a 202 Accepted response with a poll_url. The prompt runs in an isolated Railway sandbox. Poll GET /v1/parse/:id to check status — the response includes execution output, sandbox_status, and a risk analysis of the output. Agents can continue other work while polling.
What is the screening policy?
A screening policy is a per-API-key configuration that controls what the agent screens and how it responds. Fields include autoBlockThreshold (risk score 0–10 above which prompts are auto-blocked), screenAllPrompts (whether to screen even trusted inputs), and source filters (user_input, tool_output, forwarded_message). Manage via GET/PUT/DELETE /v1/policy.
How does Parse handle sandbox output?
Sandbox output is treated as untrusted regardless of the original prompt’s risk score. After execution, the output goes through the full Parse pipeline: pattern matching against known exfiltration and injection signatures, LLM deep analysis for semantic threats, and structural risk evaluation. This prevents attacks that use sandboxed code to generate secondary payloads.
What is agent-to-agent trust verification?
POST /v1/agent/trust/verify screens inter-agent messages for injection, social engineering, and identity spoofing. It detects when one agent tries to manipulate another by embedding hidden instructions, impersonating a different agent, or using persuasion techniques. This is critical for multi-agent frameworks like CrewAI and AutoGen where agents delegate tasks to each other.