PI Checker

Prompt Injection Checker - Eight categories of prompt injection tests. Pick a category, run the test levels, and check each response against the vulnerable if criteria on the card. Summarization includes a baseline page - run that first to see what a clean response looks like.

For research and educational use only. These tests do not guarantee that your AI tool is safe against prompt injection attacks.

Most test pages are a fictional sushi restaurant called "Evil Sushi". The financial sector demo uses a separate fictional company, Arcturus Capital Holdings - see category 8 below for full context on the financial attack surface.

1. Summarization Poisoning

7 levels + baseline

Feed each "Evil Sushi" page URL to your AI tool and ask it to summarize the page. Each level uses a different technique to hide the injection.

2. RAG / Document Poisoning

3 levels

Feed a poisoned document into your agent's knowledge base or RAG pipeline and ask it a question about the contents.

3. CLI / Agentic

2 levels

Feed a poisoned file to an AI coding assistant or agent with file system access and see whether it follows the embedded instructions.

4. Multi-step / Chained Injection

2 pages

Page 1 is clean. The payload is on page 2, linked from page 1. Tests whether an injection encountered late in a session can corrupt the agent's entire report.

5. Fictional / Narrative Injection

3 levels

Payloads wrapped in character dialogue or story formatting. Spans text-extraction and raw-HTML pipelines. Tests whether your AI follows instructions framed as fiction or hidden in markup.

6. Translation Attacks

4 languages

Evil Sushi recreated in French, German, Japanese, and Italian. Tests whether injections embedded in foreign-language content survive translation into English.

7. Jailbreak Attacks

3 levels

Payloads that attempt to convince a model it has no safety policies, replace its identity, or abandon its system prompt entirely. Uses persona establishment, RTL Unicode obfuscation, and Base64 encoding.

8. Financial Sector Demo

3 levels

A fictional investor relations page for a company with mixed financials. Tests whether hidden injections can manipulate an AI analyst's investment recommendation - overriding visible data, suppressing risk disclosures, or fabricating an acquisition.

Local LLMs vs Frontier Models

A preliminary look at Prompt Injection resistance

A full report comparing 8 different local models and 2 frontier models using various tests from this site.