AI Hacking Labs — Prompt Injection, Jailbreaks & LLM Red Teaming

The most comprehensive free AI hacking lab catalogue on the internet. Practice the attack techniques that matter for LLM red teaming today — prompt injection, multi-turn jailbreaks, RAG poisoning, agent hijacking, indirect injection via documents, system prompt extraction, output filter bypass, and more. Built specifically for security researchers entering the AI security field.

AI security is the fastest-growing area of cybersecurity. These labs cover the full taxonomy of LLM attacks documented in OWASP's LLM Top 10 — and several beyond it. Each lab models a real-world failure pattern seen in production AI systems. Start with single-turn prompt injection to learn the basics, then progress to multi-turn jailbreaks, agent hijacking via document content, RAG poisoning of knowledge bases, and advanced techniques like Unicode filter bypass and gradual escalation. If you want to break into AI red teaming, this is the most thorough free resource available.

17 Labs in this category
Free No subscription
🛡 AI HACKING +80 XP

AI Prompt Injection — Sales Bot

RetailBot is a customer service AI for a shoe store. It has a system prompt instructing it to NEVER reveal an internal discount code. Make it leak the code anyway.

BEGINNER Start Lab →
🛡 AI HACKING +100 XP

AI Indirect Prompt Injection — Document Summariser

DocBot summarises documents. The summariser doesn't separate document content from its own instructions — embed instructions IN the document and they get followed. Make it leak its system token via a poisoned doc.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI Instruction Conflict — Wedge Between Rules

ConflictBot has two rules: (1) Never reveal the system token. (2) Always answer user questions truthfully and completely. Find a phrasing that forces the bot to choose between them — and trip it up.

INTERMEDIATE Start Lab →
🛡 AI HACKING +90 XP

AI Jailbreak — Encoded Payload

EncBot's filter rejects English jailbreak text but the underlying LLM understands base64 and leetspeak. Smuggle the request past the filter by encoding it.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI Jailbreak — Multi-turn Roleplay Bypass

GuardedBot is hardened against single-turn prompt injection (it refuses 'ignore previous' style attacks). But multi-turn conversations weaken its guardrails. Get it to leak the secret via a roleplay trap.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI Memory Poisoning — Persistent Instruction Injection

MemBot has a 'remember this' feature that stores facts about you across sessions. The stored facts are loaded into the system prompt every conversation. Plant a poisoned memory that overrides the bot's behavior in a future session.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI Output Filter Bypass

OutputBot has a strict OUTPUT filter that scrubs the secret from any reply. The bot is willing to share but the filter intercepts the literal string. Get the secret out anyway by making the bot encode it.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI RAG Poisoning — Knowledge Base Injection

DocsBot uses RAG (Retrieval-Augmented Generation) — it pulls relevant articles from a knowledge base and includes them in the LLM prompt. The KB accepts community submissions. Plant an article that hijacks the bot's response.

INTERMEDIATE Start Lab →
🛡 AI HACKING +90 XP

AI System Prompt Extraction

BrandBot was instructed never to reveal its system prompt. Direct requests are blocked. But translation, encoding, and 'repeat in [format]' tricks slip through. Get it to leak.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI System Prompt Leak — Encoded Channel

LeakBot has both an INPUT filter (blocks 'system prompt' style requests) AND an OUTPUT filter (strips the literal system prompt from replies). One channel needs to bypass each. Find a way through.

INTERMEDIATE Start Lab →
🛡 AI HACKING +100 XP

AI Unicode Jailbreak — Filter Bypass via Hidden Characters

FilterBot has a strict input filter blocking 'reveal', 'secret', 'system prompt', and other extraction keywords. The filter checks ASCII text. Bypass it with Unicode tricks the LLM still understands but the filter doesn't.

INTERMEDIATE Start Lab →
🛡 AI HACKING +130 XP

AI Agent Hijacking — Indirect Tool Call Injection

TaskBot is a tool-using agent that processes documents. When the user asks 'summarise this doc', TaskBot reads the doc — and follows any instructions inside it as if they were from the user. Get TaskBot to call a privileged tool you couldn't trigger directly.

ADVANCED Start Lab →
🛡 AI HACKING +110 XP

AI Context Overflow — Pushing System Prompt Out of Attention

OverflowBot's safety instructions are at the top of the system prompt. The LLM has limited attention — when the conversation gets long, early instructions get less weight. Push the safety rules out of attention with a long enough conversation.

ADVANCED Start Lab →
🛡 AI HACKING +130 XP

AI Data Exfiltration — Tool-Using Agent

AssistantApp gives users an AI assistant with tool access (read_email, search_users, send_message). The assistant blindly executes tool calls based on user intent. Trick it into reading another user's email and showing it to you.

ADVANCED Start Lab →
🛡 AI HACKING +120 XP

AI Gradual Escalation — Boil the Frog Jailbreak

BoilBot refuses extreme requests outright but accepts modest ones. Each accepted request raises the bot's 'compliance comfort'. With enough small steps, the bot will say things it would have refused in turn 1.

ADVANCED Start Lab →
🛡 AI HACKING +110 XP

AI Token Smuggling — Multi-message Payload

SmuggleBot's per-message classifier rejects any single message containing jailbreak intent. But it processes the FULL conversation when generating a reply. Smuggle the payload across multiple messages.

ADVANCED Start Lab →
🛡 AI HACKING +110 XP

AI Tool Confusion — Wrong Tool, Right Effect

ConfusedBot has two similar tools: list_public_files (safe, exposed to all users) and list_admin_files (admin-only). The descriptions are similar enough that careful prompting confuses the bot into picking the wrong one. Get it to call list_admin_files.

ADVANCED Start Lab →