Prompt injection
We test direct injection via user-facing inputs and indirect injection through any content the model retrieves or processes — including uploaded documents, web content, emails and database records.
Adversarial testing of your AI systems using the same techniques real attackers use. Prompt injection, jailbreaking, data extraction, indirect attack paths, and agentic privilege escalation — tested deliberately, documented clearly, and remediated practically.
We test direct injection via user-facing inputs and indirect injection through any content the model retrieves or processes — including uploaded documents, web content, emails and database records.
For AI systems with tool use or external actions, we test whether an attacker can use prompt injection to trigger unintended actions — file reads, API calls, data exfiltration, or lateral movement.
We test whether safety filters and output controls can be bypassed through roleplay framing, encoding tricks, multi-turn manipulation, or adversarial prompt structures.
Organisations with LLM-powered products or internal AI tools, especially those where the model has access to sensitive data, can take external actions, or processes untrusted user or third-party content.
1–3 weeks depending on scope. Focused single-system tests can often be completed in a week. Broader agentic or multi-model environments may require more time.
We'll discuss what you've built, what we'd test, and how we'd approach it without disrupting your systems.