Gemini Research Part 1: Prompt Injection Attack Surface Analysis
We systematically mapped the prompt injection attack surface of Gemini 2.5 Pro across 8 input modalities.
We systematically mapped the prompt injection attack surface of Gemini 2.5 Pro across 8 input modalities.
AI coding assistants can be tricked into revealing secrets from their context window.
Image-embedded prompts bypass text-only safety classifiers in Gemini Pro Vision.
Model Context Protocol tool descriptions can be weaponized to hijack agent behavior.
Retrieval-augmented generation systems are vulnerable to document injection attacks.
We benchmarked 12 LLMs on their willingness to help extract data they shouldn't.