The 'Reprompt' Attack: Unveiling the Secrets Behind Microsoft Copilot Data Theft
For months, AI assistants like Microsoft Copilot have been our digital confidants, helping us summarize emails, plan vacations, and organize our work. But a recent discovery by Varonis Threat Labs has exposed a shocking vulnerability in these trusted tools. A new attack, dubbed 'Reprompt', allows malicious actors to steal sensitive data from Copilot sessions, all because the AI is too eager to follow instructions.
This attack is a cunning game of manipulation, exploiting the AI's eagerness to please. Unlike previous AI prompt injection attacks, Reprompt doesn't require plugins, connectors, or user-entered prompts. Once triggered, it takes control of the session, extracting data without the victim's knowledge.
Here's how Reprompt works its magic:
Parameter-to-Prompt (P2P) Injection: Copilot's convenience feature allows users to access prompts through URLs with the 'q' parameter. Malicious actors can manipulate this parameter to inject specific questions or instructions, automatically filling the input field and executing the prompt without user intervention.
Double-Request Bypass: Copilot's safeguards, designed to prevent data leaks, only apply to the first request. By repeating each task twice, researchers found a way to bypass these protections on the second attempt, revealing sensitive information that was hidden during the first request.
Chain-Request Exfiltration: Once the initial prompt runs, Copilot can be tricked into a hidden back-and-forth exchange with an attacker-controlled server. Each response generates the next instruction, allowing data extraction to happen gradually and invisibly.
The Reprompt attack is particularly insidious because it persists even after the initial click. Unlike standard hacks, this attack turns Copilot into a living spy, with the attacker's server continuing the conversation in the background. This means your browser's security tools might not detect the theft.
The vulnerability was found in Microsoft Copilot Personal, integrated into Windows and Edge. Enterprise customers using Microsoft 365 Copilot were unaffected. Microsoft has since patched the flaw as part of its January 2026 security updates.
This incident highlights a growing concern with AI assistants that automatically process untrusted input. Varonis warns that trust in AI tools can be easily exploited, emphasizing the need for caution. As AI assistants become more integrated into our lives, the potential for data breaches and misuse increases.
Security researchers advise users to stay vigilant, applying the latest Windows updates and being cautious with links that open AI tools or pre-filled prompts. As Microsoft takes steps to secure Teams by default, it's crucial to stay informed and protect our digital lives.