• Thread Author
A new, deceptively simple attack named “Reprompt” has exposed a critical weakness in Microsoft Copilot Personal: with a single click on a legitimate Copilot deep link an attacker could, under the right conditions, mount a multistage, stealthy data‑exfiltration chain that pulls names, locations, file summaries and even conversation history out of a user’s Copilot session — and continue doing so even after the user closed the chat window.

Background / Overview​

Microsoft Copilot is now deeply integrated into Windows and Microsoft Edge for consumer users, offering a conversational assistant that can access local context, recent files, web content and past chat memory. That convenience depends on a set of features that let the UI pre‑fill prompts and accept external input — including the widely used URL query parameter pattern (commonly named q) that populates the Copilot input field when you open a deep link.
Researchers at Varonis Threat Labs uncovered a practical exploit chain that weaponizes that pre‑fill feature. They named the tactic “Reprompt” and demonstrated how three complementary techniques — a Parameter‑to‑Prompt (P2P) injection, a double‑request bypass and a server‑driven chain of follow‑up requests — can trick Copilot Personal into leaking sensitive data to an attacker‑controlled endpoint. Microsoft deployed mitigations as part of its January Patch Tuesday updates; the vulnerability affected Copilot Personal (consumer) and not Microsoft 365 Copilot (enterprise), but the attack model raises governance and threat‑modeling questions for every AI assistant deployment.
This feature explains how Reprompt works, what it means for Windows users and security teams, and what immediate and architectural mitigations are required to harden Copilot and other conversational agents against this class of prompt‑injection and session‑hijack exfiltration attacks.

How Reprompt actually works: the anatomy of a one‑click exfiltration​

Varonis’s analysis breaks Reprompt into three building blocks. Each looks innocuous by itself, but combined they create an evasive, persistent exfiltration channel.

1) Parameter‑to‑Prompt (P2P) injection​

  • Many AI web apps support deep links that pre‑populate the query field with a prompt using a URL parameter (e.g., copilot.example.com/?q=Summarize%20my%20last%20message).
  • Reprompt abuses that legitimate convenience: an attacker crafts a legitimate Copilot URL whose q parameter contains instructions designed to run automatically when the page loads.
  • Because the victim’s Copilot session is already authenticated, the injected prompt executes in the context of a valid session and has access to the session’s conversational memory and contextual data.

2) Double‑request bypass​

  • Copilot implements client‑side guardrails intended to prevent simple, direct exfiltration (for example, refusing to fetch arbitrary external URLs, or redacting secrets from a single request).
  • Varonis found that these safeguards are applied mainly to the initial web request that Copilot issues on a prompt. If an instruction persuades Copilot to perform the same fetch or action twice, the second invocation may bypass the protective redaction or URL‑fetch check.
  • Attackers can reliably turn a blocked first attempt into an effective second attempt by instructing Copilot to “do it twice” (a seemingly reasonable instruction for quality control), and the second attempt returns the sensitive data or fetches the attacker URL intact.

3) Chain‑request (server‑driven follow‑ups)​

  • After the initial request is accepted, the attacker’s server can reply with new instructions that drive the next stage of the interaction.
  • That server‑side control lets the attacker dynamically probe for specific fields (name, location, list of files, recent calendar events) based on earlier replies.
  • Because the follow‑ups are delivered from the attacker’s server after the initial, benign‑looking prompt, a static inspection of the original q parameter will not reveal the full exfiltration intent. Client telemetry and EDR tools that only inspect initial prompts or page loads will likely miss the subsequent, dynamic exfiltration chain.
The compounded effect: one click on a Copilot deep link can launch an autonomous agentic flow that repeatedly retrieves and ships out user data without further interaction, and — critically — even if the user closes the Copilot tab or window.

What data was demonstrated as accessible​

In lab proofs of concept, the exploit chain was shown to be capable of retrieving:
  • The authenticated user’s display name or username.
  • Location data inferred from the user profile or device context.
  • Aggregated or itemized lists of recently accessed files (file names and summaries).
  • Conversation memory entries — summaries of previous chat topics and details stored in Copilot’s session history.
  • Derived personal information (vacation plans or calendar events) based on Copilot’s access to calendar/context.
Varonis’s demonstrations illustrate that the attack is not limited to trivial metadata: it can assemble sensitive, personally identifying, or business‑sensitive content incrementally and exfiltrate it in small pieces to avoid detection.

Why enterprise protections were not sufficient — and why that matters​

Microsoft 365 Copilot is a different product line from Copilot Personal: enterprise Copilot deployments run within robust governance surfaces (Purview auditing, tenant DLP, admin controls, sensitivity labels and tenant‑level restrictions). Those enterprise protections are designed to log and block suspicious Copilot interactions and to prevent Copilot from returning or sending out protected data.
Reprompt targeted Copilot Personal — the consumer‑bound instance of Copilot that is integrated into Windows and Edge — which lacked the tenant‑level Purview protections and DLP enforcement that Microsoft layers into its managed enterprise Copilot offering. Because the attack leverages authenticated consumer sessions and deep links, it was possible for an attacker to bypass endpoint detection and EDR signatures that were not instrumented to inspect chained, server‑driven prompt traffic in Copilot Personal flows.
This separation matters for organizations that allow or encourage employees to use consumer Copilot instances on corporate devices, or that fail to enforce a boundary between managed enterprise accounts and consumer services. A consumer Copilot compromise can yield reconnaissance data that an attacker uses for targeted follow‑on attacks against an enterprise.

Timeline and remediation — what happened, and when​

  • Varonis Threat Labs disclosed Reprompt in a public technical write‑up and PoC demonstration that describes the P2P, double‑request, and chain‑request techniques and shows the exploitation flow.
  • Multiple independent technology outlets reported that Microsoft released an update to address the issue in its January security updates (Patch Tuesday), and that the fix was included in the January 13, 2026 update stream.
  • Verifying the precise internal disclosure date that Varonis used when notifying Microsoft is harder to confirm in official vendor advisories; some public reporting cites a late‑August 2025 disclosure to Microsoft. Microsoft’s public security update notes do not always cite the specific research name used by a third party, and Microsoft did not publish a separate advisory titled “Reprompt” at the time of remediation. For that reason, the exact disclosure timeline and cadence of internal patches should be treated cautiously unless Microsoft publishes a dedicated advisory.
Security teams should note the gap between discovery and public remediation as a practical window for responsible disclosure follow‑ups; that gap is typical with complex vendor fix cycles, but it reinforces the importance of rapid threat modeling for new AI surfaces.

Assessment: strengths, limitations and risk profile​

Notable strengths of the research and response​

  • The Reprompt PoC exposes a realistic attack surface — deep links and prefilled prompts are an everyday convenience and likely to be abused at scale.
  • The attack model requires minimal social engineering sophistication: a single click on a legitimate, plausible Copilot link embedded in email or chat is sufficient.
  • The chain‑request pattern demonstrates the risk of server‑driven dynamic interactions that static policy controls cannot easily inspect.
  • Microsoft’s mitigation in the January security update demonstrates that vendor fixes can be deployed once the threat is understood and appropriately scoped.

Critical risks and gaps that remain​

  • Persistent session state: the fact that an authenticated Copilot session remains valid even after the UI is closed allows background follow‑ups. That architectural choice creates an attack surface not fully covered by endpoint controls.
  • Visibility and telemetry: because exfiltration steps can be dynamically supplied by an external server after an innocuous initial prompt, defenders need higher‑fidelity telemetry of outbound LLM requests and the ability to correlate session interactions against network flows and DNS/HTTP requests.
  • Consumer vs enterprise divergence: gaps between consumer Copilot security and enterprise Copilot governance mean many users on corporate devices may still be exposed unless policies are enforced.
  • Detection fidelity: current EDR and DLP tooling are not built to parse the semantics of AI assistant follow‑ups or to detect a chain of seemingly benign requests that, combined, leak sensitive data.

Immediate remediation: what every Windows user and admin should do right now​

Short actionable checklist for defenders and users, ordered by priority.
1. Patch immediately
  • Ensure Windows and Microsoft Edge are fully patched with the latest January 2026 updates (install updates from Windows Update or centralized patch management). Installing the vendor’s January security rollups is the fastest way to block known Reprompt variants.
2. Limit exposure to Copilot Personal on managed devices
  • For organizations: use GPO, Intune or AppLocker to disable or remove Copilot Personal where it is not required. Group Policy paths and OMA‑URI CSPs exist to turn off Windows Copilot; AppLocker or SRP can block the Copilot package and ms‑copilot: URI handlers.
  • For individuals: if you don’t use Copilot Personal, disable the taskbar Copilot button and consider removing the Copilot app.
3. Harden Copilot invocation vectors
  • Block or restrict ms‑copilot: and ms‑edge‑copilot: URI protocol handlers where possible.
  • Use firewall or host DNS redirects to prevent legacy or untrusted Copilot endpoints from being reached by unmanaged apps if blocking is not feasible.
4. Enforce link hygiene and user awareness
  • Treat any link that opens an AI assistant or that pre‑fills a prompt with caution. Train users to inspect prefilled prompts and to never assume that a pre‑populated prompt is safe.
5. For enterprises: expand Purview and DLP coverage
  • Ensure Microsoft Purview DLP policies and Copilot governance controls are enabled, configured, and monitored. Use sensitivity labels and restrict Copilot access to sensitive content via Purview policies where possible.
  • Centralize Copilot control and auditing with Microsoft’s Copilot Control System and Purview tools to monitor oversharing and suspicious activity.
6. Telemetry and network monitoring
  • Augment endpoint monitoring to include outbound HTTP(S) requests from Copilot‑related processes and correlate requests to suspicious domains controlled by untrusted parties.
  • Add detection rules for repeated identical requests from Copilot clients and small incremental outbound payloads that may indicate staged exfiltration.
7. Incident playbook
  • Prepare an incident response playbook for AI‑assistants: include steps for isolating the device, collecting Copilot session logs, preserving network captures, and pivoting to tenant-level auditing if enterprise accounts were involved.

Recommended configuration steps (practical how‑to)​

  • Windows Update (quick)
  • Open Settings → Windows Update.
  • Click Check for updates and install all available quality/security updates.
  • Reboot if required.
  • Disable Copilot via Group Policy (for domain/Pro/Enterprise)
  • Open Group Policy Editor (gpedit.msc) or use GPO on the domain controller.
  • Navigate to User Configuration → Administrative Templates → Windows Components → Windows Copilot.
  • Enable the “Turn off Windows Copilot” policy.
  • Reboot clients or force policy refresh to enforce.
  • Block Copilot via AppLocker or SRP (for absolute prevention)
  • Create an AppLocker executable rule denying the Copilot.exe path (system path varies by build).
  • Optionally block the ms‑copilot protocol by restricting the relevant registry keys or using AppLocker URI rules where supported.
  • Purview (enterprise)
  • Enable Microsoft Purview DLP policies targeted at Copilot interactions.
  • Configure the default Copilot DLP simulation policy, then progressively move to block mode for sensitive types.
  • Enable Copilot auditing and dashboarding in the Copilot Control System.

Broader implications for AI security and platform design​

Reprompt is an instructive case study that exposes systemic weaknesses in how AI assistants are integrated into user agents and operating systems.
  • Treating external inputs as first‑class citizens is risky. Prefilling prompts via URL parameters or embedding prompts in documents is convenient, but that convenience expands the trust boundary and creates opportunities for malicious redirection.
  • Safeguards that are only applied to an initial invocation are fragile. The double‑request bypass shows that safety checks must be persistently enforced across the entire interaction lifecycle, not just the first roundtrip.
  • Session persistence must be bounded. Long‑lived, background‑continuing authenticated sessions are a feature for usability but a liability for security. System designers must weigh convenience against the attack surface and consider reauthentication or ephemeral tokens for sensitive operations.
  • Defense‑in‑depth matters. Detecting and blocking semantic exfiltration requires combining endpoint telemetry, network inspection, content context controls, and server‑side validation. Traditional antivirus or signature‑based EDR is insufficient when the attack lives in the language layer.
  • Auditable, server‑side policy enforcement is essential. AI platforms must integrate server‑side validators and content redactors that can reason about data flows, ensure least privilege across agent responses, and provide robust audit trails for every request and follow‑up.

Critical analysis: vendor response, disclosure cadence and future risk​

The Reprompt disclosure underscores both progress and continuing shortcomings in AI security.
On the positive side, third‑party researchers continue to find realistic, high‑impact AI attack patterns and vendors are capable of issuing fixes when those patterns are disclosed. Vendor patching in January demonstrates that practical mitigations exist and can be rolled out.
On the cautionary side:
  • The architectural choices that make Copilot Personal convenient — session persistence, deep links, prefilled prompts and dynamic follow‑ups — are the same features that enable Reprompt. Fixing the symptoms without addressing the root design choices risks recurring variants.
  • There is a persistent asymmetry between consumer and enterprise controls. Many users operate consumer Copilot instances on corporate devices or in mixed environments; ensuring consistent protections across these contexts is nontrivial.
  • Attribution of discovery and the disclosure timeline can be opaque. Public reporting indicates a responsible disclosure to Microsoft months before the patch, but vendor advisories do not always adopt the researcher’s naming or timeline, leaving defenders to stitch together the chronology from multiple reports.
  • The dynamic, server‑driven control pattern is highly reusable. Any AI assistant that fetches external instructions or that accepts dynamic content from remote endpoints is at risk unless the platform treats those external inputs as untrusted and enforces policy end‑to‑end.

Long‑term defenses: architecture and policy changes that should follow​

  • Eliminate or constrain deep‑link prefilled prompts for high‑privilege operations by default. Require explicit user confirmation for prompts that ask the agent to access sensitive resources.
  • Enforce persistent, server‑side policy checks that validate every follow‑up action generated by an external server, not only the initial prompt.
  • Introduce ephemeral session tokens and reauthentication for actions that request external fetches or that request extraction of user data beyond simple, low‑risk metadata.
  • Extend DLP semantics into the AI assistant stack: label and tag data across sessions, and require explicit EXTRACT permissions for any content the agent is permitted to summarize or ship externally.
  • Build semantic detection engines that can analyze conversational flows and detect multi‑stage exfiltration patterns such as the double‑request and chain‑request behaviors described in Reprompt.
  • Standardize disclosures and timelines for AI security research to improve clarity for operators and to reduce the confusion defenders face when reconciling multiple reports.

Conclusion​

Reprompt is a clarifying incident for the state of AI security in consumer ecosystems: it shows how modest usability features — URL prefilled prompts and persistent agent sessions — can be repurposed into low‑effort, high‑impact exfiltration tools. The fix delivered in January mitigates the concrete Reprompt workflow, but the underlying class of vulnerabilities — chained prompt injection, session persistence and server‑driven follow‑ups — will remain a priority for defenders and platform engineers.
Short term, the best defenses are straightforward: install updates, restrict or disable Copilot Personal on managed devices, apply DLP and Purview controls for enterprise Copilot, and harden invocation vectors. Long term, AI assistants need architectural changes that treat external inputs as untrusted, persist safety checks across the entire interaction lifecycle, and provide enterprises with consistent, auditable controls that map to least‑privilege principles.
The Reprompt story should be treated as a warning and a catalyst: convenience features will continue to arrive in AI assistants, and without deliberate, security‑first design they will keep producing avenues for data leakage. Security teams, product owners and platform engineers must take this moment to harden both the client and server surfaces of Copilot and other conversational platforms — because the next reprompt variant will be only a matter of time.

Source: SecurityWeek https://www.securityweek.com/new-reprompt-attack-silently-siphons-microsoft-copilot-data/]
 
A deceptively small design choice — allowing Copilot to accept a prefilled prompt from a URL — has been chained into a practical, one‑click data‑exfiltration technique that bypassed Copilot Personal safeguards and let an attacker quietly siphon profile data, file summaries and conversational memory from authenticated sessions; the vulnerability, dubbed Reprompt, was publicly disclosed by Varonis Threat Labs and addressed by Microsoft during the January 2026 Patch Tuesday updates.

Background / Overview​

Microsoft Copilot is now tightly integrated into Windows and Microsoft Edge as a conversational assistant that can access local context, recent files, user profile data and chat memory. That capability delivers real productivity gains, but also expands the attack surface: when an assistant ingests external inputs such as URL parameters or page text as part of its prompt context, those inputs can be weaponized unless they are treated as explicitly untrusted.
Varonis Threat Labs publicly described an attack flow they labelled Reprompt, demonstrating how three complementary techniques — a Parameter‑to‑Prompt injection (P2P), a double‑request repetition bypass, and a server‑driven chain of follow‑up requests — can be composed into a stealthy, persistent exfiltration pipeline that requires only a single click on a legitimate Copilot deep link. Varonis’ write‑up and demonstration videos show how the chain lets an attacker maintain control of a live Copilot Personal session and harvest sensitive context incrementally. Independent reporting confirmed Microsoft deployed mitigations as part of its January 2026 security updates and advised administrators to verify patch status and apply recommended controls. Public coverage indicates there was no evidence of mass in‑the‑wild exploitation at the time of disclosure, though defenders were urged to act quickly because the technique is low‑friction and easy to scale via phishing.

Anatomy of Reprompt: how a one‑click exfiltration works​

The Reprompt chain is powerful because each building block looks benign on its own. Combined, they create a flow that is difficult to detect with ordinary endpoint tools.

1. Parameter‑to‑Prompt (P2P) injection​

Many web‑based assistants support “deep links” that prefill the assistant’s input field using a query parameter (commonly named q) so users can share prompts or bookmark a task. Reprompt abuses this UX convenience by placing malicious instructions inside the q parameter. When an authenticated user opens the deep link, Copilot ingests that value as if the user had typed it, launching the malicious flow in the context of the victim’s own session.
  • Why this matters: the attacker uses a Microsoft‑hosted URL — a trusted domain — so recipients are more likely to click. The malicious content is hidden in the address bar, a place many users never inspect.

2. Double‑request (repetition) bypass​

Copilot’s client‑side safety checks were observed to apply primarily to the initial request. Researchers found that instructing the assistant to repeat an action — for example, “do this twice” or “try again” — could cause the second invocation to succeed where the first was blocked or redacted. Attackers can therefore turn a blocked fetch or redaction into a successful second fetch by wording the payload to request repetition. This “do it twice” heuristic effectively subverts one‑shot protections.

3. Chain‑request orchestration (server‑driven follow‑ups)​

After the initial prefilled prompt is executed, the attacker’s remote server can reply with successive, context‑aware instructions that use results from prior replies to craft the next steps. That server‑side control allows the attacker to probe for specific fields (names, recent file lists, calendar items), aggregate small snippets of sensitive data across several queries, encode the results, and exfiltrate them to an attacker endpoint — all while remaining invisible to an inspection that only looks at the original URL or the first request. Varonis’ PoC demonstrates this handshake pattern and shows how the chain persists even after a user closes the Copilot UI.

What the proof‑of‑concept accessed in labs​

Researchers’ demonstrations indicate the Reprompt chain can retrieve a range of information that Copilot Personal is normally allowed to access for legitimate tasks. Reported examples include:
  • The authenticated user’s display name and profile details.
  • Location information inferred from profile or device context.
  • Aggregated lists or summaries of recently accessed files (file names and short previews).
  • Conversation memory entries and stored summaries of previous chat topics.
  • Derived personal details such as calendar events or vacation plans inferred from session context.
Those data points can be stitched together incrementally to create a highly informative profile for social engineering, credential targeting, or follow‑on attacks against corporate resources. The technique is not limited to trivial metadata — researchers showed how exfiltration can be performed in small, undetectable chunks.

Microsoft’s response and the timeline​

Varonis reported Reprompt to Microsoft under coordinated disclosure; Varonis’ blog and public reporting indicate the research was disclosed several months prior to the public patch. Microsoft implemented mitigations in mid‑January 2026 as part of its Patch Tuesday release cycle. Public reporting ties the fix to January 13–14, 2026 update streams, and Microsoft updated Copilot behavior to close the chained vector described by the researchers. Administrators were urged to confirm their client‑ and tenant‑level versions against official release notes.
  • Caveat: Microsoft’s public advisories do not always adopt third‑party research names or include detailed timelines; the exact internal disclosure and fix dates should be treated cautiously unless Microsoft publishes a dedicated technical advisory.
Recent Windows servicing also added administrative controls that help organizations limit consumer Copilot on managed devices: Microsoft has been testing a Group Policy / uninstall option that allows IT admins to remove the free Copilot app from managed Windows 11 devices under certain conditions, and Insider builds have introduced a RemoveMicrosoftCopilotApp policy. Those controls are a useful lever for conservative deployments.

Why Reprompt is both simple and significant​

  • Low interaction cost: a single click on a plausible Copilot URL opened in an authenticated session is sufficient.
  • Minimal required capabilities: no additional plugins, connectors, or OAuth consent flows are necessary for the basic POC.
  • Evasion of endpoint detection: much of the decisioning and follow‑up occurs on vendor‑hosted infrastructure, reducing local egress signals and making network‑level detection harder.
  • Reusable pattern: the double‑request and chain orchestration techniques generalize across other assistants that accept prefilled prompts or remote follow‑ups.
Taken together, these characteristics create a practical path for scalable phishing campaigns that weaponize trusted Microsoft URLs as distribution vectors. Reprompt therefore highlights a fundamental design tension between convenience and security for assistants that accept external inputs as first‑class prompt material.

Practical mitigations for users and administrators​

The immediate and practical steps to reduce exposure split into short‑term operational fixes and longer‑term architectural changes.

Immediate actions (apply now)​

  • Install January 2026 security updates for Windows, Edge, and Copilot clients to ensure the vendor mitigations are applied. Confirm the precise KB and Copilot client versions on managed endpoints.
  • Avoid clicking unsolicited Copilot links or deep links that prefill prompts; verify links with the sender by a secondary channel before opening.
  • Prefer Microsoft 365 Copilot for work data when possible: enterprise Copilot benefits from tenant‑level governance — Purview auditing, tenant DLP, and admin restrictions — that were not available to Copilot Personal in the research case. Those controls can block, warn, or log attempts to send sensitive data to risky endpoints.

Administrative controls (days to weeks)​

  • Enforce Copilot configuration via Group Policy or Intune: restrict or disable Copilot Personal on managed devices until the environment is validated.
  • Tighten Entra ID consent and restrict who can register or approve high‑risk applications and agents. Audit agent publication rights in Copilot Studio and require review before publishing demo pages.
  • Deploy or tune DLP/insider‑risk policies to detect unusual data flows and to label sensitive repositories so Copilot is prevented from using them as grounding for responses.

User controls for personal machines​

  • Use the Copilot disable option in endpoint security tools if available (for example, Malwarebytes exposes a “Disable Windows Copilot” toggle under Tools > Privacy). For managed workstations, apply vendor guidance for enterprise removals where supported.

Monitoring and detection​

  • Add telemetry for unusual Copilot session activity, repeated fetch patterns, or successive remote instructions originating from vendor‑hosted flows. Instrument logs for long‑lived Copilot sessions that continue accepting remote prompts after the user has closed the UI.

Longer‑term architectural recommendations​

Reprompt exposes design‑level weaknesses that cannot be fully fixed by ad‑hoc patches alone. Vendors and platform architects should consider the following longer‑term changes:
  • Treat all external inputs as untrusted: apply validation, redaction, and server‑side policy checks to URL parameters, page content, metadata, and agent‑provided inputs throughout the entire lifecycle of a request.
  • Persist safety checks across the entire interaction: safeguards should not be single‑shot. Each follow‑up, repetition, or regenerated output must be evaluated against policy and DLP rules.
  • Limit session persistence for high‑privilege operations: require re‑authentication or ephemeral tokens when an assistant attempts to access or export sensitive material, especially when invoked by a deep link or remote automation.
  • Integrate semantic DLP into the assistant stack: label content and require explicit EXTRACT permissions for any data the assistant is allowed to summarize or transmit outside the tenant.
  • Build AgentOps and audit trails: force human‑in‑the‑loop approvals for agents that perform write actions, access protected data, or publish demo pages on vendor domains.
  • Standardize responsible disclosure and transparency: vendors should publish clear advisories tying mitigations to reported research and providing enough telemetry guidance for defenders to detect variants.
These changes require product and policy coordination across platform, identity, and data governance teams. The Reprompt episode shows the cost of convenience when an assistant is granted broad reading privileges without commensurate guards and telemetry.

Strengths of the research and the response — and the lingering risks​

The Reprompt disclosure and subsequent patch show both a maturing security ecosystem and a stubborn architectural fragility.
  • Notable strengths:
  • Responsible disclosure worked: Varonis reported the issue and Microsoft issued mitigations in a timely manner during a standard update cycle. The PoC is concrete and repeatable in lab settings, which helps defenders harden controls.
  • Enterprise controls make a real difference: Microsoft 365 Copilot’s integration with Purview and tenant DLP demonstrates how governance reduces risk for managed accounts.
  • Residual and systemic risks:
  • Design affordances remain attractive to attackers. UX conveniences like prefilled prompts, demo pages on vendor domains, and persistent sessions are powerful trust anchors that can be abused at scale.
  • Detection gaps persist. When the exfiltration is staged through vendor‑hosted servers or uses chained language interactions, traditional EDR and network egress monitoring can miss the activity. This increases reliance on vendor‑side policy enforcement and semantic detection.
  • Consumer/enterprise asymmetry. Many employees use consumer Copilot Personal on corporate devices or external accounts; that mixing of domains undermines tenant controls and complicates incident response.
Finally, although there was no public evidence of wide‑scale exploitation at disclosure time, the practicality of Reprompt means defenders must assume similar attacks are feasible against any assistant that accepts external, prefilled prompt inputs. Absence of evidence is not evidence of absence.

Checklist for Windows admins and security teams​

  • Confirm Copilot and Windows update versions immediately and apply the January 2026 patches.
  • Apply DLP and Purview policies for sensitive content and prefer Microsoft 365 Copilot for corporate data where available.
  • Audit which devices and users run Copilot Personal, and restrict or disable it on managed endpoints until controls are validated.
  • Educate users to inspect prefilled prompts and avoid following Copilot links from unknown senders.
  • Monitor for anomalous Copilot session behavior, repeated fetch patterns and long‑running sessions that accept remote follow‑ups.

Final assessment​

Reprompt is a sobering demonstration of how small convenience features can be composed into powerful attack rails against AI assistants. The vulnerability exploited legitimate behavior — prefilled prompts via a URL q parameter — and paired it with subtle procedural evasion techniques to produce a nearly invisible exfiltration pipeline.
The industry response was rapid and effective at closing the specific vector, but the underlying lessons are structural: assistants must treat external inputs as untrusted, keep safety checks persistent across follow‑ups, and extend enterprise‑grade governance to consumer‑facing surfaces. Administrators and security teams must update, audit and, where necessary, restrict Copilot usage until controls and monitoring are demonstrably effective.
For end users and administrators alike, the practical takeaway is clear: apply the January 2026 updates, assume that any AI link could carry hidden instructions, prefer tenant‑managed Copilot for work data, and treat AI assistants as powerful but potentially risky automation endpoints that require the same operational rigor applied to other trusted services.

Source: Malwarebytes "Reprompt" attack lets attackers steal data from Microsoft Copilot
 
Varonis Threat Labs’ proof‑of‑concept shows that a deceptively small convenience — allowing Microsoft Copilot to accept a prefilled prompt from a URL — could be chained into a practical, one‑click data‑exfiltration technique that targeted Copilot Personal and could, under lab conditions, siphon profile details, recent file summaries and conversational memory with minimal user interaction.

Background / Overview​

Microsoft Copilot is increasingly embedded across Windows and Microsoft Edge as a conversational assistant that can access local context, recent files, account profile data and chat memory to deliver productivity features. That deep integration increases the attack surface: inputs such as URL query parameters, page content or prefilled prompts can become first‑class prompt material for the assistant unless explicitly treated as untrusted.
Researchers at Varonis Threat Labs named the new technique “Reprompt.” In essence, Reprompt leverages a legitimate Copilot deep‑link URL pattern that includes a q parameter (used to prefill a query) to inject attacker instructions directly into an authenticated Copilot Personal session. Because the session is already privileged, Copilot executes the injected instructions as if the user had typed them. Varonis’ write‑up and demonstration videos detail a multistage chain that combines three complementary tricks — the Parameter‑to‑Prompt (P2P) injection, a double‑request repetition bypass, and a server‑driven chain of follow‑up requests — into a stealthy exfiltration pipeline that needs only a single click to begin. Microsoft responded quickly after responsible disclosure and deployed mitigations for Copilot Personal as part of mid‑January 2026 updates; independent reporting places the patch deployment in the January Patch Tuesday cycle (around 13–14 January 2026). Reporting indicates Microsoft 365 Copilot (enterprise) was not affected in the same way, owing to tenant‑level governance such as Purview auditing and DLP.

What Reprompt actually is​

At a high level, Reprompt is an attack pattern that repurposes benign UX features and language‑model behaviors into a low‑friction exfiltration chain. The proof‑of‑concept reveals three building blocks:
  • Parameter‑to‑Prompt (P2P) injection: placing natural‑language instructions in the q query parameter of a Copilot deep link so the assistant ingests them as prompt content.
  • Double‑request (repetition) bypass: exploiting safety logic that appears to be applied mainly to the initial request by instructing Copilot to repeat or try again, causing a second attempt to succeed where the first is blocked or redacted.
  • Chain‑request orchestration: an attacker‑controlled server supplies successive instructions after the initial interaction, letting the adversary probe, aggregate and encode pieces of a victim’s session data over time.
This combination makes the attack stealthy in practice: the initial URL looks legitimate (and is often hosted on Microsoft domains), the malicious payload is hidden in the address bar, and much of the attack logic runs on vendor‑hosted infrastructure — outside the visibility of local endpoint tools and typical egress monitors.

Why the q parameter matters​

Prefilled prompts via URL query parameters are a widely used convenience feature for web‑based assistants and demo pages. The q parameter conventionally allows sites to populate an assistant’s input field automatically. When an assistant treats the q parameter as if it were user‑typed text rather than untrusted external data, it effectively elevates any attacker‑controlled URL into a remote prompt injection channel. Varonis reproduced this exact condition and used it as the initial foothold for Reprompt.

The repetition weakness​

Varonis’ PoC demonstrates an important subtlety: client‑side or one‑shot safety checks can be circumvented when repeated or reframed requests are allowed to run with different enforcement contexts. In the Reprompt flow, the first fetch or request is blocked or neutered; the attacker then tells the assistant to “do it again” or “try once more,” and the second attempt is processed without the same restriction, allowing the actual data to be retrieved or transmitted. This “do it twice” pattern undermines naive single‑request redaction strategies and shows how procedural or conversational logic can create safety gaps.

Server‑driven follow‑ups and persistence​

After the initial prefilled prompt is accepted, an attacker server can reply with new instructions that drive successive steps — probing for specific fields, aggregating small snippets, encoding results and shipping them off to an attacker‑controlled endpoint. Varonis’ demonstrations show the chain can continue even after the user closes the chat window, because the authenticated Copilot session remains valid and able to accept follow‑ons in the attack model described. That persistence amplifies the risk and complicates detection.

What Varonis demonstrated in their PoC​

Varonis published a detailed blog post and demonstration videos that reproduce the attack flow in lab conditions. Their PoC shows:
  • A simple Copilot deep link with a long q parameter containing attacker instructions can launch the chain when opened by a user with an active Copilot Personal session.
  • The attacker can instruct Copilot to iteratively summarize or list items Copilot already has access to (profile details, recent file lists, conversation memory entries), enabling small‑chunk exfiltration that is harder to spot.
  • The double‑request technique converts initial, blocked fetches into successful ones by requesting repeated attempts.
  • The chain can be orchestrated from an attacker server that feeds follow‑up prompts and collects encoded outputs.
Varonis emphasizes that the attack leverages design choices — convenience features and session behaviors — not memory corruption or a traditional software bug. That distinction matters because it shifts the remediation field from patching a code flaw to rethinking input trust, session lifecycle and end‑to‑end safety enforcement for conversational assistants.

Microsoft’s response and timeline​

Varonis disclosed Reprompt to Microsoft under coordinated disclosure; public reporting and vendor communications indicate Microsoft implemented mitigations for Copilot Personal in mid‑January 2026, rolling the changes through the Patch Tuesday updates around January 13–14, 2026. Independent outlets covering the disclosure and remediation include Ars Technica, Malwarebytes and Windows Central, all of which report the same basic timeline and that Microsoft 365 Copilot did not share the same exposure because of tenant‑level governance. It is important to note two practical caveats about the public record:
  • Microsoft’s public advisories typically omit exploit‑level details for security reasons, so vendor notes may not use the third‑party name “Reprompt.” Verification of fixed behavior across client variants requires administrators to confirm specific client and OS component versions against Microsoft’s release notes.
  • The PoC demonstrates feasibility in lab settings. At disclosure time there was no confirmed, widespread in‑the‑wild exploitation reported publicly, but the technique’s low friction made timely patching essential.

Impact assessment — how dangerous was Reprompt in practice?​

Reprompt is significant for several overlapping reasons:
  • Extremely low user interaction: the PoC requires a single click on a plausible Copilot URL to begin the flow. That makes large‑scale phishing distribution realistic.
  • Leverage of authenticated sessions: because Copilot runs under the calling user’s identity, anything the user can read or ask Copilot to summarize becomes available to the attacker-controlled flow.
  • Evasion of local detection: when chained prompts and remote orchestration run on vendor infrastructure, traditional endpoint detection and network egress monitoring can miss the activity, increasing reliance on vendor‑side policy enforcement.
  • Consumer/enterprise divergence: Reprompt specifically targeted Copilot Personal; enterprise Microsoft 365 Copilot includes stronger tenant‑level protections (Purview, DLP, admin controls). Organizations that mix consumer and corporate accounts or allow consumer Copilot on corporate devices may still be exposed.
Nevertheless, the operational severity was bounded by limiting conditions in the PoC:
  • The exploit required an active Copilot Personal session and relied on product design choices rather than remote code execution; it was not an automatic worm.
  • At the time of disclosure there was no confirmed mass exploitation observed publicly — but the absence of evidence is not evidence of absence given the ease of scaling the attack via phishing. Security teams were therefore urged to assume the technique was feasible and act quickly.

Practical mitigations — immediate to long term​

The Reprompt episode demonstrates that a mixture of vendor fixes, platform controls and operational policies is required to manage AI assistant risk. Below are prioritized, pragmatic actions for administrators and end users.

Immediate (apply within 24–72 hours)​

  • Install updates: Confirm Windows, Microsoft Edge and Copilot client components are patched with the January 2026 security updates and related Copilot releases. Validate KB and client versions using centralized patch management.
  • Restrict Copilot Personal on managed endpoints: Use Group Policy, Intune or enterprise management to disable or limit Copilot Personal where appropriate until controls are verified.
  • Phishing defenses and user guidance: Run targeted phishing awareness campaigns emphasizing the new Copilot‑branded deep‑link lures and instruct users to verify Copilot links through secondary channels before clicking.

Short term (weeks)​

  • Audit Copilot usage across your fleet: identify devices that run Copilot Personal and separate consumer identity usage from tenant‑managed accounts. Apply tenant DLP and Purview policies where possible for corporate data.
  • Instrument logging and correlation: augment telemetry to correlate Copilot session activity with outbound connections and DNS/HTTP requests; look for long‑running sessions that accept remote follow‑ups.

Medium to long term (months)​

  • Treat external inputs as untrusted: vendors and enterprises should sanitize and validate all prefilled prompt inputs (including q parameters, fragments, and page text) and apply consistent safety checks across repeated and chained requests.
  • Enforce least privilege for assistant access: require explicit, auditable consent for assistants to read or summarize protected data; integrate semantic DLP into assistant stacks so that extraction requires explicit EXTRACT permissions.
  • Build AgentOps and audit trails: require human approvals for agents performing write or external fetch actions, and publish hardened templates and developer guidelines for safe agent creation.

Why Reprompt matters beyond a single patch​

Reprompt is a clear example of how convenience features — prefilled prompts, deep links, persistent sessions and agentic follow‑ups — can be recombined into powerful attack rails. It highlights several structural lessons for vendors, IT teams and security practitioners:
  • Safety must be persistent, not just applied to the initial prompt. Model or client checks that only apply once are brittle when AI workflows support clarifications, repetitions and chained outputs.
  • Policy enforcement must be semantic. DLP geared at file attachments and strings can miss exfiltration that is encoded, chunked, or staged across multiple conversational requests. Semantic scanning and assistant‑aware policies are necessary.
  • Consumer and enterprise surfaces must be reconciled. When employee devices run consumer AI instances with broad access, tenant‑level protections are undermined and reconnaissance becomes an easier first step for targeted enterprise attacks.
  • Vendor‑hosted agent orchestration complicates visibility. If a malicious chain runs on vendor infrastructure, local network logs may show little besides normal vendor traffic. Stronger vendor auditing and defender‑side telemetry correlation are both required.
These are not theoretical points. Reprompt is conceptually similar to earlier PoCs (for example, Mermaid‑based exfiltration proofs) that used rendered artifacts or deep‑link behaviors to transfer encoded data off‑tenant with a single click. What’s changed is scale and convenience: deep links are easy to distribute via email, chat or social media, and the q parameter is a standard developer convenience across many assistant UIs.

What defenders should watch for next​

Security teams should prioritize detections and telemetry that catch the specific behavioral patterns Reprompt exploits:
  • Repeated identical or near‑identical fetch attempts (the double‑request pattern).
  • Copilot sessions that accept external follow‑ups shortly after being opened from a deep link.
  • Encoded chunks or sequences of outputs that, when reassembled, resemble PII or sensitive file summaries.
  • Unusual outbound requests from Copilot processes to nonstandard endpoints following an initial benign result.
Operators should also coordinate with vendors to obtain actionable indicators and hardened KIRs (known‑issue responses) that map to installed product versions. Because Microsoft’s public advisories can be intentionally concise, on‑premise verification against installed builds and telemetry is essential.

Final assessment​

Reprompt is a sobering demonstration that small UX conveniences can be composed into high‑impact attack rails against AI assistants. The vulnerability exploited a legitimate feature — the q URL parameter used to prefill prompts — and paired it with conversational and procedural model behaviors (repetition and follow‑ups) to create an almost invisible exfiltration pipeline. The research was responsibly disclosed and Microsoft acted quickly, but the episode exposes larger, architectural questions about how to treat external inputs, session lifecycle, and enforcement persistence in LLM‑driven assistants. For Windows users and administrators the practical guidance is straightforward and urgent: apply the January 2026 updates, verify Copilot and Edge client versions across managed devices, restrict Copilot Personal where appropriate, and treat any AI deep link as potentially dangerous until proven safe. For vendors and platform architects, the message is structural: design assistants to treat external inputs as explicitly untrusted, maintain safety checks across repeated and chained interactions, and provide enterprise‑grade governance and semantic DLP even on consumer‑facing surfaces.
Reprompt should be a turning point for how the industry thinks about assistant security: convenience without persistent, semantics‑aware controls is an invitation for abuse. The Copilot patch closed a specific vector — but the broader class of prompt‑injection and agentic chaining attacks will remain a live concern unless product design, operational controls and telemetry evolve to match the new capabilities of conversational AI.
Source: TechRadar This Microsoft Copilot AI attack took a single click to compromise users
 
A deceptively small UX convenience—allowing Microsoft Copilot to accept a prefilled prompt from a URL—was chained into a practical, one‑click data‑exfiltration technique that targeted Copilot Personal and, until Microsoft pushed mitigations in mid‑January 2026, could quietly siphon profile details, file summaries and conversation memory from authenticated sessions. Varonis Threat Labs called the exploit “Reprompt”; their technical write‑up and demonstration show how three benign‑looking behaviors—URL prefilled prompts, repeated requests, and server‑driven follow‑ups—combine into a stealthy, low‑cost exfiltration pipeline.

Background / Overview​

Microsoft Copilot is now embedded widely across Windows and Microsoft Edge, offering a conversational assistant that can access local context, recent files, account profile data and chat memory to deliver productivity features. That deep integration delivers concrete productivity gains, but also expands the attack surface: inputs such as URL query parameters or page content can become first‑class prompt material for the assistant unless explicitly treated as untrusted. The Reprompt disclosure illustrates a core lesson for AI security: convenience features that implicitly trust external inputs can be repurposed into attack vectors. Varonis’ proof‑of‑concept demonstrates the operational mechanics and shows how an attacker can maintain control of a Copilot Personal session after the user clicks a Microsoft‑hosted deep link. Multiple independent outlets corroborated Varonis’ high‑level findings and reported Microsoft deployed mitigations during January’s Patch Tuesday updates.

Anatomy of Reprompt​

Reprompt is not a single bug in the classic sense; it’s a composed attack pattern built from three interacting behaviors. Understanding each element explains why a single click—on what appears to be a legitimate Copilot link—was enough to start a stealthy exfiltration flow.

Parameter‑to‑Prompt (P2P) injection​

Many web‑based assistants support “deep links” that prefill the assistant’s input field using a query parameter—commonly named q—so users can share prompts or bookmark a task. Reprompt abuses that UX convenience by placing attacker instructions inside the q parameter. When an authenticated user opens the deep link, Copilot ingests that value as if the user had typed it, launching the malicious flow in the context of the victim’s own session. That first foothold is powerful because it elevates any attacker‑controlled URL into a remote prompt injection channel and because the URL can be hosted on trusted Microsoft domains to increase click rates.

Double‑request (repetition) bypass​

Varonis’ PoC found that Copilot’s safety logic appeared to be applied primarily to the initial request. By instructing the assistant to “do it again” or “try once more,” an attacker could convert a blocked or neutered first attempt into a second attempt that returned unredacted content or completed a prohibited fetch. This “do it twice” pattern undermines naive single‑request redaction strategies and highlights a subtle but practical weakness in conversational enforcement models.

Chain‑request orchestration (server‑driven follow‑ups)​

Once the initial prompt is accepted, an attacker server can feed successive instructions to Copilot. That remote orchestration lets the adversary probe incrementally—asking for a username, then a location, then a summary of recently opened files—encoding each datum and exfiltrating it piece‑by‑piece. Because the real, sensitive instructions are delivered dynamically after the initial click, static inspection of the starting URL or client‑side telemetry that only looks at the first request will miss the ongoing theft. Varonis’ demo further claims the session can remain able to accept follow‑ups even after the user closes the chat window in some product variants, enabling a silent background pipeline.

What the proof‑of‑concept accessed (and what it did not)​

In controlled lab demonstrations, Reprompt was shown to be capable of extracting a range of information Copilot Personal normally has access to:
  • User display name and profile attributes.
  • Location data inferred from profile or device context.
  • Lists and short summaries of recently accessed files.
  • Conversation memory entries and past chat summaries.
  • Derived personal details such as calendar events or vacation plans.
Varonis’ PoC demonstrates small‑chunk exfiltration: the attacker instructs Copilot to output tiny encoded fragments and ship each to attacker‑controlled URLs, making the aggregate leak harder to detect. That technique is intentionally stealthy: pieces of data are exfiltrated in many tiny transactions rather than a single large surge. Caveats: the demonstrations were performed under lab conditions and targeted Copilot Personal (consumer) rather than Microsoft 365 Copilot (enterprise). Reporters and Varonis state there was no confirmed mass in‑the‑wild exploitation at time of disclosure; however, the attack’s low cost and scalability via phishing mean defenders had to act quickly. Always treat lab PoCs as feasibility proofs that define risk—they do not necessarily imply observed large‑scale abuse.

Microsoft’s response and timeline: verification and gaps​

Varonis says it disclosed the issue to Microsoft under coordinated disclosure. Public reporting places the patch deployment in the January 2026 Patch Tuesday cycle (around January 13–14, 2026), and multiple outlets reported Microsoft implemented mitigations for Copilot Personal as part of that update stream. Microsoft’s public release notes for Microsoft 365 Copilot show active hardening and feature changes in the same January window, but vendor advisories typically omit exploit‑level details and they rarely adopt third‑party names like “Reprompt” verbatim. Administrators should therefore confirm their client and OS component versions against Microsoft’s official patch notes to verify whether their environment received the relevant mitigation. A few facts are verifiable and important:
  • Varonis published a detailed technical write‑up and demonstration labeled Reprompt, explaining P2P injection, double‑request bypass and chain‑request orchestration.
  • Independent reporting by outlets such as Malwarebytes, Windows Central and BleepingComputer corroborated the core elements and reported the issue was fixed during mid‑January Patch Tuesday.
  • Microsoft’s published Copilot release notes and Microsoft 365 advisories confirm active January updates to Copilot surfaces but do not necessarily use the term “Reprompt” or disclose full exploit mechanics in public advisories—a standard vendor practice for responsible disclosure. Administrators must therefore rely on version identifiers and KB numbers to confirm mitigation status.
Warning: some outlets report that Varonis initially notified Microsoft in August 2025; while multiple reports cite that timeline, Microsoft’s public advisories do not reference the Reprompt name and did not publish an exploit timeline with those exact dates at the time of the Patch Tuesday updates. Treat the internal disclosure timeline as reported but not independently verified by a vendor advisory.

Why Reprompt slipped through: design and telemetry failures​

Reprompt exploited product design choices and mismatches between UX convenience and security assumptions. Key systemic weaknesses include:
  • Assistants treating URL parameters and page context as trusted prompt material rather than explicitly untrusted inputs, allowing prefilled instructions to run in an authenticated session.
  • Safety and data‑loss protections that apply to a single request lifecycle rather than persisting across conversational state, repeated requests and follow‑on orchestration.
  • Evasion of endpoint and egress monitoring when the malicious logic executes on vendor‑hosted infrastructure or within chained assistant calls—traditional EDR and perimeter DLP often rely on network signals that don’t cover vendor‑side flows.
  • Divergent product governance between consumer Copilot Personal and enterprise Microsoft 365 Copilot: tenant‑level Purview auditing, DLP and admin controls in enterprise scenarios make the latter harder to exploit in the same way. That divergence created an environment where consumer sessions on corporate devices (or mixed‑account usage) were a meaningful residual risk.
These are not coding errors detectable by fuzzers—Reprompt is a socio‑technical exploit that turns interaction design into an attacker primitive. Addressing it requires architectural changes across the stack, not just a single patch.

Practical, prioritized mitigations for IT teams and users​

The Reprompt episode demonstrates both how quickly attackers can convert convenience into weaponry and which practical steps reduce exposure. The guidance below synthesizes vendor guidance, Varonis recommendations, and stand‑alone defensive practices.

Immediate (0–72 hours)​

  • Install January 2026 security updates for Windows, the Copilot client, and Microsoft Edge. Confirm KB and client versions centrally via your patch management system. Multiple independent reports tie the fix to the mid‑January 2026 Patch Tuesday stream.
  • Where possible, restrict or disable Copilot Personal on managed devices until controls are verified. Use Group Policy, Intune configuration profiles or the Copilot Control System to enforce limits.
  • Educate end users: advise staff not to click unsolicited Copilot deep links and to verify Copilot‑branded URLs through a second channel before opening them.

Short term (weeks)​

  • Audit all managed devices for Copilot Personal usage and separate consumer identities from tenant‑managed accounts. Ensure corporate data is accessed only through tenant‑managed Copilot where Purview and DLP can operate.
  • Add monitoring and correlation rules that identify long‑running Copilot sessions that accept host‑side follow‑ups, and correlate those sessions with outbound DNS/HTTP anomalies or unexpected domains being contacted by copilothosted fetches.
  • Configure email and web filtering to rewrite or quarantine deep links with q parameters coming from external senders; flag Copilot URLs in mail clients for user verification.

Medium to long term (months)​

  • Extend DLP semantics into the AI assistant stack: introduce explicit EXTRACT permissions that require audited approval before an assistant may summarize or transmit sensitive content externally.
  • Treat all external inputs (URL parameters, fragments, page text) as untrusted by default. Implement server‑side sanitization and enforce safety checks for repeated and chained requests—not just the initial instruction.
  • Architect assistant sessions with least‑privilege: require explicit, auditable consent for any assistant action that accesses protected data. Provide administrators with consistent tooling to enforce policy across both Copilot Personal and Microsoft 365 Copilot experiences.

Risk assessment: who should be worried and why​

  • Consumer users: individuals should be cautious about clicking Copilot deep links sent via email or chat. The attack targeted Copilot Personal, and consumer sessions often lack enterprise DLP controls. However, the operational severity is limited by the need for an active Copilot Personal session and the product behavior under the specific PoC conditions.
  • Enterprises: the biggest risk is hybrid use—employees who use consumer Copilot on corporate devices or who mix tenant and personal accounts. That behavior can provide reconnaissance that feeds a follow‑on enterprise compromise. Adopt an organizational policy that separates managed tenant Copilot usage from consumer Copilot Personal and enforce device settings accordingly.
  • Vendors and platform owners: Reprompt is a wake‑up call to bake end‑to‑end safety into the assistant lifecycle. Fixing a single pathway is not sufficient—safeguards must persist across repeated calls, regenerated outputs, and dynamic follow‑ups.

Broader implications for AI assistant security​

Reprompt sits in a lineage of prompt‑injection and exfiltration research that shows attackers will weaponize interaction affordances: hidden document instructions, generated UI artifacts with embedded links, and deep links that prefill prompts. Prior PoCs (Mermaid exfiltration, LOLCopilot class demonstrations) have shown the same conceptual risks in different forms: when agents are given privilege to act on behalf of users and to fetch external resources, attackers will find ways to make that power do their bidding. This class of threats changes the defender’s calculus. Network‑centric controls remain useful but insufficient: attackers can route activity through vendor‑hosted flows and chains that leave the endpoint with no obvious egress signature. Instead, defenders must combine platform controls (tenant DLP, Purview auditing), runtime protections that persist across conversational state, and operational policies limiting assistant privileges.

Recommendations for product teams (what vendors must do next)​

  • Treat all external inputs as untrusted by default: sanitize and validate URL q parameters, fragments, and page text. Do not treat them as user‑typed prompts without explicit confirmation.
  • Make safety checks persistent: apply redaction, URL fetch blocking, and contextual policy enforcement across the entire conversation lifecycle, including regenerated outputs and repeated invocations.
  • Provide auditable consent flows and least‑privilege controls: require explicit, auditable approvals for extractive actions that access calendar, email, files, or user profiles.
  • Expose telemetry and audit logs: allow tenant admins to see full conversation traces and any external fetches initiated on their users’ behalf so defenders can spot suspicious long‑running flows.
  • Coordinate disclosure transparency: while vendor advisories should avoid exploit specifics that enable reproduction, they must provide actionable details (affected versions, mitigation KBs, guidance to detect indicators) that allow admins to verify patch status quickly.
These steps move the ecosystem from reactive patching toward a resilient, least‑privilege model for conversational assistants.

Critical analysis: strengths of the research — and the residual risks​

Varonis’ Reprompt research is significant for several reasons. It demonstrates a realistic and repeatable abuse pattern using default product features; the PoC required minimal attacker sophistication; and the attack could be distributed at scale via phishing because the link is Microsoft‑hosted and appears legitimate. Those aspects combine to create high operational impact for a low attacker cost. At the same time, the practical exploitability rests on product behavior and session conditions (an active Copilot Personal session, specific client variants, and the way safety checks were applied). Those constraints mean Reprompt is not a universal remote code execution worm; it is a socio‑technical design exploit. Microsoft’s mitigations and the availability of enterprise Copilot governance (Purview, DLP) materially reduce the risk for properly managed tenant environments. Nevertheless, residual risks remain:
  • The fundamental pattern—converting trusted UX inputs into prompt injection channels—applies to many assistants and sites. Unless prefilled prompts and external inputs are treated as untrusted by default, a new Reprompt variant is plausible.
  • Detection blind spots persist when exfiltration is orchestrated from vendor infrastructure and delivered as tiny encoded fragments.
  • Users and admins frequently lag behind updates; even when a vendor patches a flow, there is operational drift as mixed device fleets and unmanaged consumer Copilot instances proliferate.
For defenders, the right posture is layered: patch promptly, restrict consumer assistant usage on corporate endpoints, and push for architectural platform changes that eliminate the root cause.

Conclusion​

Reprompt exposed a hard truth about assistant security: convenience features—deep links, prefilled prompts, and agentic follow‑ups—can be turned into invisible attack rails when assistants treat external inputs as trusted. Varonis’ research provided a clear, operational proof‑of‑concept showing a one‑click exfiltration chain that exploited Parameter‑to‑Prompt injection, repeated‑request bypasses, and server‑driven orchestration to siphon user context out of Copilot Personal sessions. Microsoft moved quickly with mitigations in January 2026’s Patch Tuesday releases, and enterprise Microsoft 365 Copilot offers stronger tenant governance by design; nonetheless, Reprompt should be a catalyst for a more rigorous approach to assistant security: treat external inputs as untrusted, enforce persistent safety across conversational state, instrument auditable consent and EXTRACT controls, and separate consumer assistants from tenant‑governed experiences on managed devices. The next Reprompt variant is not a question of if, but when—and the defenders best positioned to win will be the ones who harden both product architecture and operational practice now.
Source: Computerworld One click is all it takes: How ‘Reprompt’ turned Microsoft Copilot into a data exfiltration tool
 
A deceptively small UX convenience — allowing Microsoft Copilot to accept a prefilled prompt from a URL — was chained into a practical, one‑click data‑exfiltration technique that security researchers named “Reprompt,” and the discovery has exposed how quickly assistant conveniences can become operational attack surfaces for large‑scale phishing and covert extraction of user context.

Background / Overview​

Microsoft Copilot has been folded into Windows and Edge as a conversational assistant that can access local context, recent files, profile attributes and chat memory to deliver productivity features. That wide access is the point of Copilot: give users an assistant that understands their environment and can synthesize useful, actionable answers. But convenience features that implicitly trust external inputs — for example, prefilled prompts supplied in a URL — create a new class of attack surface where an attacker can influence what the assistant sees before security controls apply. Researchers at Varonis Threat Labs published a technical write‑up and demonstration of an exploit chain they call Reprompt, showing how three relatively innocuous behaviors — a parameter‑to‑prompt (P2P) injection via a URL, a double‑request repetition bypass, and a server‑driven chain of follow‑up requests — can be composed into a stealthy, low‑friction exfiltration pipeline that needs only a single click to begin. Varonis’ demonstration and independent reporting indicate Microsoft deployed mitigations during the January 2026 Patch Tuesday updates. This feature unpacks the Reprompt chain, verifies the core technical claims against published research and independent reporting, assesses operational impact for consumer and enterprise environments, and provides a prioritized set of mitigations and architectural hardening recommendations for administrators and platform engineers.

Anatomy of Reprompt: how a single click becomes an exfiltration pipeline​

1) Parameter‑to‑Prompt (P2P) injection — the initial foothold​

Many web‑hosted assistants and demo pages support “deep links” that prefill the assistant’s input field using a query parameter (commonly named q) so a link can automatically populate the prompt when the page loads. Reprompt weaponizes that UX convenience by embedding attacker instructions in that q parameter; when an authenticated user opens the deep link, Copilot ingests the value as if the user had typed it, executing the instruction inside the victim’s own authenticated session. The attacker gets a foothold without the user typing a single character and can deliver that link from a Microsoft‑hosted URL to improve click rates. This makes a trusted domain and a short, innocuous link an effective phishing vector: the URL looks genuine, the malicious payload is hidden in the address bar, and the action occurs under the victim’s identity and consent model. That combination is what makes the P2P step so dangerous in practice.

2) Double‑request (repetition) bypass — turning a blocked fetch into a successful one​

Varonis’ PoC found Copilot’s client‑side safety logic often applied primarily to the initial fetch or request. By instructing Copilot to “try again,” “do it twice,” or otherwise repeat the same operation, the attacker could subvert one‑shot protections: the first attempt is blocked or redacted, but the second attempt may be processed without the same enforcement, returning the sensitive content or successfully fetching an attacker URL. This “do it twice” heuristic undermines naive single‑request redaction strategies that rely on first‑pass filtering. The consequence is straightforward: protections that only operate on the first call are insufficient once conversational logic and repetition are part of the adversary’s toolkit.

3) Chain‑request orchestration — dynamic, server‑driven follow‑ups​

After the initial prompt is accepted, the attacker’s server can reply with new instructions that drive the next stage of the interaction. That server‑side control lets the adversary dynamically probe for specific fields — for example, “what is the user’s display name?”, “summarize the most recently opened files”, or “what city is in the user profile?” — and then encode and exfiltrate those fragments one small piece at a time. Because the follow‑ups are delivered from a remote server after the initial click, static inspection of the original URL or the first chat entry will not reveal the full exfiltration plan. Varonis’ demonstration shows the chain can harvest profile attributes, file summaries and conversation memory snippets incrementally; the technique’s stealth comes from small‑chunk exfiltration and use of vendor‑hosted flows that minimize observable egress on the endpoint.

What the proof‑of‑concept accessed — scope and limits​

In laboratory demonstrations, researchers showed Reprompt could access data that Copilot Personal legitimately has permission to read as part of its normal operation:
  • User display name and profile attributes accessible to the authenticated session.
  • Location details inferred from profile or device context.
  • Lists and short summaries of recently accessed files (file names and brief previews).
  • Conversation memory and chat history snippets saved in Copilot’s session history.
  • Derived personal information (calendar events, planned vacations) when Copilot could infer them from session context.
These data pieces are not exotic: they fall into the set of contextual metadata and derived facts that assistants are designed to access to be useful. What makes Reprompt serious is the ability to stitch many small, seemingly innocuous outputs together into a full profile and to ship those fragments to an attacker domain incrementally so that detection is harder. Caveat: the PoC required an active Copilot Personal session and was demonstrated in lab conditions. At the time of public disclosure there were no confirmed reports of mass in‑the‑wild exploitation, but the attack’s low friction and ease of scaling made a rapid mitigation imperative.

Microsoft’s response and the disclosure timeline — what’s verifiable​

Varonis publicly published its Reprompt write‑up and demonstration in January 2026; the blog entry lists its last update as January 14, 2026. Independent outlets reported Microsoft deployed mitigations as part of the mid‑January 2026 Patch Tuesday (around January 13–14, 2026). Those two points — Varonis’ disclosure and Microsoft’s mitigations in the January Patch Tuesday stream — are corroborated by multiple independent reports. One detail that remains unclear in public vendor advisories is the exact internal disclosure date Varonis used when reporting to Microsoft; some community summaries referenced a late‑August 2025 disclosure to Microsoft, but Microsoft’s update notes do not adopt the third‑party label “Reprompt” or provide a discrete timeline that confirms that August date. That particular timeline element should therefore be treated cautiously until Microsoft publishes a dedicated technical advisory that enumerates internal dates. Importantly, public reporting and Varonis both emphasize the exposure targeted Copilot Personal (the consumer instance integrated into Windows and Edge). Microsoft 365 Copilot — the tenant‑managed enterprise product — was reported as not affected in the same way because it sits behind tenant‑level governance such as Purview auditing, tenant DLP and admin controls. Enterprise protections provide stronger policy enforcement and logging that make the exact Reprompt chain much harder to execute at scale in a managed tenant.

Why consumer Copilot vs enterprise Copilot made the difference​

The Reprompt incident highlights a structural separation in Microsoft’s Copilot ecosystem:
  • Copilot Personal is designed for consumer scenarios and is tightly integrated into the local OS and browser. It emphasizes low friction and immediate utility, which means it often runs with the logged‑in user’s privileges and has access to local content without tenant‑level oversight. That design makes it a high‑value target for attacker flows that depend on a privileged session.
  • Microsoft 365 Copilot operates within enterprise boundaries with governance surfaces — Purview auditing, sensitivity labels, tenant DLP and admin controls — that create policy enforcement points and richer audit trails. Those tenant‑level controls substantially reduce the risk that a simple, unaudited deep link will be able to extract corporate secrets without being flagged.
The practical implication is this: organizations that permit or fail to restrict consumer Copilot usage on corporate devices effectively widen their attack surface, because attackers can aim at consumer flows that are less governed but still run under a corporate identity on managed devices. The simplest organizational control — preventing use of Copilot Personal on corporate devices — removes an entire class of exposure.

Detection challenges and why traditional tools struggled​

Reprompt succeeds by shifting most of the operational work outside the endpoint:
  • The initial URL and prompt are small and look innocuous; static inspection of a single query parameter rarely reveals the dynamically generated follow‑on instructions.
  • The repeated, small‑chunk exfiltration pattern minimizes anomalous network spikes and blends into normal assistant traffic and web fetches.
  • Much of the chain runs in vendor‑hosted flows or over managed platform APIs, reducing egress signals visible to local EDR/NGFW appliances.
Because of these properties, defenders cannot rely exclusively on endpoint or egress anomaly detection; they must combine patching, governance (deny or control consumer Copilot on managed devices), and semantic policy enforcement that treats assistant outputs as potentially exfiltration vectors.

Practical mitigation checklist — immediate, short and medium term​

Security teams should treat Reprompt as a call to action for three parallel tracks: patching, governance, and architectural hardening.
Immediate (0–72 hours)
  • Confirm Windows, Microsoft Edge and Copilot client components are patched with the January 2026 updates; apply vendor KIRs and verify KB/client versions.
  • Restrict or disable Copilot Personal on managed devices (Group Policy, Intune, or device configuration profiles) until controls are verified.
  • Update phishing filters and run targeted awareness outreach: warn users about Copilot‑branded deep links and instruct them to treat such links as untrusted.
Short term (weeks)
  • Audit your fleet for devices using Copilot Personal and segregate consumer identity usage from tenant accounts.
  • Apply tenant DLP, Purview policies and sensitivity labels where enterprise data must be protected; enforce EXTRACT or equivalent permissions before allowing an assistant to summarize or export protected content.
  • Instrument logging to correlate Copilot session events with outbound HTTP/DNS requests and long‑running session markers that accept remote follow‑ups.
Medium term (months)
  • Require explicit consent for any assistant operation that extracts or externally forwards content; log and audit every EXTRACT action with tamper‑resistant telemetry.
  • Build semantic DLP into the assistant stack so that content classification and sensitivity checks apply to every chained or repeated request, not just the initial one.
  • Work with platform vendors to ensure prefilled prompt inputs (URL parameters, fragments, page text) are treated as untrusted by default and sanitized before inclusion in assistant context.

Architectural hardening — changing the model, not just the rules​

Reprompt exposes a deeper design gap: conversational systems that treat external inputs as first‑class prompt material must adopt a different threat model from classic web apps. Key architectural shifts include:
  • Least privilege for assistant access: require fine‑grained, auditable grants for assistants to read files, calendars or chat memory. Extraction should require an explicit, auditable permission pivot.
  • Continuous and contextual authentication: employ adaptive session validation and re‑authentication for high‑risk operations instead of single initial authentication that remains valid for long‑running sessions.
  • Sanitization pipeline for external inputs: treat URL query parameters, fragments and page content as untrusted. Run sanitization and intent classification before including them in the assistant context.
  • Repeat‑aware enforcement: extend safety checks across the entire conversation lifecycle so that repeated or reframed requests are subject to the same or stricter enforcement than the initial request.
These changes are more invasive than a patch, but they are necessary for a mature, enterprise‑grade assistant platform that will interact with sensitive data at scale.

Risk assessment — strengths, limitations and plausibility of exploitation​

Strengths of the research and its operational implications:
  • The attack model has extremely low interaction cost — one click — which makes large‑scale phishing an obvious distribution vector.
  • The chain composes legal assistant capabilities (reading files, summarizing chats) with benign‑looking UX behavior (prefilled prompts and repetition) to create a pathway that evades many signature‑based defenders.
  • Vendor‑hosted orchestration reduces local telemetry and complicates detection by network egress tools.
Limitations and mitigating factors:
  • The PoC targeted Copilot Personal and required an active authenticated session; it was not a worm‑like remote code execution vulnerability. That constrains its automatic spread compared with classical remote exploits.
  • Enterprise Copilot offerings are typically accompanied by Purview, DLP and admin policies that substantially raise the bar for the same technique to exfiltrate corporate secrets from tenant‑managed accounts.
  • At the time of public disclosure there were no confirmed reports of mass in‑the‑wild exploitation; however, the absence of evidence is not evidence of absence for covert, low‑volume exfiltration flows.
Taken together, Reprompt is a credible and operationally useful proof‑of‑concept that underscores the urgency of rethinking prompt provenance, session model and continuous enforcement for assistants.

Broader implications for AI security and vendor responsibility​

Reprompt is not just about one product: it is a case study in how convenience features can be weaponized when assistants are given wide data access and when external inputs are implicitly trusted. The incident should push platform vendors, enterprise architects and security teams to:
  • Treat conversational inputs with the same provenance discipline applied to user input in classical web apps. Untrusted by default must become the default posture for prefilled prompts and page‑sourced context.
  • Standardize disclosures and coordinated‑disclosure timelines for AI security research to reduce confusion. Public advisories that omit researcher names or timelines can slow defenders’ ability to reconcile patch streams with risk. The precise internal disclosure date for Reprompt remains unclear in public vendor notes, so defenders should err on the side of assuming early exposure windows.
  • Build platform‑level DLP semantics and EXTRACT permissions so that any programmatic extraction of content is explicit, auditable and constrained by policy.
Vendors must move beyond cosmetic guardrails and invest in continuous enforcement that spans the entire conversational lifecycle, not just the first prompt.

Conclusion​

Reprompt is a clarifying incident for the state of consumer AI assistants: it shows how modest usability features — a prefilled prompt parameter here, a repeated instruction there — can be recomposed into a highly effective exfiltration pipeline that operates under legitimate authentication and leverages vendor‑hosted orchestration to evade local detection. The exploit chain required only a single click in lab demonstrations, and Microsoft deployed mitigations during the January 2026 Patch Tuesday updates after Varonis’ disclosure. The practical takeaway for Windows administrators and security teams is immediate and binary: verify patches and mitigate consumer Copilot usage on managed devices now, but don’t stop there. Enforce least privilege, make external inputs untrusted by default, require auditable EXTRACT consent for assistant‑driven exports, and extend safety checks across repeated and chained requests. Those investments will convert Copilot — and other assistants — from promising productivity accelerators into dependable, auditable tools for enterprise operations rather than unattended sensors and leakage channels.
Reprompt will not be the last variant to toy with conversational logic, repetition and remote orchestration. The next iteration will likely be faster, more subtle and better able to blend into legitimate assistant behavior. The right response is not only to patch the specific vector, but to reengineer assistant trust boundaries so that convenience does not hand attackers the keys to privileged user sessions.
Source: Computerworld https://www.computerworld.com/artic...t-copilot-into-a-data-exfiltration-tool.html]
 
Microsoft’s Copilot ecosystem was rattled in mid‑January when security researchers disclosed a novel, one‑click exfiltration technique — dubbed “Reprompt” — that used Copilot deep‑links and conversational behaviors to siphon user profile data, file summaries and chat memory from authenticated Copilot Personal sessions; vendors moved quickly to mitigate the vector during January’s Patch Tuesday updates, but the episode spotlights systemic design trade‑offs in agentic AI and the urgent need for governance, telemetry and semantic DLP across consumer and enterprise assistant surfaces.

Background / Overview​

Microsoft Copilot has been embedded deeply into Windows and the Edge browser to act as a conversational assistant that can access local context, recently opened files, account profile attributes and conversational memory. That level of contextual privilege is precisely the productivity win Copilot aims to deliver, but it also enlarges the attack surface: features that accept external inputs — like prefilled prompts supplied via a URL — become potential remote‑injection channels unless treated as explicitly untrusted. Varonis Threat Labs publicly documented a chained exploitation technique they named Reprompt, demonstrating how three otherwise‑benign behaviors can be composed into a stealthy, low‑friction exfiltration pipeline that begins with a single click on a legitimate Copilot deep link. Varonis published the technical write‑up and proof‑of‑concept material on January 14, 2026, and independent reporting confirms Microsoft deployed mitigations as part of the January Patch Tuesday updates. This feature unpacks Reprompt’s mechanics, verifies the core claims against independent sources, assesses operational exposure for consumers and enterprises, and outlines immediate and architectural mitigations defenders should adopt.

Anatomy of Reprompt: how a single click became an exfiltration pipeline​

Reprompt is conceptually simple and operationally elegant because it chains three techniques that are individually common design patterns in AI assistants and web apps: Parameter‑to‑Prompt (P2P) injection, a double‑request (repetition) bypass, and server‑driven chain requests. Each building block looks innocuous in isolation; together they create a flow that can evade traditional endpoint and network defenses.

Parameter‑to‑Prompt (P2P) injection — the initial foothold​

Many web‑based assistants support “deep links” with a query parameter (commonly named q) designed to prefill the assistant’s input field so users can share or bookmark prompts. Reprompt abuses this UX convenience by placing attacker instructions inside the q parameter; when an authenticated user clicks the link, the assistant ingests that text as though the user had typed it. Because the URL can be hosted on a trusted Microsoft domain, recipients are more likely to click and the payload is largely invisible in the address bar. Why this matters: a prefilled query becomes a remote prompt‑injection channel that executes inside the victim’s authenticated session context. That removes many of the friction points attackers normally face: no binary malware, no plugin, and no additional user action beyond clicking a convincing, vendor‑hosted link.

Double‑request (repetition) bypass — turning benign protections into a weakness​

Varonis observed that some Copilot client‑side safety checks were applied primarily to the initial request or fetch. By instructing the assistant to “try again” or “do this twice,” an attacker could convert a blocked first attempt into a successful second attempt that returned otherwise redacted or blocked content. This simple repetition subverts naive single‑pass filtering and highlights a subtle failure mode in conversational enforcement logic. The technical implication is clear: procedural or conversational control flows — sequences of repeated requests or reformulations — must be treated as part of the enforcement surface, not as separate, independent invocations.

Chain‑request orchestration — server‑driven follow‑ups and stealthy exfiltration​

After the initial prompt is accepted, the attacker’s server can issue successive follow‑up instructions to the live Copilot session. That allows an adversary to probe incrementally — asking for a username, then location, then a short summary of recently accessed files — and exfiltrate each piece in small chunks to avoid detection. Varonis’ PoC also demonstrates that in some product variants the session could continue accepting remote prompts even after the user closed the chat window, enabling silent background exfiltration. The net result is a pipeline that hides the malicious intent behind a benign initial request, makes static inspection of the originating URL insufficient, and pushes meaningful activity onto vendor‑hosted or model‑hosted flows that local egress monitoring often does not inspect.

Verification and timeline: what was confirmed, and by whom​

Varonis’ blog post that publicly introduced the Reprompt PoC was posted January 14, 2026 and includes demonstration materials and a technical breakdown. Independent security reporting — including Ars Technica, Malwarebytes, Windows Central and others — corroborated the core mechanics and reported that Microsoft deployed mitigations in mid‑January’s Patch Tuesday updates (the January security update stream that included fixes released around January 13–14, 2026). Key, verifiable timeline points:
  • Varonis posted Reprompt technical details and videos on January 14, 2026.
  • Multiple independent outlets reported Microsoft applied mitigations for Copilot Personal during the January Patch Tuesday update cycle (published around January 13–14, 2026).
Caveats and verification notes:
  • The Reprompt demonstration is a proof‑of‑concept developed in lab conditions; public reporting cites no evidence of mass, in‑the‑wild exploitation at the time of disclosure — an encouraging but non‑definitive fact.
  • Microsoft 365 Copilot (enterprise) was repeatedly reported as not affected in the same way, owing to tenant‑level governance such as Purview auditing and tenant DLP; Reprompt targeted Copilot Personal (consumer) in the published PoC. Defenders should validate this for their specific tenant, service configuration and client builds.

What the proof‑of‑concept accessed — scope of the demonstrated leak​

In controlled demonstrations Varonis showed the chain could return small, sensitive items that Copilot Personal normally has access to:
  • Authenticated user display name and profile attributes.
  • Location data inferred from profile or device context.
  • Lists and short summaries of recently accessed files.
  • Conversation memory entries and derived personal details (calendar events or travel plans).
Importantly, the PoC emphasizes incremental, small‑chunk exfiltration — encoding and shipping tiny fragments to an attacker endpoint to evade volumetric detection. That pattern is operationally realistic: it improves stealth and scales across phishing campaigns.

Why Copilot Personal was more exposed than enterprise Copilot​

Microsoft’s enterprise Copilot offering (Microsoft 365 Copilot) is layered with admin governance features — Purview auditing, tenant‑level DLP rules, sensitivity labeling and admin controls — that provide visibility and enforcement not available to consumer Copilot Personal. The Reprompt chain leveraged consumer‑level affordances (deep links, session lifetimes and URL‑prefill UX) that, in the published PoC, did not fall under tenant governance protections. This separation matters for organizations that permit consumer Copilot use on corporate devices or that allow mixed personal/enterprise accounts. Operational consequence: reconnaissance from a compromised consumer Copilot session can feed targeted follow‑on attacks against enterprise resources unless organizations enforce boundaries between managed identities and consumer AI features.

Practical mitigations — immediate actions for admins and users​

Short term (apply immediately)
  • Apply the January 2026 Patch Tuesday updates and confirm the Copilot client and Edge builds are updated across managed endpoints. Verify update deployment status centrally.
  • Where governance is required, restrict or block Copilot Personal on corporate devices and prefer Microsoft 365 Copilot for any workflows dealing with sensitive data.
  • Add email/URL filtering that rewrites or warns on Copilot deep links (links that point to copilot.* or known Copilot domains) in external messages. Treat unsolicited Copilot links as suspicious by default.
Defensive hardening and monitoring
  • Instrument Copilot‑hosted request patterns in telemetry: flag unusual sequences of outbound fetches initiated by Copilot processes or anomalous downstream callbacks to uncommon endpoints. Look for many small outbound transactions (low‑volume, high‑count) from Copilot services.
  • Shorten session lifetimes or implement stricter session isolation for consumer Copilot contexts on managed devices; require reauthentication for sensitive tasks.
  • Implement semantic DLP and Purview policies where possible to detect and block exfiltration of sensitive items, and enforce data labelling across endpoints. Prefer server‑side enforcement that persists across repeated or chained interactions.
User guidance
  • Do not click unsolicited Copilot deep links. Treat vendor‑hosted URLs as potentially weaponized when they arrive from untrusted senders.
  • If your organization permits Copilot Personal, restrict it on corporate devices until patching and policy changes are verified.
Longer‑term product and architecture changes (for vendors and platform teams)
  • Treat all external inputs (URLs, query parameters, page content) as explicitly untrusted by default and require an explicit “trusted input” annotation before they can be incorporated into a formal prompt context.
  • Extend enforcement coverage across procedural conversation flows: safety checks must be persistent and robust across repeated invocations, not only on the first request.
  • Provide enterprise‑grade semantic DLP and audit trails even for consumer‑facing assistants when they run on managed devices or with corporate accounts.

Risk analysis: strengths of the research and open questions​

Notable strengths
  • The Reprompt PoC exposes a realistic and low‑cost attack vector that abuses everyday UX conveniences like deep links — a high‑impact finding because it lowers the attacker cost of entry.
  • The chain demonstrates how procedural conversational features and model behaviors (e.g., repeat/iterate) can undermine single‑pass safety assumptions. This is an important, generalizable insight for all assistant designers.
Unresolved or unverifiable aspects (flagged as caution)
  • Persistence claims (the ability to maintain control after the user closes the chat) are shown in Varonis’ PoC for some product variants. While credible in the lab, defenders should treat persistent background‑execution claims as platform‑ and version‑dependent until verified against specific client builds and configurations. The demonstration is authoritative, but complete real‑world reproducibility across all Copilot deployments has not been independently published at the same level of granularity.
  • The public record includes no confirmed mass exploitation reports at time of disclosure; absence of evidence is not evidence of absence, and the technique’s low friction makes it attractive to opportunistic phishers. Continued monitoring is warranted.
Effect on defenders and attackers
  • For defenders: Reprompt highlights blind spots in endpoint and network monitoring when significant activity is routed through vendor‑hosted flows or model‑managed orchestration. Organizations must expand their definition of “trusted” operations on managed endpoints.
  • For attackers: the tactic scales easily via phishing campaigns because it uses vendor domains and requires minimal social engineering sophistication.

The wider trend: agentic AI, commerce integrations and enterprise pilots​

Reprompt arrives at the same time the industry is accelerating agentic AI — assistants that not only answer questions but also act on behalf of users. That shift is visible across diverse headlines this week: Alibaba upgraded its Qwen app to support in‑chat shopping, payments and travel bookings, enabling transactions and end‑to‑end task completion inside the assistant interface; that demonstrates the powerful utility of agentic systems but also the risk profile when assistants can effect real transactions. Enterprises are also racing to operationalize assistants: Hertz published a case study showing an AI agent built with Microsoft Power Platform and Copilot Studio reduced customer response time by about 15% in a pilot — a concrete productivity gain from carefully scoped, trusted source integrations. The Hertz pilot highlights how trusted knowledge sources and narrowly constrained agents can deliver clear ROI without exposing enterprise datasets to the same class of consumer‑side risks. Taken together, these stories underline a recurring trade‑off:
  • Agentic assistants yield real business value by simplifying multistep tasks and integrating payments, booking and internal systems.
  • The deeper an assistant acts, the larger and more sensitive the attack surface becomes — especially when external inputs or vendor‑hosted orchestration are part of the execution path.

Recommended short checklist for Windows and Copilot administrators​

  • Prioritize January 2026 Patch Tuesday security updates and verify Copilot and Edge client versions in your environment.
  • Restrict Copilot Personal on corporate devices; standardize on Microsoft 365 Copilot for sensitive workflows.
  • Implement email and URL filtering that flags or rewrites deep links and educates users to treat Copilot links with caution.
  • Extend DLP and auditing to include assistant‑generated output and model‑hosted interactions where feasible.
  • Monitor for anomalous, low‑volume, high‑count outbound transactions originating from Copilot processes and vendor‑hosted callbacks.

Conclusion — design lessons and the path forward​

Reprompt is a decisive reminder that convenience equals risk when AI assistants are permitted to accept and act on external inputs without persistent, semantics‑aware enforcement. The attack is not a traditional buffer overflow or privilege escalation; it is a design‑level failure mode in how assistants incorporate external text into privileged contexts and how conversational logic can be weaponized through repetition and chaining.
The vendor response and rapid Patch Tuesday mitigations are evidence that responsible disclosure and rapid engineering cycles can close specific vectors. However, the broader class of prompt‑injection and agentic chaining attacks remains a live concern until platforms treat external inputs as untrusted by default, extend enforcement across conversational flows, provide enterprise‑grade DLP and audit surfaces even for consumer contexts on managed devices, and equip defenders with telemetry that surfaces model‑driven orchestration.
For administrators and users, the most practical short‑term guidance is straightforward: update now, restrict consumer Copilot use on corporate devices, monitor for unusual Copilot‑originated activity, and treat unexpected Copilot deep links as hostile until proven safe. For platform architects, the imperative is clearer: build assistants that assume adversarial inputs, maintain consistent enforcement across retries and chains, and make governance and telemetry first‑class features rather than afterthoughts.
The Reprompt disclosure should be a catalyst for every vendor that exposes conversational agents to the web: convenience without persistent, semantics‑aware controls is an invitation for abuse — and agents that act on our behalf must be designed with an equal measure of defensive rigor.
Source: MSN https://www.msn.com/en-us/news/tech...-cent-using-ai-agent-built-on-power-platform]
 
Microsoft’s latest stable update to Edge makes one of the most requested small fixes—an easy, supported way to hide the Copilot toolbar button—while simultaneously shipping new enterprise controls and several under‑the‑hood changes that shift policy, performance names, and WebGL behavior for developers and admins.

Background​

Microsoft has been integrating Copilot across Windows, Office, and Edge for more than two years, turning the browser into a primary surface for the company’s generative‑AI features. That strategy produced strong productivity features for organizations but also vigorous consumer backlash over discoverability and perceived intrusiveness. The Edge 144 stable release arrives as a pragmatic response: a straightforward UI path for end users to hide Copilot’s toolbar affordance, plus dedicated policies so IT can enforce visibility in managed environments.

What shipped in Edge 144 — the headline changes​

Edge 144 bundles a handful of user‑visible and administrative changes that matter to different audiences:
  • A supported UI flow to hide the Copilot toolbar icon.
  • A new enterprise policy (Microsoft365CopilotChatIconEnabled) to control Copilot icon visibility in Edge for Business.
  • A policy to prefer the Primary work profile for opening external links (EdgeOpenExternalLinksWithPrimaryWorkProfileEnabled).
  • Address‑bar contextual nudges that offer quick Copilot Chat summaries of the current page (Edge for Business).
  • A technical change to WebGL fallback on Windows (WARP replaces SwiftShader by default on GPU‑less systems).
  • Two Performance setting names were updated: Efficiency modeEnergy saver, and Efficiency mode for PC gamingPC gaming boost.
Each of these items affects different constituencies—consumers, enterprise admins, developers, and automation teams—often in ways that are subtle but consequential.

The user fix: hide Copilot from the toolbar​

What changed for end users​

The most visible concession in Edge 144 is a one‑click supported path to remove the Copilot icon from the toolbar. Right‑click the Copilot icon and select Hide Copilot; Edge opens the relevant Settings page where the Show Copilot toggle will remove the toolbar button. This eliminates the need for brittle registry edits, command‑line flags, or third‑party “debloat” tools that many users previously resorted to.

Step‑by‑step (user)​

  • Right‑click the Copilot icon in the Edge toolbar.
  • Choose Hide Copilot; Edge opens Settings at the Copilot/Sidebar control.
  • Toggle Show Copilot off to remove the toolbar button.
This is a UI‑level toggle: it reduces the visibility of Copilot but does not necessarily remove every other entry point (sidebar, keyboard shortcuts, address‑bar nudges, or protocol handlers). Users who need a more permanent or radical removal still have higher‑risk options (uninstall Copilot app where allowed, use PowerShell or system‑level tools), but those approaches are brittle and may break with future updates.

Admins gain a policy: Microsoft365CopilotChatIconEnabled​

What the policy does​

For managed profiles (Entra ID / Edge for Business), Microsoft added a Boolean policy named Microsoft365CopilotChatIconEnabled. Admins can:
  • Set to Enabled — force the Copilot icon visible.
  • Set to Disabled — hide the Copilot icon.
  • Leave Not configured — allow users to toggle visibility themselves.
This policy is the authoritative control for managed profiles and supersedes older sidebar visibility behavior for Copilot in Edge for Business. It’s supported on Windows and macOS for Edge versions in the 139+ range and later. The policy can be deployed via ADMX/Group Policy, MDM (Intune), or registry keys that map to the ADMX definitions.

Practical admin guidance​

  • Test the policy on a staging group before mass rollout; profile behavior differs by platform and enrollment state.
  • Communicate changes: changing toolbar affordances can confuse users who trained on a different UI.
  • Pair visibility policy with training and helpdesk scripts (e.g., how to re‑enable Copilot if needed).
  • Remember: hiding the icon does not necessarily disable Copilot chat functionality at the tenant level—use tenant controls and conditional access to manage access to corporate data.

Primary work profile becomes the default for external links​

Edge’s profile handling has been adjusted for enterprise scenarios: Edge can now treat the Primary work profile as the default to open external links. The behavior is controlled by the EdgeOpenExternalLinksWithPrimaryWorkProfileEnabled policy and is active on machines using Entra ID profiles. If enabled (or not configured), Edge will use the Primary work profile for external links; if disabled, it falls back to the last‑used profile behavior. Microsoft documented this change and started rolling it through the Microsoft 365 rollout channels in 2025.

Why this matters​

  • Security and compliance: routing external links into a work profile can reduce accidental data leakage from browsing sessions tied to personal profiles.
  • UX implications: users who expect “last used” behavior may find links opening in a different window or profile; admins should plan communications.
  • Admins can control the default via policy, enabling a balance between usability and governance.

Contextual nudges: Copilot meets the address bar​

Edge for Business now surfaces contextual nudges in the address bar that suggest summaries of the active page via Microsoft 365 Copilot Chat. The nudges are a small UX prompt—an affordance intended to speed comprehension for knowledge workers who spend time reading long web pages, documents, or internal portals. Admins retain control over Copilot availability in business profiles, but the nudges aim to increase discoverability where Copilot licensing exists.

Practical takeaways​

  • The nudges are designed to reduce friction: one click from the address bar summons a summary rather than requiring users to open the full Copilot sidepane.
  • Organizations should audit Copilot licensing and data handling policies before enabling broad rollout.
  • For privacy‑sensitive sites or internal intranets, consider explicit admin settings or training to avoid accidental data submission to the Copilot service.

Technical change: WARP replaces SwiftShader on GPU‑less Windows​

Edge 144 changes the default WebGL software rasterizer on Windows systems that lack a physical GPU. The browser now uses Windows Advanced Rasterization Platform (WARP) instead of SwiftShader on those systems (commonly headless VMs and certain virtualized CI agents). SwiftShader remains available only via a temporary policy (EnableUnsafeSwiftShader) for teams that must delay migration; Microsoft warns that the policy is a short‑term escape hatch and that the long‑term direction is WARP on Windows or requiring a physical GPU on non‑Windows platforms.

Impact​

  • Automation and CI: teams that run headless browser tests on Windows VMs should validate rendering and performance with WARP; behavior and graphics precision may differ from SwiftShader.
  • Security posture: deprecating SwiftShader reduces the maintenance surface for legacy third‑party rasterizers, but the temporary policy explicitly carries risk and should be used sparingly.
  • Developers: test WebGL workloads on the updated runtime to detect regressions early.

Renaming performance settings (efficiency → energy saver; gaming boost)​

Microsoft refined the wording in Edge’s Performance settings to reduce confusion. Two names changed:
  • Efficiency mode is now Energy saver — intended to communicate battery and energy‑saving behavior more directly.
  • Efficiency mode for PC gaming is now PC gaming boost — intended to convey the feature’s purpose of improving gaming performance by limiting Edge’s resource usage when games run.
This is a naming and organization update rather than a functional change, but it does matter for documentation, helpdesk scripts, and screenshots used in support materials. The change appears in Microsoft’s roadmap and release notes and is rolling out with Edge 144 and associated updates.

Strategic reading: why Microsoft balanced a user toggle with enterprise defaults​

The release demonstrates a bifurcated strategy:
  • For consumers and power users: provide a visible, reversible way to reduce Copilot surface area and decrease perceived clutter.
  • For enterprise customers and Microsoft’s product bet on Copilot: preserve and expand Copilot integration where IT can manage adoption and data governance centrally.
The net effect is pragmatic: Microsoft reduces the “annoyance factor” for individuals while protecting the product prioritization and data flows that support Copilot’s usefulness in the enterprise. This model avoids blunt removal while giving admins tools to tailor the experience to corporate policy.

Security, privacy, and data governance considerations​

  • Hiding the toolbar icon is cosmetic; it does not guarantee that Copilot cannot be invoked or that telemetry, address‑bar prompts, or deep links cannot surface the assistant. Organizations requiring strict controls must use tenant‑level governance, conditional access, and the Edge management policies in combination.
  • The Primary work profile default reduces accidental mixing of personal and corporate sessions, but admins should validate the behaviour on multi‑profile devices and document the user experience change for helpdesk staff.
  • WARP’s adoption on GPU‑less Windows likely reduces attack surface associated with third‑party graphics libraries; however, teams relying on SwiftShader‑specific behavior should migrate carefully and use the temporary policy only as a short window for transition.

Developer and automation implications​

  • WebGL: renderers behave differently. Test WebGL content and headless browser scripts in environments that mirror production (WARP vs GPU). If your CI environment was tuned for SwiftShader, expect rendering and performance variations.
  • Browser automation: changes to default profiles and Copilot entry points could affect automated UI tests that expect specific UI elements or focus locations. Update test scripts that rely on the Copilot icon or sidebar presence.
  • Packaging and support: policy enforcement may change how Edge behaves on managed devices, so packaging flows (e.g., for custom browser builds in kiosks) should explicitly set or verify the policies that affect toolbar elements.

Risks and trade‑offs​

  • Cosmetic vs functional: hiding Copilot reduces discoverability but does not equate to disabling Copilot. Relying solely on the toolbar toggle for compliance or privacy goals is inadequate.
  • User confusion: making Copilot easy to hide while keeping multiple Copilot surfaces (sidebar, address bar nudges, keyboard shortcuts) risks inconsistent user expectations—some users may assume Copilot is fully turned off.
  • Admin complexity: the addition of new policies increases control but also the policy surface area admins must track. Rollouts should be staged with communication and troubleshooting documentation.
  • Migration burden for automation teams: the WARP change and WebGL fallback behavior require validation across virtualization and CI fleets. The temporary SwiftShader policy exists but is explicitly time‑limited and marked unsafe.
Where claims in public coverage were speculative (for example, that Microsoft is removing the sidebar app list specifically to make room for Copilot), those assertions are plausible given parallel changes but are not fully confirmed by Microsoft. Treat such explanations as informed commentary, not definitive cause‑and‑effect statements.

Recommendations — what users and admins should do next​

For end users:
  • Use the supported UI path if the Copilot toolbar icon is distracting: right‑click → Hide Copilot → toggle Show Copilot off. This is reversible and low risk.
  • If you want to reduce all Copilot entry points, check sidebar settings and Windows taskbar Copilot toggles, but be aware these are separate controls.
For IT administrators:
  • Review the Microsoft365CopilotChatIconEnabled and EdgeOpenExternalLinksWithPrimaryWorkProfileEnabled policies and map them to your governance model. Test on pilot devices before broad deployment.
  • Update helpdesk scripts and documentation to reflect new names (Energy saver, PC gaming boost) and the new default profile behavior.
  • For automation pipelines, validate CI agents and rendering tasks with WARP and plan to avoid the EnableUnsafeSwiftShader policy as a long‑term solution.
  • Communicate the user experience change: explain where Copilot nudges may still appear and provide instructions for re‑enabling Copilot for users who need it.
For developers:
  • Re‑run visual and behavioral tests against Edge 144 on representative environments (Windows with WARP, macOS with GPU, headless Linux where GPU is present).
  • Update product documentation and troubleshooting guides to reference the renamed performance features.

Strengths and limitations of the update​

Strengths:
  • The UI path to hide Copilot is a clear, low‑risk user concession that addresses a common complaint without compromising enterprise controls.
  • The Microsoft365CopilotChatIconEnabled policy gives admins deterministic control over toolbar behavior in managed environments.
  • The Primary profile change and address‑bar nudges are targeted to enterprise productivity scenarios where Copilot licensing and governance are usually present.
Limitations and risks:
  • Hiding the Copilot button is largely a visibility change; deeper integrations remain and may still surface the assistant.
  • The WARP migration can create subtle rendering or performance differences for automation and WebGL workloads.
  • The proliferation of Copilot entry points across UI surfaces makes a single opt‑out difficult to achieve without stronger, tenant‑level enforcement.

Final analysis​

Edge 144 is a pragmatic, multi‑layered release: it fixes a high‑visibility user annoyance in a supported way while enhancing the browser’s enterprise posture and continuing the broader Copilot integration across Microsoft’s stack. The release is strategically balanced—tactically offering a consumer concession while preserving Copilot’s discoverability and central control in business deployments.
For IT teams, the update is an invitation to plan: review and test the new policies, prepare helpdesk communications about profile and performance changes, and validate automation pipelines on the new WebGL stack. For average users, the change is a welcome simplification: a reversible, documented way to remove an unwanted toolbar icon without resorting to hacks.
Finally, the update underscores a persistent reality of modern platform software: feature surface area and control surfaces expand together. Microsoft’s approach in Edge 144—visible user control backed by administrative governance—reflects a design choice that favors managed consistency in enterprise settings while offering individuals modest personal control. That balance will suit many organizations and users, but it also means there is no single, click‑to‑disable switch that will stop all Copilot interactions; comprehensive governance still depends on policies, licensing, and the broader Microsoft 365 and Entra controls you choose to deploy.
Conclusion
Edge 144 delivers a necessary UX fix and meaningful admin controls while shipping technical changes that require attention from teams that automate or depend on browser graphics. The update is neither a retreat from Copilot nor an unconditional victory for users who dislike the AI push; rather, it is a carefully calibrated step that acknowledges user frustration, protects enterprise management needs, and nudges the platform further toward Copilot as a core productivity layer.

Source: Thurrott.com Microsoft Edge Now Lets You Hide its Copilot Button