Microsoft’s Copilot for Microsoft 365 was supposed to make AI agents safer to run at enterprise scale; instead, recent reports show a control-plane failure that left some agents discoverable and installable despite tenant-level policy locks—forcing administrators into time-consuming, per-agent PowerShell workarounds and increasing the risk of data exposure, incomplete audit trails, and unmanaged automation.
Microsoft built the Copilot agent model to let organizations extend productivity with prebuilt, partner, and custom agents while offering admin controls to limit who can discover, install, or run those agents. Admin consoles and the Copilot Control System are the intended governance surfaces; Purview audit logging is the intended source of truth for any Copilot access to tenant content.
Despite that design, administrators and independent researchers have documented three related failure classes in recent months: (1) a policy-enforcement gap where a tenant-level “No users can access agent” setting is being ignored for some agents; (2) an audit-telemetry gap in Purview where certain prompt-driven flows produced user-visible outputs without emitting the expected resource-reference attributes; and (3) a sandbox misconfiguration in Copilot’s live Python/Jupyter environment that allowed a path‑hijack privilege escalation. These failures are distinct but mutually reinforcing and have material operational and security consequences.
Administrators have reported that the only practical remediation (until platform fixes are confirmed) has been manual, per‑agent revocation or blocking using PowerShell or REST/SDK calls—an approach that does not scale and increases the chance of human error.
Administrators should treat Copilot agents like any other third‑party application in their environment: insist on explicit deny semantics, test enforcement from representative user contexts, require phishing‑resistant authentication for AI services, and maintain cross-correlated telemetry pipelines that do not rely on a single vendor-supplied signal. Where vendor fixes are in flight, take the manual, time‑consuming steps required now—because the cost of not doing so is not merely a productivity nuisance; it is a governance and compliance liability that can be measured in exposed data, failed audits, and extended incident response timelines. (cybersecuritynews.com, learn.microsoft.com)
Source: Digital Watch Observatory Copilot policy flaw allows unauthorized access to AI agents | Digital Watch Observatory
Background
Microsoft built the Copilot agent model to let organizations extend productivity with prebuilt, partner, and custom agents while offering admin controls to limit who can discover, install, or run those agents. Admin consoles and the Copilot Control System are the intended governance surfaces; Purview audit logging is the intended source of truth for any Copilot access to tenant content. Despite that design, administrators and independent researchers have documented three related failure classes in recent months: (1) a policy-enforcement gap where a tenant-level “No users can access agent” setting is being ignored for some agents; (2) an audit-telemetry gap in Purview where certain prompt-driven flows produced user-visible outputs without emitting the expected resource-reference attributes; and (3) a sandbox misconfiguration in Copilot’s live Python/Jupyter environment that allowed a path‑hijack privilege escalation. These failures are distinct but mutually reinforcing and have material operational and security consequences.
What exactly happened
Policy enforcement failure: “NoUsersCanAccessAgent” ignored
Multiple tenant reports and security write-ups reproduced cases where administrators configured global restrictions (for example, setting agent availability to “No users can access agent”) yet some agents remained visible or installable to end users. Observers reported this behavior most frequently with Microsoft-published prebuilt agents and a subset of third‑party agents provisioned via Copilot Studio. The platform’s admin UI and the enforcement code paths appear to be operating on divergent state, or some privileged provisioning flows are bypassing tenant-level UI checks—producing inconsistent behavior across product surfaces (web, Teams, Outlook). (cybersecuritynews.com, learn.microsoft.com)Administrators have reported that the only practical remediation (until platform fixes are confirmed) has been manual, per‑agent revocation or blocking using PowerShell or REST/SDK calls—an approach that does not scale and increases the chance of human error.
Flag: a specific, widely-reported number—“107 Copilot agents” introduced during a May rollout—appears in investigative threads but is not confirmed in Microsoft’s public release notes. Treat that figure as an investigative assertion rather than a canonical platform count until a tenant inventory export validates it.
Audit telemetry gap: missing resource references in Purview
Separately, security researchers reproduced a trivial prompt variation that caused Copilot to summarize a document while the corresponding Purview audit record omitted the AccessedResources/file-reference attribute that would normally tie the response back to the underlying file. The practical effect is an audit blind spot: Copilot presented a summary derived from a tenant file, but standard Purview-driven SIEM correlations and eDiscovery playbooks could not link the action to the file. Microsoft acknowledged and applied server-side remediation for the observed case, but the incident underlines the fragility of relying exclusively on vendor telemetry without compensating signals.Sandbox misconfiguration: the pgrep path-hijack
Independent research showed how Copilot’s live Python/Jupyter sandbox could be provoked such that a writable path earlier in $PATH allowed a crafted file to shadow a system binary (for example, a local pgrep), enabling root execution inside the container. Eye Security’s write‑up demonstrates the mechanics: an unqualified command invocation combined with writable directories created a classic search-path hijack. Microsoft patched the environment following responsible disclosure, but the finding illustrates how simple configuration details can yield powerful privileges inside otherwise isolated sandboxes.Why this matters to enterprises
Short paragraphs aid scanning—here are the real-world risks:- Data exfiltration by stealthy agents. If an agent that should be blocked stays discoverable, it can perform semantic retrievals from SharePoint, OneDrive, or Exchange and return excerpts or summaries to users who shouldn’t have that access. Where agents expose “search” or “export” features, the consequence can be direct leakage.
- Shadow automation and uncontrolled actions. Agents tied to Power Platform workflows or RPA can execute actions (move files, call connectors, run flows) outside change-control processes if non‑admin users can invoke them. That creates operational and legal risk—particularly for regulated data.
- Audit, compliance, and forensics erosion. Missing AccessedResources attributes in Purview events can void the chain-of-evidence that compliance teams and incident responders rely on. A missing audit link is not a minor inconvenience; it can materially extend investigation timelines or fail legal discovery requirements.
- Operational overhead and human error. The temporary workaround—per‑agent PowerShell revocation—means large tenants must enumerate agents, run revokes one-by-one, and maintain an auditable list of actions. That process is slow, manual, and error-prone.
- Downstream escalation risk from sandbox root. Although container root does not necessarily mean host escape, root inside a sandbox enables telemetry tampering, local log modification, and lateral research that can compound other failures if additional misconfigurations exist. The pgrep issue underscores brittle assumptions in sandbox design.
Microsoft’s response and timeline (verified items)
- Microsoft documented agent governance surfaces and encouraged Conditional Access and DLP as primary compensating controls for agent behavior. The corporate docs describe admin‑controlled agent installation modes (Microsoft-installed, admin-installed, user-installed) and point administrators to the Copilot Control System in the Microsoft 365 admin center. (learn.microsoft.com, microsoft.com)
- Microsoft applied server-side fixes for at least two of the issues reported by researchers: the Purview audit emission gap and the Jupyter/live‑Python sandbox path‑hijack. Those remediations were applied without tenant action required in the documented cases. Microsoft classified the sandbox fix at a moderate importance level and acknowledged the reporting researchers publicly. (research.eye.security, learn.microsoft.com)
- Microsoft added—or announced—administrative controls to block agents created in Copilot Studio and provided PowerShell and REST/SDK pathways for quarantining agents; feature-message channels describe general availability timelines for these admin controls (PowerShell cmdlets to block/unblock agents became available in mid‑2025). However, until those controls were fully available and validated in every tenant, administrators still needed per‑agent workarounds in many environments. (app.cloudscout.one, practical365.com)
Technical anatomy: why enforcement can fail
The problem is rarely a single bug. Analysis and incident reports point to at least three interacting platform characteristics that together create the observed failure modes:- Control-plane desynchronization. Inventory and UI surfaces that list agents in admin centers may not be synchronized with the enforcement code path that determines runtime discoverability, creating race conditions where an agent appears to users even when policy state should hide it.
- Privileged provisioning or publisher differentiation. Microsoft‑published agents and agents provisioned through privileged flows may be routed through backend channels that do not evaluate the same tenant-level UI checks, yielding inconsistent behavior across product surfaces.
- Policy semantics vs. hard-deny enforcement. Some admin toggles are implemented as scoping hints or UX-layer filters instead of global, platform-wide prohibitions. If an admin control is applied only at the UI layer, other surfaces or privileged provisioning paths can still expose the agent.
- Telemetry and UI divergence. UI rendering paths can diverge from audit emission paths; suppressing a link in the user-facing response may inadvertently short-circuit the code path that emits AccessedResources metadata into Purview, creating audit gaps.
Practical, prioritized mitigation checklist for administrators
The following is a prioritized list designed for busy security teams to regain control and reduce risk quickly. Each step is practical and grounded in platform capabilities.Immediate triage (0–48 hours)
- Export and reconcile the Copilot Agent inventory from the Microsoft 365 admin center and Copilot Control System. Treat any unknown publisher or agent as high priority for investigation; create a quarantine list.
- Reproduce enforcement checks from representative non‑admin accounts (including guest profiles where applicable) across web, Teams, Outlook, and mobile surfaces. Document any discrepancies (screenshots, timestamps) and open support tickets with Microsoft if tenant behavior contradicts admin settings.
- If your tenant sees agents that should be blocked but remain discoverable, apply per‑agent blocking via the documented admin PowerShell or REST/SDK controls (the Copilot Studio admin block command and REST API were announced as admin options; use these to temporarily quarantine agents until the platform-side fixes are validated). (app.cloudscout.one, mwpro.co.uk)
Short-term hardening (48 hours – 2 weeks)
- Enforce Conditional Access for Copilot and AI services: require phishing‑resistant MFA (FIDO2/Windows Hello for Business), require compliant devices, and apply session controls for high-sensitivity roles. Conditional Access is a recommended compensating control in Microsoft’s guidance.
- Apply Purview sensitivity labels and DLP to restrict Copilot’s ability to process high‑sensitivity content. Where feasible, restrict Copilot access to high‑sensitivity repositories and require approval gates for HR, legal, or regulated data. Validate the effect by simulating Copilot queries in a safe, controlled tenant. (learn.microsoft.com, lewisdoes.dev)
- Harden maker and uploader privileges in Copilot Studio: restrict who can publish agents, require connector approvals, and apply connector management policies to limit inbound/outbound data flows.
Detection and monitoring (2–6 weeks)
- Correlate Purview CopilotInteraction/AIAppInteraction records with Microsoft Graph read counters, SharePoint/OneDrive access logs, and agent invocation logs in your SIEM. Create detection rules for anomalous agent usage, unexpected cross‑connector retrievals, or sudden spikes in agent-driven exports.
- Simulate the “link suppression” edge case in a non‑production tenant: ask Copilot to “summarize without including a link” and confirm that a complete AccessedResources event is generated. If not, augment detection with secondary signals (file read counters, API calls).
- Route Purview audit logs to immutable storage or a hardened SIEM export to preserve evidence continuity in case of telemetry tampering attempts. Microsoft documents pay-as-you-go and retention controls for copilot/audit exports—configure these according to your compliance needs.
Operational controls and process (ongoing)
- Maintain an auditable register of agent revocations and a playbook for per‑agent blocking/unblocking. Automate agent inventory reconciliation and scheduled checks (daily/weekly) so manual steps scale. Community tooling and scripts exist to help automate uninstall/unregister workflows; vet any third‑party scripts carefully before running them in production.
- Harden oversight consoles (Responsible AI Operation panels etc.) with strict admin separation, vaulted credentials, and dedicated monitoring. Treat any console that can alter telemetry or enforcement rules as a high-value target.
Risk assessment: strengths, gaps, and the governance deficit
Notable strengths
- Microsoft’s architecture layers (containerized sandboxes, network constraints, and centralized telemetry) reflect modern, defense-in-depth thinking. When some vulnerabilities were responsibly disclosed, Microsoft issued server-side remediations reasonably quickly. These are meaningful operational facts for defenders. (research.eye.security, learn.microsoft.com)
- Microsoft has incrementally released admin tooling (PowerShell, REST/SDK, Copilot Control System) and guidance (Conditional Access, DLP, connector policies) to let tenants apply compensating controls while platform fixes are validated. (learn.microsoft.com, microsoft.com)
Significant gaps and risks
- The platform’s multi-surface complexity (web, Teams, Outlook, Studio) combined with privileged provisioning paths increases the likelihood of policy intent ≠ runtime reality—a governance deficit, not just a coding bug.
- Telemetry divergence—the possibility that UI decisions can suppress or alter audit emissions—breaks a core compliance assumption: that every material read by an AI assistant produces a durable, attributable audit record. That failure has outsized consequences for regulated sectors.
- Manual per‑agent remediation is operationally expensive at scale. Large tenants must reconcile hundreds (or more) of agents across multiple work surfaces; human fatigue and process drift increase the probability of missed blocks or incomplete remediation.
Longer-term implications and recommendations for vendors
- Vendors must treat admin toggles as authoritative enforcement primitives, not as UX scoping hints. A tenant’s security posture depends on the platform honoring explicit deny semantics across all product surfaces.
- Telemetry must be robust and independently verifiable. Relying solely on vendor-supplied audit streams without immutable export or cross-correlation fosters systemic blind spots. Vendors should provide explicit guarantees about which telemetry attributes are emitted for every data access and make those guarantees programmable and testable.
- Disclosure and bounty practice matters. Researchers who perform coordinated disclosure and produce reproducible proofs-of-concept should be incentivized responsibly; classification and payout choices influence the likelihood and timeliness of vulnerability reporting. The Eye Security path-hijack case illustrates both the value of independent research and the debate around severity classification.
Conclusion
The Copilot enforcement and telemetry failures form a compact case study in modern cloud‑AI risk: user-facing convenience features, paired with complex multi-surface rollouts and multiple backend provisioning paths, create unanticipated governance gaps. The technical fixes (sandbox patches, admin PowerShell controls, improved telemetry) are necessary and meaningful, but the enterprise takeaway is operational and organizational as much as technical: assume that vendor controls can fail, validate them regularly, and put layered compensating controls in place—Conditional Access, DLP, immutable audit exports, agent inventory reconciliation, and per‑agent quarantine processes.Administrators should treat Copilot agents like any other third‑party application in their environment: insist on explicit deny semantics, test enforcement from representative user contexts, require phishing‑resistant authentication for AI services, and maintain cross-correlated telemetry pipelines that do not rely on a single vendor-supplied signal. Where vendor fixes are in flight, take the manual, time‑consuming steps required now—because the cost of not doing so is not merely a productivity nuisance; it is a governance and compliance liability that can be measured in exposed data, failed audits, and extended incident response timelines. (cybersecuritynews.com, learn.microsoft.com)
Source: Digital Watch Observatory Copilot policy flaw allows unauthorized access to AI agents | Digital Watch Observatory