• Thread Author
Microsoft’s push to weave Copilot into the fabric of Microsoft 365 has hit a trust-defining snag: for months, under specific prompting conditions, the AI assistant’s access to source documents could be absent from Microsoft 365 audit logs, leaving security teams with empty entries where definitive reads should have been. The behavior, first observed in early July 2025 and traced back to earlier warnings in mid‑2024, was quietly corrected only recently and without customer notification, reigniting questions about how Microsoft classifies, communicates, and remediates cloud vulnerabilities that directly impact compliance and incident response.

A man monitors holographic blue dashboards and a glowing globe on a desk.Overview​

Copilot for Microsoft 365 has become the marquee feature of Microsoft’s AI strategy, promising to summarize long reports, draft emails, and extract insights from SharePoint, OneDrive, Teams, and Exchange content. In regulated environments, this convenience rests on a bedrock assumption: every content access—human or machine—must be logged with enough fidelity to support forensics, compliance audits, and insider‑risk inquiries.
The newly exposed flaw undermined that assumption. If a user asked Copilot to summarize a document and specifically instructed the assistant not to include a link back to the source, the audit trail could show an empty or incomplete log entry rather than a clean, attributable “read” of the file. That subtle omission created a gap large enough to frustrate investigations and potentially mask misuse.
The episode culminates a year of simmering concern about AI observability in enterprise suites. It also puts a harsh spotlight on Microsoft’s security response playbook—particularly the decision not to inform customers and the stance that the issue did not merit a CVE, despite earlier public commitments to assign identifiers to cloud flaws even when customers need take no action. For organizations leaning on Copilot to transform knowledge work, the lesson is immediate and sobering: AI features must be held to the same auditability standards as any enterprise app, and exceptions cannot hide in prompt‑driven edge cases.

Background​

Microsoft 365 audit logs are the canonical record for who accessed what, when, and how. They underpin:
  • Regulatory obligations (SOX, HIPAA, GDPR, PCI DSS, ISO/IEC 27001).
  • Incident response workflows and forensic timelines.
  • Insider‑risk investigations and legal holds.
  • Continuous monitoring policies in zero‑trust architectures.
Copilot interacts with tenant data through Microsoft Graph and service‑side retrieval mechanisms to ground its responses in the user’s permissions and context. In principle, this should generate auditable signals equivalent to a human read. In practice, the flaw allowed at least one pathway—summarization without linking—to yield incomplete entries, contradicting expectations that every material read leaves a durable trace.
This isn’t merely an academic nuance. Audit record integrity is the difference between proving that a sensitive file was accessed at 11:03 AM by a specific identity and arguing from circumstantial analytics that “something” probably happened. When the records are wrong—or missing—regulatory exposure and investigation cost escalate rapidly.

What happened and why it matters​

The discovery and the timeline​

  • Early July 2025: A startup CTO, Zack Korman, notices that certain Copilot prompts produce empty or incomplete audit entries in Microsoft 365—despite clearly deriving information from a specific document. His test case: ask Copilot to summarize a file while instructing it not to include a link to the document in the response.
  • Within days: The Microsoft Security Response Center (MSRC) begins reproducing the behavior. By July 10, Korman is informed that a fix has been rolled out—seemingly ahead of a formal disclosure process and without notification to affected customers.
  • August 2: Microsoft communicates that a broader cloud update will complete mid‑August, with publication timing discussed but a CVE seemingly off the table because customers don’t need to take action.
  • August 14–17: Microsoft confirms its decision not to notify customers and proceeds with the Microsoft 365 update cadence.
Korman’s tests indicate that the issue had existed for longer than initially suspected, intersecting with a 2024 Black Hat presentation by security researcher Michael Bargury that highlighted weaknesses in how AI‑mediated file accesses were reflected in Microsoft 365 logging. If accurate, that chronology implies months of potential logging deficiencies—more than enough time for prompt‑savvy insiders or attackers to learn and abuse the edge case.

Why this is different from “just another bug”​

  • It targets the audit layer, not the application feature set. When audit data becomes unreliable, security teams lose the authoritative source of truth they rely on during breaches, legal discovery, and regulatory audits.
  • The triggering condition is a prompt, not an exploit chain. That means it could be invoked by an otherwise normal user during legitimate work, intentionally or accidentally, without tripping classical intrusion alarms.
  • The operational impact is hidden. There are no error banners, failed tasks, or broken UX—just missing or empty entries in systems that few end users see, but which every security program depends on.

How the flaw manifested​

The problematic prompt pattern​

  • A user asks Copilot to summarize a known file (for example, “Summarize the Q2 2025 annual report”).
  • The user adds a constraint: do not include a link to the source in the answer.
  • Copilot retrieves content to produce the summary, but the Microsoft 365 audit trail records an empty or incomplete entry for the underlying file read.
The scenario points to a divergence between two internal pathways:
  • A “show your work” path that composes an answer and includes a source link, emitting a clear “read” event tied to the file and the requesting identity.
  • A “no links” path that still grounds on the file’s content but fails to emit the same audit signal, resulting in a log artifact that is ambiguous or blank.
From a security lens, both paths are equivalent in sensitivity: content was accessed and used to satisfy a user’s request. The audit record should be equally explicit in both cases. The inconsistency is what created the blind spot.

What this likely means in practice​

  • If an insider wanted to learn from a restricted file without leaving an obvious breadcrumb, this prompt nuance may have reduced their audit exposure.
  • Investigators reviewing a suspected data leak could find Copilot‑authored summaries referencing confidential content yet be unable to tie those summaries to a definitive “read” event. That complicates timeline reconstruction and attribution.
  • Compliance teams validating control effectiveness may have erroneously concluded that certain content wasn’t accessed during test windows, when in fact it was.

Microsoft’s response under the microscope​

Classification and CVE policy​

Microsoft reportedly classified the issue as “Important,” below the “Critical” threshold it uses to assign CVE identifiers to cloud flaws that require no customer action. This is contentious for two reasons:
  • The company had publicly committed to issuing CVEs for certain cloud vulnerabilities even when customer action is not required, as a transparency and tracking mechanism.
  • Audit log integrity is itself a security control. A flaw that degrades it undermines detection and response, and arguably merits broader disclosure and traceability.
Whether this case met Microsoft’s internal criteria is ultimately Microsoft’s call—but the decision not to notify customers is hard to square with the scale of potential operational impact.

Communication and the Secure Future Initiative​

After a series of cloud security stumbles, Microsoft unveiled a “Secure Future Initiative” to improve product security, operational rigor, and communication. In this case, the optics cut against those goals:
  • Silent or low‑visibility remediation.
  • No customer notification for an issue that affects compliance posture.
  • Disagreement between researcher expectations and the vendor’s stated process.
The result is a widening trust gap. Enterprises don’t just need bugs fixed; they need to know that a bug existed, understand its scope, and assess whether it affected them. Without that signal, they cannot fulfill their own obligations to regulators, auditors, and customers.

The compliance and risk implications​

Auditability is a control, not a convenience​

Many frameworks explicitly require detailed access logging:
  • SOX: Evidence for change control and access to financial reporting systems.
  • HIPAA: Access logs for protected health information.
  • GDPR: Accountability and record‑keeping for processing activities.
  • ISO/IEC 27001: Operations security and event logging control objectives.
If Copilot’s accesses were mislogged in some cases, organizations may have:
  • Underreported or mischaracterized access during audits or incident reports.
  • Lost the ability to conclusively attribute actions during investigations.
  • Increased their exposure in litigation due to weaker evidentiary trails.

Insider risk and industrial espionage​

An employee with legitimate access to Copilot but not to certain files could use AI features to synthesize information from restricted documents if the guardrails are weak. If audit logs fail to reflect these interactions, insider‑risk detections—often dependent on behavioral correlation—lose a key signal.

Threat actor playbooks​

Since at least August 2024, conference talks and community research have discussed ways AI interfaces can blur audit trails. It is reasonable to assume that sophisticated attackers tested or adopted similar prompt patterns. Even if exploitation wasn’t widespread, the mere possibility raises the bar for due diligence in post‑incident reviews.

Technical analysis: where might the logging gap live?​

Caveat: The following analysis discusses plausible mechanics based on how Microsoft 365 services and AI grounding typically work. Exact internal details remain proprietary.

Dual‑path content grounding​

  • With link: Copilot retrieves content, constructs a citation, and surfaces a hyperlink back to the document. This flow likely calls a service boundary that generates both user‑context read events and application telemetry, resulting in auditable “FileAccessed” or equivalent operations.
  • Without link: Copilot still retrieves content but refrains from producing a citation. If the citation generation were the trigger for writing a canonical “read” audit record, skipping it might have suppressed or altered the event emission. Alternatively, the no‑link path could have used a different retrieval method that logged less detail or recorded it under a different operation name not exposed in standard searches.

Service‑side summarization and event correlation​

Copilot blends client orchestration with service‑side summarization. If the “no link” path confined some operations to transient service layers, audit events could have been tied to an opaque application identity or lost in aggregation. The visible result would be an empty or ambiguous audit record that fails to bind the file, the user, and the time in a single, queryable line.

Consequences for downstream tools​

  • Microsoft Purview Audit (Standard/Premium) relies on consistent operation schemas. Divergent paths break parsers and dashboards.
  • SIEM and XDR rules ingest “FileAccessed,” “SearchQueryPerformed,” “MessageGenerated,” and similar operations. If Copilot emitted nonstandard or partial events during “no link” summarization, correlation rules would miss them.
Even if Microsoft’s July and August updates normalized these differences, the historical data gap persists.

What Microsoft 365 administrators should do now​

1. Establish the window of concern​

  • Identify when Copilot features were enabled in your tenant and when the mid‑August 2025 updates reached your regions. This anchors the bracket for log review.
  • If you adopted Copilot in late 2024 or early 2025, consider a 9–12 month retrospective to capture the period discussed by independent researchers.

2. Turn on the richest logs you can​

  • Ensure Microsoft Purview Audit is enabled at the highest available tier for your licensing (Premium if possible) to capture extended properties and longer retention.
  • Validate retention settings: many tenants default to 90 days for standard audit; Premium extends this significantly. Confirm your retention meets investigative timelines.

3. Hunt for “empty access” patterns​

Use multiple pivots, because the missing audit entry may not present as a simple “FileAccessed.”
  • Look for Copilot activities (“AI‑generated,” “MessageGenerated,” or product‑specific operations) that reference or summarize known sensitive documents without a corresponding file read in a reasonable time window.
  • Correlate user sessions: Entra ID sign‑in logs → Copilot/Office 365 service principals → application activities → file access. A gap between application response and file read is a red flag.
  • Compare Teams or Outlook messages that contain Copilot summaries with SharePoint/OneDrive access logs for implicated files.

4. Tighten AI‑data controls​

  • Enforce sensitivity labels and encryption on confidential files and ensure Copilot respects label policies.
  • Review tenant‑level Copilot controls, including scoping for high‑risk sites and ongoing public preview features that may alter logging behavior.
  • Revisit Conditional Access policies for “app‑enforced restrictions” and “use terms of use” to remind users of acceptable AI usage.
  • Consider Microsoft Defender for Cloud Apps (App Governance) to monitor and alert on unusual AI application telemetry patterns.

5. Strengthen insider‑risk analytics​

  • Add detections for content summarization of high‑value files, not just downloads or shares.
  • Weight “AI‑assisted data discovery” as a signal in insider‑risk models, particularly when paired with anomalous sign‑ins or newly elevated privileges.

6. Document and disclose as needed​

  • Record your findings, the potential impact window, and the steps taken to mitigate. If you operate in regulated sectors, consult counsel on whether notifications are appropriate.
  • Update your risk register: “AI auditability gap” should be a discrete item with owners and timelines.

Practical queries and checks​

The specific operation names and schemas in Microsoft 365 can vary by product and license, but the following patterns help you find gaps.

PowerShell (Unified Audit Log)​

  • Start by confirming audit is on:
  • Get‑AdminAuditLogConfig
  • Enumerate operations related to file access and AI activities around a time window:
  • Search‑UnifiedAuditLog -StartDate "2025‑05‑01" -EndDate "2025‑08‑20" -Operations FileAccessed, FilePreviewed, FileAccessedExtended, SearchQueryPerformed, MessageGenerated -ResultSize 5000
  • Pivot on users with high Copilot usage:
  • Filter for records where the UserId or UserKey matches those users and inspect whether file reads exist near Copilot activity timestamps.
Export the results to CSV and join on TimeGenerated and UserId to locate summaries without reads.

Microsoft Defender (Advanced Hunting with KQL)​

  • Copilot or service‑principal‑centric patterns:
  • Query AppEvents or CloudAppEvents for operations attributed to Copilot service principals.
  • SharePoint/OneDrive correlation:
  • Join SharePointFileOperation with application events on UserId and TimeGenerated in a ±5 minute window.
  • Flag gaps:
  • Identify application events that reference specific sites or document IDs without a matching FileAccessed event.
Even if naming varies, the principle is constant: find AI activity that plausibly required a read, then prove whether the read is recorded.

Governance: update your AI acceptance criteria​

Require “full‑fidelity audit” as a hard gate​

When approving AI features or pilots:
  • Insist that every data retrieval by the AI assistant produces a standard audit event equivalent to a human read.
  • Require that “no link” or “anonymized” responses still log the underlying access in a consistent, queryable format.

Mandate negative‑testing in proofs of concept​

  • Test prompts designed to suppress citations, strip sources, or request paraphrase‑only outputs.
  • Verify that logging remains complete across those cases, and retain evidence from test runs.

Build SLAs for transparency​

  • Include contractual obligations for vulnerability disclosure in AI features, including:
  • Timely customer notification.
  • Clear severity classification.
  • Tenant‑specific impact statements or mitigation guidance when feasible.

The broader pattern: AI features, enterprise controls, and trust​

Enterprise AI is racing ahead, and vendors rightly evangelize productivity gains. But every new AI capability touches long‑standing controls in subtle ways:
  • Data minimization versus context‑rich grounding.
  • Explainability versus user experience.
  • Audit verbosity versus cost and performance.
This incident reveals how a small UX choice—whether to include a link—can ripple into a control failure if logging is tied to presentation rather than the underlying data access. It also shows why security teams must add “AI‑path parity” to their validation checklist: any functional variation (summarize, rewrite, translate, extract) that touches sensitive content must yield the same audit fidelity.
Microsoft is not alone in grappling with this. Every major productivity suite adding generative AI faces the same challenge: build features that feel magical while maintaining the unglamorous, nonnegotiable plumbing of enterprise security. The organizations that thrive will be those that make audit and transparency first‑class citizens of their AI architectures.

Questions Microsoft still needs to answer​

  • Scope: Which Copilot scenarios and apps were affected (Word, Excel, PowerPoint, Teams, Outlook, SharePoint) and for how long?
  • Signals: Which operation types were suppressed or altered? Can customers query historical data to infer impacted events?
  • Tenants: Were any geographies, cloud instances (Commercial, GCC, GCC High), or licensing tiers disproportionately affected?
  • Detection: Will Microsoft deliver retroactive detections or advisories to help customers find likely exposures?
  • Policy: How does this case align with the company’s commitment to assign CVEs to cloud vulnerabilities where customers need take no action?
  • Process: What changes will ensure that audit logging is validated across all prompt variants, including “no citation” and “private mode” experiences?
Providing clear, tenant‑impactable answers would convert a reactive fix into a durable trust‑building moment.

What “good” should look like for AI auditability​

Principled requirements​

  • Event parity: Every data access by AI produces the same audit event as a human read, regardless of prompt phrasing or UI options.
  • End‑to‑end traceability: Each AI response is traceable to the identity, source files, time, and scope used to generate it, with optional privacy‑preserving redactions for user display but not for audit storage.
  • Tamper‑evident logs: Audit records are immutable, cryptographically anchored, and retained long enough to support investigations.
  • Customer‑visible schemas: Operation names and properties are documented and stable, enabling SIEM correlation and third‑party analytics.
  • Disclosure discipline: When audit fidelity is degraded, customers are promptly notified with clear remediation guidance.

Practical enhancements vendors can deliver​

  • A “Show full audit context” button for administrators that reveals the files and scopes used to generate a Copilot answer, even if the end‑user view omits links.
  • An “AI access parity” policy toggle: When enabled, Copilot responses are blocked if the system cannot emit a canonical audit event for the underlying retrieval.
  • Built‑in hunts and workbooks in security portals that detect AI‑generated responses referencing sensitive locations without paired read events.
  • A compliance mode that refuses prompts which would produce unverifiable or partially auditable outputs.

The path forward for enterprise teams​

This incident does not argue against AI assistants in the enterprise. It argues for disciplined engineering and governance that keep classic controls intact as the user experience evolves. For security and compliance leaders, the agenda is clear:
  • Re‑validate that your Microsoft 365 audit pipeline captures all AI‑mediated access patterns, including edge‑case prompts.
  • Tighten data classification and label‑based access rules so Copilot cannot silently aggregate from unlabeled, high‑risk repositories.
  • Establish an AI change‑management board that treats new features like any line‑of‑business app rollout, with explicit test plans and go/no‑go criteria focused on auditability.
  • Update your incident response runbooks to include AI‑specific hunts and “audit gap” playcards.
  • Push for stronger commitments from vendors—beyond marketing language—to codify audit fidelity, disclosure practices, and tenant‑specific impact assessments.

Conclusion​

Copilot’s audit‑logging flaw is a reminder that security is not compromised only by breaches and buffer overflows; it can be compromised by the invisible erosion of accountability. When a single prompt modifier can move an operation off the main logging highway, the risk is not just technical—it is organizational. Investigations misfire, compliance attestation weakens, and trust in shared cloud controls frays.
Microsoft has reportedly fixed the behavior, but the episode exposes two gaps that only customers can close: an architectural gap, where AI experience variants must produce identical, high‑fidelity audit events; and a governance gap, where vendors must proactively notify customers when controls like auditing falter. Enterprises should treat this as a catalyst to harden their AI observability strategies, re‑baseline Copilot’s behavior in their tenant, and demand process transparency that matches the ambition of the AI features they are being asked to trust.

Source: heise online AI assistant: Microsoft's Copilot falsified access logs for months
 

Back
Top