When the AI Reads the Mail It Was Never Supposed to See: The Microsoft 365 Copilot Confidential Email Incident

In late January 2026, something quietly broke inside Microsoft 365. For weeks, the company’s flagship AI productivity assistant, Copilot, was reading and summarizing emails that organizations had explicitly labeled as confidential and barred from automated processing. It did not break down doors to do it. There were no stolen credentials, no external attacker, no malware. The problem was a software defect buried in Copilot’s own retrieval pipeline, and it meant that data loss prevention policies, sensitivity labels, and the governance controls that enterprise customers had spent considerable time and money configuring were, for a period of weeks, simply being ignored.

The incident, tracked internally by Microsoft under service advisory CW1226324, has since been patched. But the technical fix resolves only the narrowest version of the problem. The broader questions it raises — about AI governance, the reliability of enterprise data controls when AI enters the stack, and the pace at which organizations are deploying intelligent assistants without the governance frameworks to match — remain wide open.

This article examines what happened, why it matters, what organizations should do in response, and what the incident signals about the future of enterprise AI security.

Understanding Microsoft 365 Copilot and How It Is Supposed to Work

Microsoft 365 Copilot is an AI-powered productivity assistant embedded across Outlook, Word, Excel, PowerPoint, Teams, and OneNote. It became available to paying Microsoft 365 business customers in September 2025. As of January 2026, Microsoft reported 15 million paid Microsoft 365 Copilot seats across its 450 million Microsoft 365 commercial subscriber base. Its core value proposition is simple: give employees a natural-language interface to their organizational data, so they can get summaries of email threads, draft documents, find information across SharePoint, and ask questions without leaving their workflow.

To deliver that experience, Copilot uses a retrieval-augmented generation (RAG) architecture. This means it does not rely solely on its training data to answer questions. Instead, it first retrieves relevant content from the user’s organizational environment — emails, documents, Teams chats, SharePoint files — using Microsoft Graph, and then feeds that retrieved content into a large language model (LLM) to generate a response. The quality and safety of the experience depend entirely on which content gets retrieved and whether the right guardrails are applied before any content reaches the model.

That retrieval layer is where enterprise governance controls are supposed to operate. Microsoft Purview sensitivity labels allow administrators to classify content — applying designations such as Confidential, Highly Confidential, or Internal — and configure enforcement actions that travel with the data. When a sensitivity label is applied to an email, it can trigger encryption, restrict forwarding, and, critically for AI use cases, instruct automated tools including Copilot to exclude that content from summarization and processing.

Data Loss Prevention (DLP) policies, configured through Microsoft Purview, add a second layer of enforcement. DLP policies can be configured specifically to prevent Copilot from accessing content bearing certain sensitivity labels. According to Microsoft’s own documentation, when a DLP policy is in place, Copilot and Copilot Chat will not summarize protected content, though it may still surface a link to the item so the user can open it manually. The system is designed so that the AI acts as a productivity accelerator while fully respecting the organization’s data governance posture.

That is what was supposed to happen. What actually happened in late January 2026 was something different.

What Went Wrong: The Timeline and Technical Failure

On January 21, 2026, customers began reporting anomalous behavior in Microsoft 365 Copilot Chat. Specifically, the Copilot ‘work tab’ chat experience — the interface through which users interact with organizational content — was returning summaries that drew on emails stored in users’ Sent Items and Drafts folders, even when those emails had sensitivity labels applied and DLP policies configured to block exactly that kind of access.

Microsoft’s own service alert, later obtained and reported by BleepingComputer, stated the issue plainly: ‘The Microsoft 365 Copilot work tab Chat is summarizing email messages even though these email messages have a sensitivity label applied and a DLP policy is configured.’ The company attributed the cause to ‘a code issue’ that was ‘allowing items in the sent items and draft folders to be picked up by Copilot even though confidential labels are set in place.’

The technical failure appears to have occurred in the integration layer between Copilot’s retrieval pipeline and Microsoft Purview’s enforcement framework. In a correctly functioning system, label-based exclusions and DLP rules act as a gate: the retrieval pipeline checks whether a content item is protected before including it in the context that gets passed to the language model. When that gate fails — whether due to a logic error in when the label check runs, a caching defect that doesn’t propagate label status correctly, or a misconfiguration in how the pipeline handles specific mailbox folders — the model receives content it was never supposed to see, and produces summaries from it.

The scope of the bug was specific but consequential. Items in the Inbox and other standard folders appeared to be unaffected. The failure was confined to Sent Items and Drafts. But that specificity does not reduce the severity. Sent Items and Drafts are precisely the folders that contain the most sensitive organizational content: outbound communications about ongoing negotiations, pricing decisions, legal strategy, pending transactions, security incidents, personnel matters, and unfinished documents that have not yet been reviewed or approved for distribution. The impact was significant enough that the UK’s National Health Service flagged the issue internally, logging it as incident INC46740412 — a signal of how far the failure reached into regulated healthcare environments.

Microsoft began rolling out a server-side fix in early February 2026. By February 20, the company updated its service alert to state that the root cause had been addressed and that the targeted code fix had ‘saturated across the majority of affected environments.’ A full remediation confirmation was expected by February 24, 2026. Throughout the incident, Microsoft maintained that the bug ‘did not provide anyone access to information they weren’t already authorized to see,’ framing it as a policy-enforcement deviation rather than an access-control breach. In a statement to The Register, a Microsoft spokesperson said: ‘We identified and addressed an issue where Microsoft 365 Copilot Chat could return content from emails labeled confidential authored by a user and stored within their Draft and Sent Items in Outlook desktop.’

That framing is technically accurate. Copilot only surfaced content within the signed-in user’s own account. No other user gained access to someone else’s mailbox. But security and compliance professionals have immediately and correctly pushed back on this characterization, because it misses the point of DLP controls entirely.

Why ‘No One Saw Your Emails’ Is Not the Right Reassurance

Microsoft’s official statement — that access controls and data protection policies ‘remained intact’ and that the bug did not grant unauthorized parties access to anyone’s mailbox — is correct as far as it goes. No external attacker read anyone’s confidential email. No competitor gained access to a pricing strategy. The content stayed within the permission boundary of the account’s owner.

But sensitivity labels and DLP policies exist precisely to restrict automated processing, not just to control human-to-human access. Organizations apply the Confidential label to content because they want to prevent that content from being ingested, indexed, summarized, or otherwise processed by automated systems — including AI assistants built by the same company that hosts their email. The DLP policy for Copilot is not designed to keep other users out of a mailbox. It is designed to keep Copilot itself from processing certain content. When that policy fails, the control fails, regardless of whether any additional human accessed the data.

There is also the question of AI-generated output as a new exposure vector. When Copilot summarizes a confidential email, the summary itself becomes a new artifact. If that summary is copied into a shared document, pasted into a Teams message, or used as the basis for a draft response that is then forwarded — all things Copilot is designed to facilitate — the confidential information has now traveled beyond its intended boundary. DLP policies guard the original; they do not automatically govern AI-generated derivatives of that original.

For organizations operating in regulated industries — healthcare (HIPAA), financial services (GLBA, SOX), legal services (attorney-client privilege), and government (FISMA, FedRAMP) — the stakes are higher still. These organizations use DLP controls not just as internal best practice, but as a legal and regulatory requirement. When an AI system bypasses those controls, even within a user’s own permission boundary, it may constitute a compliance failure that requires notification, documentation, and potentially regulatory reporting.

The Broader Pattern: AI and the Inconsistency of Label Enforcement

What makes the CW1226324 incident particularly instructive is not that it happened, but that it reveals an architectural challenge that extends well beyond this single bug. Microsoft’s own documentation acknowledges that sensitivity labels do not function identically across all Microsoft 365 applications. The same label applied to an email in Outlook may be enforced differently in Teams, in Microsoft 365 Copilot Chat, or in other integrated tools. Administrators who assume uniform enforcement across all surfaces are operating on an incorrect premise.

This inconsistency is not new, but AI amplifies its consequences dramatically. In a traditional software environment, a label-enforcement gap might mean that a user could see a document they should not have been able to open. The impact, while concerning, is bounded by what that individual user does with the information. In an AI-assisted environment, the same gap can result in the AI synthesizing and repackaging restricted content at scale, across every user in the tenant, every time they interact with Copilot Chat. What was a narrow exposure surface becomes a broad one.

Security researchers analyzing the incident have also noted a pattern of prior incidents. VentureBeat reported that Microsoft Copilot has now ignored sensitivity labels in multiple separate incidents within an eight-month period, and that in neither case did existing EDR (endpoint detection and response) or WAF (web application firewall) tools produce alerts. The earlier incident, disclosed and patched in June 2025, was a critical zero-click vulnerability known as EchoLeak (CVE-2025-32711, CVSS 9.3), discovered by researchers at Aim Security. In that case, a malicious email could manipulate Copilot’s retrieval-augmented generation pipeline into accessing and transmitting internal data to an attacker-controlled server — without any user interaction. The two incidents had fundamentally different root causes — an adversarial exploit chain versus an internal code defect — but produced the same consequence: data the system was not supposed to process was processed, and the tenant security stack did not generate an observable alert. Retrospective detection in both cases depended entirely on Microsoft Purview telemetry — meaning organizations without comprehensive Purview logging would have had no visibility into the exposure at all.

Separately, on February 17, 2026, the European Parliament disabled built-in AI features on work devices used by lawmakers and staff, citing unresolved concerns about data security and privacy. The Parliament’s IT department concluded it could not guarantee the security of data processed by cloud-based AI tools, including email summarization and virtual assistant features. The timing — coinciding with the CW1226324 disclosure — illustrates that institutional confidence in AI governance controls is eroding in real time, and that organizations handling highly sensitive information are choosing precautionary restrictions over continued exposure.

A 2026 survey by Cybersecurity Insiders and Saviynt, based on responses from 235 CISOs and senior security leaders at large enterprises, found that 47% have already observed AI agents exhibit unintended or unauthorized behavior. That figure reflects a larger truth: the governance frameworks for enterprise AI have not kept pace with the speed of deployment. Organizations are installing AI productivity assistants that have broad retrieval access to their most sensitive data, in some cases before they have established the monitoring, audit procedures, and policy controls necessary to govern them safely.

What Organizations Should Do Now

The patch is rolling out. But administrators and security teams cannot simply move on. This incident has surfaced a set of operational requirements that should be addressed regardless of whether any individual organization was directly affected.

Audit and Forensics

Organizations should run a targeted review of Microsoft Purview audit logs covering the period from January 21 through mid-February 2026. Specifically, look for Copilot Chat interactions in the work tab that returned content from messages labeled Confidential, or that referenced items stored in Sent Items and Drafts folders. Export metadata — sender, recipients, timestamps, and content references — and preserve it under legal hold.

If your logging posture was incomplete during the exposure window, document that gap formally. For organizations subject to regulatory oversight, an undocumented AI data access gap during a known vulnerability period is exactly the kind of finding that surfaces in audits. Request tenant-level confirmation from Microsoft about whether your environment was included in the set of tenants contacted for remediation validation.

Validate DLP and Label Enforcement — Do Not Assume It

The most important operational lesson from this incident is that governance controls must be tested, not assumed. Create test emails in Sent Items and Drafts, apply confidentiality labels, configure the relevant DLP policies, and then directly query Copilot to verify that the content is excluded from summaries. Perform these tests across environments — Outlook desktop, Outlook on the web, and mobile — and document the results as evidence of compliance.

Microsoft’s own Purview documentation acknowledges that label behavior can vary by application. Administrators should review those exceptions explicitly and understand where enforcement gaps may still exist in their environment, even after the CW1226324 fix has been applied.

Enable Restricted Content Discovery for High-Risk SharePoint Sites

Restricted Content Discovery (RCD) is a SharePoint Advanced Management feature available to tenants with Microsoft 365 Copilot licenses. When enabled for a site, RCD removes that site’s content from Copilot’s retrieval pipeline entirely. This is a harder enforcement mechanism than sensitivity labels because it operates at a different layer: the content never enters the model’s context window in the first place, regardless of whether label enforcement functions correctly. Organizations should identify their highest-sensitivity SharePoint sites — legal, HR, finance, executive communications — and enable RCD for those locations as a defense-in-depth measure.

Credential and Secret Rotation

Organizations where sensitive email frequently contains credentials, API keys, tokens, or other secrets — a practice that should be eliminated but remains common — should conduct a targeted review of Sent Items and Drafts content from the exposure window. If messages containing sensitive credentials were present in those folders during the period when Copilot was incorrectly processing them, those credentials should be rotated as a precautionary measure.

Review Downstream Content

AI-generated summaries can be copied, forwarded, and embedded in other documents. If Copilot produced summaries from confidential emails during the exposure window, those summaries may have traveled downstream into shared collaboration spaces. Organizations should review whether any Copilot-generated content created between late January and mid-February 2026 was shared into Teams channels, saved to SharePoint, or sent externally, and assess whether that content may have contained information derived from labeled emails.

What This Means for the Future of Enterprise AI Governance

The CW1226324 incident is not a story about a single software bug. It is a signal about the structural challenges of deploying AI in enterprise environments where data governance is a legal and operational requirement.

Traditional DLP systems were designed for a world of human actors and deterministic software. A human forwarding an email can be flagged. A file upload to an unauthorized destination can be blocked. These are bounded, predictable actions that existing rule sets can anticipate. AI assistants operate differently. They synthesize. They summarize. They repackage information in new forms that existing rules may not have been designed to govern. The output of a Copilot summarization session is not a forwarded email. It is a new artifact, and whether DLP policies govern that new artifact — and all the places it might travel — depends on how comprehensively the governance framework has been extended to cover AI-generated content.

Dr. Ilia Kolochenko, CEO of ImmuniWeb and a member of the Europol Data Protection Experts Network (EDEN), told Cybernews that incidents like this one will likely surge in 2026, possibly becoming the most frequent type of security incident at both large and small companies around the globe. That prediction reflects a straightforward dynamic: the number of AI tools with access to enterprise data is increasing rapidly, the governance frameworks for controlling those tools are still maturing, and the intersection of retrieval-augmented AI with complex, layered data classification systems creates enforcement challenges that are not fully solved by any vendor today.

It is also worth noting that this structural challenge is not unique to Microsoft. Any AI assistant with retrieval access to enterprise data operates through the same architectural pattern: a retrieval layer selects content, an enforcement layer gates what reaches the model, and a generation layer produces output. If the enforcement layer fails — for any reason, whether a code bug, a configuration gap, or a novel adversarial input — restricted data enters the model context, and the output reflects it. Organizations deploying Google Gemini for Workspace, Salesforce Einstein, or any third-party AI tool with document and email access face the same fundamental governance challenge.

The Microsoft Copilot incident argues for a fundamental shift in how organizations approach AI governance. Rather than treating AI tools as subject to existing controls and assuming those controls will transfer cleanly, security teams need to verify AI-specific enforcement, build audit capabilities that cover AI-generated content and not just raw data access, and recognize that AI introduces new exposure vectors — summarization, synthesis, and reformatting — that require policy coverage in their own right.

Implications for Cybersecurity Training and Awareness

For cybersecurity educators and awareness professionals, this incident provides a powerful and concrete teaching case for several important concepts.

First, it illustrates why ‘trust but verify’ is inadequate as a governance posture for enterprise AI. Many organizations rely on vendor assurances and assume that certified, enterprise-grade tools honor their security policies. The CW1226324 incident demonstrates that even well-resourced vendors with dedicated security engineering teams can ship code that silently violates declared data governance behavior. Verification must be active, documented, and repeated.

Second, it surfaces the distinction between access control and processing control — a distinction that end users, and even many IT professionals, frequently conflate. Access control asks: who can open this file? Processing control asks: what systems are permitted to ingest and transform this content? DLP policies for AI tools operate in the processing control domain, and that domain requires its own governance logic, separate from and in addition to traditional access control.

Third, the incident is a practical demonstration of why data classification hygiene matters. Organizations that had not diligently applied sensitivity labels to their most sensitive email content had no mechanism — even a potentially broken one — through which Copilot’s behavior could have been restricted. Data governance frameworks only protect what they govern. That means consistent, accurate classification of sensitive information is a prerequisite for any AI safety control to function.

Finally, for executives and business leaders, this incident illustrates a risk category that frequently goes unacknowledged in AI adoption conversations: the risk that AI tools deployed for productivity may inadvertently circumvent the compliance and governance controls on which the organization’s regulatory standing depends. As AI adoption accelerates across every sector, governance must be a first-order consideration, not an afterthought to be addressed after rollout.

Conclusion

The Microsoft 365 Copilot confidential email incident is, on its surface, a story about a code bug that was identified, logged as CW1226324, and fixed. Microsoft has patched it. The fix is rolling out. For some organizations, the matter ends there.

But for security professionals who understand the trajectory of enterprise AI deployment, the incident is something more: a concrete demonstration of the governance gaps that emerge when AI retrieval systems intersect with layered data classification frameworks that were not designed with AI in mind. It shows that existing security tools — EDR, WAF, traditional DLP alerts — did not catch the problem. It shows that the failure mode was silent, operating for weeks before broad external reporting brought it to light. And it shows that the AI’s technically accurate output — summaries drawn from content the user was already authorized to see — can still represent a meaningful control failure from a compliance and governance standpoint.

The organizations best positioned to navigate this moment are those that have stopped assuming their AI tools inherit their security posture and have started treating AI governance as its own discipline — one that requires dedicated policy design, active verification, comprehensive logging, and a clear understanding of where the vendor’s responsibility ends and the organization’s begins.

The AI did not break in. The door was already unlocked. The question now is who else has a key, and whether anyone will notice when it is used.

Sources: BleepingComputer, The Register, TechRadar, Office 365 IT Pros, VentureBeat, Windows Forum, Cybernews, Cybersecurity News, SOC Prime, Cybersecurity Insiders / Saviynt, TechCrunch, Microsoft FY26 Q2 Earnings, Microsoft Learn documentation

← all articles