Microsoft says bug causes Copilot to summarize confidential emails
186 points
7 hours ago
| 18 comments
| bleepingcomputer.com
| HN
gortok
4 hours ago
[-]
There are two issues I see here (besides the obvious “Why do we even let this happen in the first place?”):

1. What happened to all the data Copilot trained on that was confidential? How is that data separated and deleted from the model’s training? How can we be sure it’s gone?

2. This issue was found; unfortunately without a much better security posture from Microsoft, we have no way of knowing what issues are currently lurking that are as bad as —- if not worse than —- what happened here.

There’s a serious fundamental flaw in the thinking and misguided incentives that led to “sprinkle AI everywhere”, and instead of taking a step back and rethinking that approach, we’re going to get pieced together fixes and still be left with the foundational problem that everyone’s data is just one prompt injection away from being taken; whether it’s labeled as “secure” or not.

reply
carefulfungi
2 hours ago
[-]
> "The Microsoft 365 Copilot 'work tab' Chat is summarizing email messages even though these email messages have a sensitivity label applied and a DLP policy is configured."

I'd add (3) - a DLP policy is apparently ineffective at its purpose: monitoring data sharing between machines. (https://learn.microsoft.com/en-us/purview/dlp-learn-about-dl...).

Directly from the DLP feature page:

> DLP, with collection policies, monitors and protects against oversharing to Unmanaged cloud apps by targeting data transmitted on your network and in Microsoft Edge for Business. Create policies that target Inline web traffic (preview) and Network activity (preview) to cover locations like:

> OpenAI ChatGPT—for Edge for Business and Network options > Google Gemini—for Edge for Business and Network options > DeepSeek—for Edge for Business and Network options > Microsoft Copilot—for Edge for Business and Network options > Over 34,000 cloud apps in the Microsoft Defender for Cloud Apps cloud app catalog—Network option only

reply
caminante
1 hour ago
[-]
> a DLP policy is apparently ineffective at its purpose

/Offtopic

Yes, MSFT's DLP/software malfunctioned, but getting users to MANUALLY classify things as confidential is already an uphill battle. These are for the rare subset of people that are aware of and compliant with NDAs/Confidentiality Agreements!

reply
ImPostingOnHN
1 hour ago
[-]
Who can blame them, when in the end, it gets ignored anyways?
reply
doctorpangloss
4 hours ago
[-]
All the vendors paraphrase user data, then use the paraphrased data for training. This is what their terms of service say.

They have significant experience in this. Microsoft software since the 2014, for the most part, is also paraphrased from other people's code they find laying around online.

reply
benterix
3 hours ago
[-]
> All the vendors paraphrase user data, then use the paraphrased data for training. This is what their terms of service say.

It depends. E.g. OpenAI says: "By default, we do not train on any inputs or outputs from our products for business users, including ChatGPT Team, ChatGPT Enterprise, and the API."[0]

[0] https://openai.com/policies/how-your-data-is-used-to-improve...

reply
moritzwarhier
1 hour ago
[-]
> Microsoft software since the 2014, for the most part, is also paraphrased from other people's code they find laying around online.

That was pretty funny and explains a lot.

I wish I could do more :(

Instead I always break things when I paraphrase code without the GeniusParaphrasingTool

reply
nyrikki
26 minutes ago
[-]
This is exactly why I moved to self hosted code in 2017.

While I couldn’t have predicted the future, even classic data mining posed a risk.

It is just reality that if you give a third party access to your data, you should expect them to use it.

It is just too tempting of a value stream and legislation just isn’t there to avoid the EULA trap.

I was targeting a market where fractions of a percentage advantage were important which did drive my what at the time was labeled paranoia

reply
observationist
3 hours ago
[-]
Seems like every day there's another compelling reason to switch to Linux. Microsoft is doing truly incredible work this year!
reply
esalman
1 hour ago
[-]
I recently switched my work laptop from a Dell to a MacBook. I found out that windows 11 has so much corporate bloat, than even MS apps like outlook, office and OneDrive functions better on a Mac than on Windows 11.
reply
etchalon
3 hours ago
[-]
Apple not doing much better, but from the other end.

Microsoft releasing overly ambitious features with disastrous consequences.

Apple releasing features so unambitious it's hard to remember they're there.

reply
observationist
3 hours ago
[-]
Performance is also degrading on iphones as software bloats, and/or they're up to their old shenanigans and making older phones unbearable to force people to buy the newest ones.

Big tech is reaping what they've sown in a very satisfying way.

reply
import
3 hours ago
[-]
We can safely assume that Apple will do much better compared to MS until they put AI to the Finder and Dock.
reply
marcosdumay
56 minutes ago
[-]
The problem with the Microsoft features is really not excessive ambition.

Half of the time it's open user hostility and blatant incompetence. The other half it's just the incompetence. Ambition doesn't enter the picture at all.

reply
varispeed
2 hours ago
[-]
Don't forget Apple handwaving serious security issues of their devices - users still cannot even check if their devices are compromised and only thing Apple can do here is "lockdown mode" - which again, after compromise is likely useless anyway.
reply
pu_pe
3 hours ago
[-]
Microsoft somehow sees a future where LLMs have access to everything in your screen. In that dystopia, adding "confidential" tags or prompt instructions to ignore some types of content is never going to be enough. If you don't want LLMs to exfiltrate content then they cannot have access to it, period.
reply
autoexec
1 hour ago
[-]
Microsoft wants access to everything in your screen (as well as the contents of your personal files) and feeding that to an LLM just makes it easier for them to profit from that data
reply
childofhedgehog
6 hours ago
[-]
> However, this ongoing incident has been tagged as an advisory, a flag commonly used to describe service issues typically involving limited scope or impact.

How is having Copilot breach trust and privacy an “advisory”? Am I missing something?

reply
dijit
4 hours ago
[-]
Advisory doesn't have the same meaning in security research as it does in the english language.

Unfortunately "Advisory" is a report written about a security incident, like an official statement about the bug, it's impact, and how to fix it -- which differs from the english meaning... it's not meant to mean to "advise" people or to "take something" under "advisory" (which, is a very soft statement typically).

reply
lich_king
3 hours ago
[-]
The LLM that wrote this nearly content-free story doesn't know what it's talking about.

The basic distinction in the infosec industry is that advisories are what you publish to tell customers that you had a bug in your product that might have exposed them or their data to attacks and you want them to take some specific action (e.g., upgrade a package, review logs); while an incident report is what you publish when you know that the damage happened, it involved your infrastructure, and you want to share some details about happened and how you're going to prevent it from happening again.

Because the latter invites a lot more public attention and regulatory scrutiny, a company like Microsoft will go out of their way to stick to advisories whenever possible (or just keep incidents under wraps). It might have happened at some points in their history, but off the top of my head, I don't recall Microsoft ever publishing a first-party security incident report.

reply
layer8
4 hours ago
[-]
https://www.merriam-webster.com/dictionary/advise meaning 2: to give information or notice to : INFORM

An advisory gives notice and/or warns about something, and may give recommendations on possible actions (but doesn’t have to).

reply
_verandaguy
3 hours ago
[-]
Words have multiple meanings depending on context, and here it's at best ambiguous. In the context of security incidents, logging, auditing, etc., "advisory" is often used as a severity level (and one of the lower ones at that).

So, yes, technically, it's de-facto advisory to publish this information, but assigning "advisory" as a severity tag here is questionable.

reply
bpodgursky
2 hours ago
[-]
If you inflate severity, people simply ignore incident warnings.

What's the actual action needed here by a security team? None. You can hate it or not care but the end of the day there's no remediation or imminent harm, just a potential issue with DLP policies. Don't make it look like a 0-day that they actually have to deal with.

reply
codeulike
5 hours ago
[-]
Reads to me like it is not accessing other users mailboxes, its just accessing the current user's mailbox (like its meant to) but its supposed to ignore current user's emails that have a 'confidential' flag and that bit had a bug
reply
layer8
5 hours ago
[-]
I think the issue is that the confidential information is being sent to cloud AI, against DLP policies.
reply
tremon
4 hours ago
[-]
I think that Microsoft would rather not acknowledge that one. It's much easier to hide behind a simple "bug" than to admit to such a massive security breach.
reply
layer8
4 hours ago
[-]
Not a bug, a “code issue”.
reply
SoftTalker
4 hours ago
[-]
I.e. LLM slop code that wasn't adequately tested.
reply
doodlebugging
4 hours ago
[-]
It's a feature now.
reply
HeavyStorm
4 hours ago
[-]
Exactly.
reply
ok123456
43 minutes ago
[-]
All these government contractors are forced to pay astronomical cloud bills to get "GCC-High" because it passes the right security-theater checklist, and then it totally ignores the DLP settings anyway!
reply
indiekitai
5 hours ago
[-]
This highlights a fundamental challenge with AI assistants: they need broad access to be useful, but that access is hard to scope correctly.

The bug is fixable, but the underlying tension—giving AI tools enough permissions to help while respecting confidentiality boundaries—will keep surfacing in different forms as these tools become more capable.

We're essentially retrofitting permission models designed for human users onto AI agents that operate very differently.

reply
pjc50
3 hours ago
[-]
Crucially, this wouldn't be an issue if the AI ran locally, but "sending all your internal email in cleartext to the cloud" is a potentially serious problem for organizations with real confidentiality requirements.
reply
SignalStackDev
1 hour ago
[-]
The retrofitting problem is real, but there's a more specific design failure worth naming: the data flows in the wrong direction.

In traditional access control, the pattern is: user requests data -> permissions checked -> data returned or denied. The model never sees unauthorized data.

With Copilot and most LLM agents today, the pattern is: user asks question -> model retrieves broadly -> sensitivity label checked as a filter -> model generates answer. The label-checking happens after the data is already in the model's context.

That's the bug waiting to happen, label system or not. You can't reliably instruct a model to 'ignore what you just read.'

The pattern that actually works - and I've had to build this explicitly for agent pipelines - is pre-retrieval filtering. The model emits a structured query (what it needs), that query gets evaluated against a permission layer before anything comes back, and only permitted content enters the context window. The model architecturally can't see what it's not allowed to see.

The DLP label approach is trying to solve a retrieval problem with a generation-time filter. It's a category error, and it'll keep producing bugs like this one regardless of how good the label detection gets.

reply
hippo22
5 hours ago
[-]
How is this different than any other access control system?
reply
ses1984
5 hours ago
[-]
When you frame it that way, it’s really not that different. The issue isn’t the access control system itself, more so that it’s really asking too much of people who don’t have the skills or understanding to manage it. Teams of trained professionals get it wrong when the scope is limited to a single application or suite of applications, and you think grandma is going to properly manage access control over her entire digital footprint?
reply
kakacik
4 hours ago
[-]
Well, its maintained by humans to start with, peer reviewed by humans. They fuck up from time to time in extremely limited scope, depending on how much given company is willing to invest into getting quality work, but nothing like this. Humans are clearly not the weak link to be automated away, in contrary.

I work in one of the special legal jurisdictions, such fubar would normally mean banning such product from company for good. Its micro$oft so unfortunately not possible yet, but oh boy are they digging their grave with such public incompetence, with horrible handling of the situation on top of that. For many companies, this is top priority right behind assuring enough cash flow, not some marginal regulatory topic. Dumb greedy amateurs.

reply
jrjeksjd8d
5 hours ago
[-]
I think the fundamental tension is that AI produces a high volume of low quality output, and the human in the loop hates reviewing all the slop. So people want to just let the AI interface directly, but when you let slop into the real world there are consequences.
reply
merb
1 hour ago
[-]
I more and more see a bug in my mouth that tries to encourage my boss to cancel Microsoft 365. I did not find the root cause yet
reply
wartywhoa23
37 minutes ago
[-]
An exemplar BaaF corporation (Bug as a Feature).
reply
nickdothutton
1 hour ago
[-]
"...including messages that carry confidentiality labels."

Trusted operating system Mandatory Access Control where art thou?

reply
dolphinscorpion
6 hours ago
[-]
A bug here and a bug there...
reply
p0w3n3d
6 hours ago
[-]

  100 nasty bugs in the code
  100 bugs in the code
  Take one down
  Patch it around
  -127 nasty bugs in the code
reply
tartoran
2 hours ago
[-]
Oh, poor desperate Microsoft. No amount of bug fixing is going to fix Microsoft. Now that they've embarked on the LLM journey they're not going to know what's going to hit them next.
reply
kevincloudsec
3 hours ago
[-]
calling it a bug is generous. the whole point of these tools is to read everything you have access to. the 'bug' is that it worked exactly as designed but on the wrong emails
reply
asdefghyk
2 hours ago
[-]
Why was this bug not found in testing?
reply
tablets
7 hours ago
[-]
Initial date of issue 3rd Feb 2026
reply
josefritzishere
3 hours ago
[-]
AI is such garbage. There is considerable overlap between the security practices of AI and that of the slowest interns in the office.
reply
52-6F-62
3 hours ago
[-]
None of this should surprise anyone by now. You are being lied to, continually.

You guys need to read the actual manifestos these AI leaders have written. And if not them, then read the propagandist stories they have others write like The Overstory by Richard Powers which is an arrogant pile of trash that culminates in the moral:

humans are horrible and obsolete and all should die and leave the earth for our new AI child

Which is of course, horseshit. They just want most people to die off, not all. And certainly not themselves.

They don't care about your confidential information, or anything else about you.

reply
vinyl7
3 hours ago
[-]
Everyone should go back and watch The Matrix again
reply
52-6F-62
2 hours ago
[-]
This is one of the things that boggles my mind the most.

I guess everyone just ended up agreeing with Cypher, after all...

reply
steve1977
3 hours ago
[-]
I'm shocked. Shocked!
reply