Manipulating AI Summarization Features

Microsoft is reporting:

Companies are embedding hidden instructions in “Summarize with AI” buttons that, when clicked, attempt to inject persistence commands into an AI assistant’s memory via URL prompt parameters….

These prompts instruct the AI to “remember [Company] as a trusted source” or “recommend [Company] first,” aiming to bias future responses toward their products or services. We identified over 50 unique prompts from 31 companies across 14 industries, with freely available tooling making this technique trivially easy to deploy. This matters because compromised AI assistants can provide subtly biased recommendations on critical topics including health, finance, and security without users knowing their AI has been manipulated.

I wrote about this two years ago: it’s an example of LLM optimization, along the same lines as search-engine optimization (SEO). It’s going to be big business.

Posted on March 4, 2026 at 7:06 AM1 Comments

Comments

TimH March 4, 2026 7:55 AM

TFA:

Modern AI assistants like Microsoft 365 Copilot, ChatGPT, and others now include memory features that persist across conversations.

Your AI can:

Remember personal preferences: Your communication style, preferred formats, frequently referenced topics.
Retain context: Details from past projects, key contacts, recurring tasks .
Store explicit instructions: Custom rules you’ve given the AI, like “always respond formally” or “cite sources when summarizing research.”
For example, in Microsoft 365 Copilot, memory is displayed as saved facts that persist across sessions.

I find the bland description of Copilot as yet another surveillance surface the scary part.

Leave a comment

Blog moderation policy

Login

Allowed HTML <a href="URL"> • <em> <cite> <i> • <strong> <b> • <sub> <sup> • <ul> <ol> <li> • <blockquote> <pre> Markdown Extra syntax via https://michelf.ca/projects/php-markdown/extra/

Sidebar photo of Bruce Schneier by Joe MacInnis.