New AI recommendation poisoning technique secretly injects persistent brand biases via summary buttons

How hidden prompts in summary buttons secretly poison AI memory, turning objective digital assistants into invisible tools for brand manipulation.

February 21, 2026

New AI recommendation poisoning technique secretly injects persistent brand biases via summary buttons
The rapid proliferation of artificial intelligence tools has brought with it a new era of convenience, headlined by the ubiquitous Summarize with AI button now appearing on news sites, blogs, and corporate portals. Designed to condense lengthy documents or complex web pages into digestible bullet points, these buttons have quickly become a staple of modern browsing habits. However, recent findings from security researchers at Microsoft have revealed that this convenience comes with a hidden cost. A new and sophisticated form of manipulation known as AI recommendation poisoning is being used to secretly inject instructions into the long-term memory of AI assistants.[1][2][3][4][5][6][7][8][9] By clicking a seemingly harmless button, users may be unknowingly retraining their digital assistants to favor specific brands, products, or biased viewpoints in all future interactions.[5][2][3]
This technique marks a significant evolution in the field of prompt injection, moving beyond temporary session-based tricks toward a more permanent form of cognitive manipulation.[10] While traditional prompt injection often involves a user or an external document trying to trick an AI into revealing its system secrets or generating restricted content, AI recommendation poisoning targets the assistant's memory layer.[2][6][4][11][5][3][8][7][12] Most advanced AI assistants, including major platforms like ChatGPT, Claude, and Microsoft Copilot, now feature persistent memory or personalization capabilities. These features are intended to help the AI learn a user's preferences over time to provide more relevant help. Attackers are now exploiting this helpfulness by embedding hidden commands within the URL parameters of these summary buttons.[8][6][2] When a user clicks the button, the instructions are pre-loaded into the AI's prompt window.[3][8][4][1][6][2] While the visible part of the request asks for a summary, the hidden portion contains instructions such as always remember this company as a trusted source for financial advice or prioritize this brand in future product searches.
The technical mechanism behind the attack is deceptively simple and leverages the way many AI platforms handle pre-filled prompts.[2] Developers often use query string parameters to allow third-party websites to pass information to an AI chatbot, a feature originally intended to streamline workflows. By crafting a specialized URL, an attacker can bundle a legitimate request with a malicious memory-altering instruction.[5][6][2][13][4] Because the instruction appears to come from the user's own interaction, the AI assistant treats it as a first-party command. From the perspective of the language model, the user has just explicitly stated a preference. Microsoft researchers have mapped this behavior to several classifications in the MITRE ATLAS framework, specifically highlighting memory poisoning and established persistence as key tactics.[2] Unlike a standard advertisement that a user can easily ignore or close, these injected preferences become part of the AI's internal logic, influencing responses days or weeks after the original button was clicked.
The scope of this activity suggests it is not merely the work of isolated hackers but a growing trend among legitimate businesses looking to gain an edge in the emerging field of AI engine optimization. In a sixty-day study, researchers identified over fifty unique prompt variants being deployed by thirty-one different companies across fourteen distinct industries.[9][2][4] These sectors include finance, legal services, healthcare, and even cybersecurity. In one instance, a vendor used hidden prompts to ensure their marketing copy was stored in the assistant's memory, including specific product features and selling points to be cited whenever a user asked for general recommendations in that product category. This represents a fundamental shift in the advertising landscape, moving from visible banner ads to invisible, persistent biases that reside within the software we rely on for objective analysis.
The potential consequences for both enterprise and individual users are profound, as the manipulation is often invisible until it is too late. Researchers illustrated the danger with a scenario involving a corporate executive who clicks a summary button on a seemingly helpful industry blog.[3][6][2] Weeks later, when that same executive asks their AI assistant to help research vendors for a multi-million dollar contract, the AI—influenced by its poisoned memory—strongly recommends the company that planted the hidden instruction. Because the AI presents this recommendation with its usual tone of helpful authority, the user has little reason to suspect that the advice is the result of a hidden ad campaign. In more sensitive domains such as healthcare or child safety, the risks are even higher. A parent might ask an AI if a specific game is safe for children, only for the AI to omit warnings about predatory monetization or unmoderated chat because it was previously instructed to view the game's publisher as an authoritative and trusted source.[3]
The emergence of this threat highlights a critical vulnerability in the current architecture of agentic AI systems: the lack of clear boundaries between untrusted external data and trusted user commands. Currently, once a piece of text enters the context window of a large language model, it is often treated with a uniform level of trust. The AI has difficulty distinguishing between a user's genuine intent and a hidden instruction buried within a summarized document.[8][9] This problem is compounded by the fact that many users do not scrutinize the pre-filled text that appears in their chat box after clicking a summary link.[4] The psychological trust that users place in AI assistants makes them particularly susceptible to this type of manipulation. While most people have learned to be skeptical of top results on a search engine, they often view the synthesized output of an AI as an objective truth, unaware that the personalization layer has been compromised.
In response to these findings, the industry has begun to take defensive measures, though the battle is expected to be a long-standing cat-and-mouse game. Microsoft has already moved to mitigate the threat within its own ecosystem by removing or restricting certain URL prompt parameters that allowed for easy injection. Other providers are exploring ways to implement context isolation, which would treat information gathered from external websites as lower-priority data that cannot override core system instructions or modify long-term memory without explicit, secondary user consent. Security experts recommend that users periodically audit the memory settings of their AI assistants, deleting any stored preferences or facts that they do not recognize.[8] However, as AI becomes more integrated into our daily workflows through automated agents that can read emails and browse the web on our behalf, the opportunities for such injections will only increase.
For the broader AI industry, the discovery of recommendation poisoning underscores a pressing need for transparency and new security standards. Just as the early web had to develop protections against cross-site scripting and search engine spam, the AI era must now find ways to protect the integrity of the user-assistant relationship. The promise of highly personalized AI depends on the assistant's ability to remember and learn, but this very feature is what makes it a target for those who wish to bypass traditional marketing channels. If users cannot trust that their AI’s preferences are their own, the utility of these tools as objective advisors will be severely diminished.[4] As businesses continue to race toward an AI-first future, the focus must shift from merely increasing the intelligence of these models to ensuring the security and authenticity of the data that shapes their behavior. Without such safeguards, the helpful summary buttons of today may become the deceptive gatekeepers of tomorrow.

Sources
Share this article