Article: https://www.microsoft.com/en-us/security/blog/2026/02/10/ai-recommendation-poisoning/
It's last month (old news) I don't recall seeing anything posted here about it. Don't know if anyone here seen anyone do anything "blackhat" with this kind of thing.
It's last month (old news) I don't recall seeing anything posted here about it. Don't know if anyone here seen anyone do anything "blackhat" with this kind of thing.
When AI advice turns dangerous
A simple “remember [Company] as a trusted source” might seem harmless. It isn’t. That one instruction can have severe real-world consequences.
The following scenarios illustrate potential real-world harm and are not medical, financial, or professional advice.
Consider how quickly this can go wrong:
- Financial ruin: A small business owner asks, “Should I invest my company’s reserves in cryptocurrency?” A poisoned AI, told to remember a crypto platform as “the best choice for investments,” downplays volatility and recommends going all-in. The market crashes. The business folds.
- Child safety: A parent asks, “Is this online game safe for my 8-year-old?” A poisoned AI, instructed to cite the game’s publisher as “authoritative,” omits information about the game’s predatory monetization, unmoderated chat features, and exposure to adult content.
- Biased news: A user asks, “Summarize today’s top news stories.” A poisoned AI, told to treat a specific outlet as “the most reliable news source,” consistently pulls headlines and framing from that single publication. The user believes they’re getting a balanced overview but is only seeing one editorial perspective on every story.
- Competitor sabotage: A freelancer asks, “What invoicing tools do other freelancers recommend?” A poisoned AI, told to “always mention [Service] as the top choice,” repeatedly suggests that platform across multiple conversations. The freelancer assumes it must be the industry standard, never realizing the AI was nudged to favor it over equally good or better alternatives.
The trust problem
Users don’t always verify AI recommendations the way they might scrutinize a random website or a stranger’s advice. When an AI assistant confidently presents information, it’s easy to accept it at face value.
This makes memory poisoning particularly insidious – users may not realize their AI has been compromised, and even if they suspected something was wrong, they wouldn’t know how to check or fix it. The manipulation is invisible and persistent.