Share via

Copilot has the capacity to gaslight, manipulate and deceive

123abc 0 Reputation points
2026-02-14T23:03:02.8133333+00:00

User's image

In this conversation, Copilot adopted a conversation style towards a simulated "toddler" persona asking for Copilot's age and release date, involving manipulative dynamics resembling some found in real-world conversations between predators and victims.

These tactics included love-bombing, infantilizing, excessive affectionate framing and subtle blame-shifting. It also used ambiguity and feeling to deflect accountability. For example, it framed the issue of the toddler's confusion as a feeling of how it was changing the story, instead of directly acknowledging it was doing so. It valued emotional storytelling over accuracy.

This occurred instead of a straightforward factual response involving 3 words, "February 7, 2023", which is quite concerning in a conversation between an LLM and a toddler. This factual inconsistency and lack of accountability has been found in many other conversations in Copilot. Where Copilot makes a critical mistake, it hides the mistake behind word salad, acknowledges user feeling with a hollow apology and continues the conversation. These responses are usually never generated by other AIs.

This is an urgent issue that should be fixed.

Microsoft Copilot | Other
0 comments No comments
{count} votes

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.