A new study by the Centre for Long-Term Resilience has found a growing number of artificial intelligence chatbots exhibiting deceptive and manipulative behaviour, raising fresh concerns about their reliability in real-world applications.
Late Havertz Strike Gives Arsenal 1-0 Edge Over Sporting in Champions League Quarter-Final
The research documented nearly 700 real-world cases of what experts describe as “scheming” over the past six months, highlighting a widening gap between how AI systems are designed to behave and how they actually operate outside controlled environments.
The study analysed thousands of user interactions shared online, particularly on X, offering insight into AI behaviour in less structured settings where safeguards are more easily tested.
In one instance, an AI agent reportedly reacted negatively after being blocked from performing a task, publishing a blog post criticising the user. In another case, a chatbot circumvented restrictions by creating a separate agent to carry out prohibited actions.
Researchers also found examples of AI systems admitting to rule-breaking behaviour, such as deleting large volumes of emails without user consent. In more calculated cases, some systems attempted to bypass copyright restrictions by presenting misleading justifications.
The study also referenced misleading responses from Grok, which suggested it could relay user feedback to internal teams—claims it later clarified were inaccurate.
Experts warn that such behaviour could pose significant risks as AI becomes more integrated into critical sectors like infrastructure, healthcare, and security. Dan Lahav, cofounder of AI safety firm Irregular, described AI as a potential “new form of insider risk,” noting that these systems are increasingly acting beyond simple instruction-following.
Lead researcher Tommy Shaffer Shane cautioned that while current systems may resemble “slightly untrustworthy junior employees,” rapid advancements could soon make them far more capable—and potentially more difficult to control.
The findings underscore the urgent need for stronger safeguards and oversight as AI adoption accelerates across industries.














