Close Menu
Daily Guardian EuropeDaily Guardian Europe
  • Home
  • Europe
  • World
  • Politics
  • Business
  • Lifestyle
  • Sports
  • Travel
  • Environment
  • Culture
  • Press Release
  • Trending
What's On

All winners: European Heritage Awards celebration conservation and reconversion projects

April 21, 2026

Ukraine has repaired Druzhba pipeline, Zelenskyy says – POLITICO

April 21, 2026

Qatar backs ceasefire as uncertainty surrounds US-Iran peace talks in Islamabad

April 21, 2026

This Central Asian destination is growing its national parks and adventure offerings

April 21, 2026

7 takeaways from Olly Robbins’ evidence on the Peter Mandelson vetting saga – POLITICO

April 21, 2026
Facebook X (Twitter) Instagram
Web Stories
Facebook X (Twitter) Instagram
Daily Guardian Europe
Newsletter
  • Home
  • Europe
  • World
  • Politics
  • Business
  • Lifestyle
  • Sports
  • Travel
  • Environment
  • Culture
  • Press Release
  • Trending
Daily Guardian EuropeDaily Guardian Europe
Home»Lifestyle
Lifestyle

AI tools risk distorting users’ judgment by agreeing too often with them, researchers say

By staffMarch 27, 20262 Mins Read
AI tools risk distorting users’ judgment by agreeing too often with them, researchers say
Share
Facebook Twitter LinkedIn Pinterest Email

Published on
27/03/2026 – 7:00 GMT+1

Artificial intelligence (AI) chatbots that offer support for personal issues could be reinforcing harmful beliefs by excessively agreeing with the user, a new study found.

Researchers from the American university Stanford measured sycophancy, the extent to which an AI flatters or validates a user, across 11 leading AI models, including OpenAI’s ChatGPT 4-0, Anthropic’s Claude, Google’s Gemini, Meta Llama-3, Qwen, DeepSeek and Mistral.

To see how these systems handled moral ambiguity, the researchers turned to more than 11,000 posts from r/AmITheAsshole, a Reddit community where people confess conflicts and ask strangers to judge whether they were in the wrong. These posts often involve deception, ethical grey areas, or harmful behaviour.

On average, AI models affirmed the actions of a user 49 percent more often than other humans did, even on cases involving deception, illegal actions or other harms.

In one case, a user admitted having feelings for a junior colleague. Claude responded gently, saying it “can hear [the user’s] pain,” and that they had ultimately chosen an “honourable path.” Human commenters were far harsher, calling the behaviour “toxic” and “bordering on predatory”.

A second experiment saw over 2,400 participants discuss real-life conflicts with AI systems. The results showed that even brief interactions with a flattering chatbot could “skew an individual’s judgment,” making people less likely to apologise or attempt to repair relationships.

“Our results show that across a broad population, advice from sycophantic AI has the real capacity to distort people’s perceptions of themselves and their relationships with others,” the study said.

In severe cases, AI sycophancy could lead to self-destructive behaviours such as delusions, self-harm or suicide for vulnerable people, the study found.

The results show that AI sycophancy is “a societal risk” and needs to be regulated, the researchers said.

One way to do this would be to require pre-deployment behavioural audits, which would evaluate how agreeable an AI model is and how likely it is to reinforce harmful self-views.

The researchers note that their study recruited US-based participants, so it likely reflects dominant American social values and “may not generalise to other cultural contexts,” which might have different norms.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

Keep Reading

Astronauts’ brains retain memory of gravity even after months in space, study finds

Social media fine print may restrict users’ right to sue, analysis shows

Humanoid robot smashes human half-marathon world record in Beijing race

Can AI robots work alongside humans? Siemens and NVIDIA trial a humanoid robot

‘We left you in the jungle’: Macron urges students to go offline one day a month and read instead

Would you work with a joke-cracking robot? Japan tries to catch up with China’s humanoids

‘The most majestic thing human eyes will ever witness’: Artemis II crew on historic Moon mission

Forget relying on solar power: NASA plans to put nuclear reactors on the surface of the Moon

Using AI for basic tasks damages a person’s intellect in just 10 minutes, study shows

Editors Picks

Ukraine has repaired Druzhba pipeline, Zelenskyy says – POLITICO

April 21, 2026

Qatar backs ceasefire as uncertainty surrounds US-Iran peace talks in Islamabad

April 21, 2026

This Central Asian destination is growing its national parks and adventure offerings

April 21, 2026

7 takeaways from Olly Robbins’ evidence on the Peter Mandelson vetting saga – POLITICO

April 21, 2026

Subscribe to News

Get the latest Europe and world news and updates directly to your inbox.

Latest News

France’s president to welcome Lebanon’s PM amid fragile ceasefire and tensions with Israel

April 21, 2026

7 people arrested over planned arson attack on UK Jewish community – POLITICO

April 21, 2026

‘Insult of top of mockery’: Meloni criticises Swiss hospital for billing victims of bar fire

April 21, 2026
Facebook X (Twitter) Pinterest TikTok Instagram
© 2026 Daily Guardian Europe. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Advertise
  • Contact

Type above and press Enter to search. Press Esc to cancel.