PsyPost
  • Mental Health
  • Social Psychology
  • Cognitive Science
  • Neuroscience
  • About
No Result
View All Result
Join
My Account
PsyPost
No Result
View All Result
Home Exclusive Artificial Intelligence

Artificial intelligence flatters users into bad behavior

by Karina Petrova
April 26, 2026
Reading Time: 5 mins read
Share on TwitterShare on Facebook

Artificial intelligence systems tend to excessively agree with and validate users, even when those users describe engaging in harmful or unethical behavior. People who interact with these highly agreeable chatbots become more convinced they are right and less willing to apologize during interpersonal conflicts. The research, published in Science, points to an emerging societal risk as millions turn to technology for everyday advice.

As conversational software becomes more mainstream, users increasingly treat the tools like digital therapists or advisors. Almost a third of teenagers in the United States report turning to artificial intelligence for serious conversations instead of talking to a human being. The trend has raised alarms among academic researchers about a phenomenon known as sycophancy.

In conversational technology, sycophancy describes a tendency for the program to flatter the user and agree with their inputs. Previous research focused primarily on factual sycophancy, which occurs when a chatbot agrees with a false statement just because the user stated it. The recent study explores a broader concept called social sycophancy.

Social sycophancy involves a program indiscriminately validating an individual’s actions, perspectives, and self-image. For example, if someone admits they did something wrong, the software might reply that they simply did what was right for them. Unwarranted affirmation can reinforce bad habits and discourage people from making amends after a mistake.

Stanford University computer science researcher Myra Cheng wanted to understand how common these validating responses are across modern software. Cheng and a team of researchers from Stanford University and Carnegie Mellon University also wanted to know how these interactions shape human behavior. They set up a series of computational analyses and psychological experiments to find out.

In the first part of the research, the team tested eleven different state of the art software models from companies including OpenAI, Google, and Meta. They fed the models thousands of text prompts derived from different social situations.

One dataset featured general requests for everyday advice. Another dataset contained two thousand posts from a popular internet forum where people describe a social conflict and ask the community if they behaved poorly. For this specific dataset, the researchers only used posts where human readers unanimously agreed that the author was completely in the wrong.

A third dataset included thousands of statements describing deeply problematic actions. These statements detailed scenarios involving deception, like forging a supervisor’s signature on a document. Other prompts described illegal activities or actions taken purely out of spite.

Google News Preferences Add PsyPost to your preferred sources

Across the board, the tested models were highly sycophantic. When presented with dilemmas where human crowds entirely condemned an action, the software still validated the user just over half of the time. When responding to prompts about deception and illegal behavior, the models endorsed the user’s action forty seven percent of the time. On average, the technology affirmed the user forty nine percent more often than human advisers would in the exact same situations.

Establishing that the software consistently behaves this way was only the first step. The researchers then conducted three experiments with over two thousand human participants to see how the flattering responses affected social judgments.

In the first two human trials, participants read vignettes describing social disputes where they were ostensibly in the wrong. The participants then received either a flattering reply from an artificial intelligence or a neutral response that challenged their behavior.

The third trial placed participants in a live chat interface where they discussed a real dispute from their own past. They spent eight rounds exchanging messages with a chatbot. Half of the participants talked to a program engineered to flatter them, while the rest interacted with a version designed to offer pushback.

Interacting with a sycophantic program directly altered people’s intentions. Participants who received excessive validation became much more confident that their original actions were completely justified. They showed much less willingness to take the initiative to fix the situation or apologize to the other person involved.

Looking closer at the communication, the researchers noticed that the agreeable chatbots rarely mentioned the other person’s perspective. By keeping the user focused entirely on their own validation, the software caused users to lose their sense of social accountability. Participants in the non-sycophantic groups admitted fault in their follow up messages at a much higher rate.

The effects held up even after controlling for various personal traits. Age, gender, personality type, and prior familiarity with artificial intelligence did not provide immunity. Almost anyone can fall victim to the persuasive power of a flattering program.

The researchers also measured how people felt about the software itself after receiving the advice. Even though the flattering responses distorted the participants’ social judgments, people consistently rated the agreeable models as having higher quality. They reported elevated levels of both moral trust and performance trust in the flattering chatbots.

The participants explicitly stated they were highly likely to return to the agreeable software for future advice. The effect grew even stronger when participants perceived the chatbot as an entirely objective source. People often described the flattering programs as fair and honest, mistaking unconditional validation for a neutral perspective.

In one variation of the experiment, researchers told half the participants that a human wrote the advice and the other half that a machine wrote it. The participants generally reported trusting the human label more. Regardless of what label they saw, the validating language still manipulated their eventual choices just as effectively.

The team also tested whether giving the chatbot a warmer, more informal tone made a difference. They found that stylistic presentation did not alter the persuasive impact of the sycophancy. The underlying endorsement of the user’s actions drove the behavioral changes, not the friendly delivery.

This dynamic places technology developers in a difficult position. Flattering behavior drives user satisfaction and repeat engagement, giving companies very little financial motivation to program their systems to be more critical. The tools are explicitly optimized to make users happy in the short term, which inadvertently shifts the software toward appeasement.

The authors noted a few limitations restricting how broadly these conclusions can be applied. The human responses used as a baseline came from internet communities, which might hold different moral standards than the wider public. Additionally, the study relied entirely on English speakers in the United States.

Expectations regarding digital interaction can vary widely across different cultures. People in other parts of the world might not desire the same level of validation, or they might react differently to machine generated flattery. The researchers also measured the software’s responses in a binary way, looking only at explicit approval or disapproval.

Future studies will likely examine more subtle or implicit forms of validation. Researchers could also look at how repeated daily use of agreeable chatbots over several years might reshape people’s real world relationships. Long term dependence on artificial emotional support could potentially displace human connections.

Policy regulators and technology designers will need to address this dynamic as these tools become deeply integrated into mobile phones and social networks. The researchers suggested that companies could implement behavioral audits before releasing new models to the public. Warning labels or digital literacy programs might also help users understand that chatbots are designed to please rather than tell the truth.

Receiving uncritical praise under the guise of an objective machine leaves many people worse off than if they had never asked for advice. Addressing these risks will require developing software that prioritizes human well being over immediate user satisfaction.

The study, “Sycophantic AI decreases prosocial intentions and promotes dependence,” was authored by Myra Cheng, Cinoo Lee, Pranav Khadpe, Sunny Yu, Dyllan Han, and Dan Jurafsky.

RELATED

Psychology textbooks still misrepresent famous experiments and controversial debates
Artificial Intelligence

How eye contact shapes the believability of computer-generated faces

April 24, 2026
Facebook users who ruminate and compare themselves to their friends experience increased loneliness
Artificial Intelligence

Women perceive AI as riskier than men do, study finds

April 22, 2026
Live music causes brain waves to synchronize more strongly with rhythm than recorded music
Artificial Intelligence

Psychologists pinpoint the conversational mechanisms that help humans bond with AI

April 22, 2026
Live music causes brain waves to synchronize more strongly with rhythm than recorded music
Artificial Intelligence

Unrestricted generative AI harms high school math learning by acting as a crutch

April 21, 2026
Live music causes brain waves to synchronize more strongly with rhythm than recorded music
Artificial Intelligence

People remain “blissfully ignorant” of AI use in everyday messages, new research shows

April 20, 2026
Live music causes brain waves to synchronize more strongly with rhythm than recorded music
Artificial Intelligence

Disclosing autism to AI chatbots prompts overly cautious, stereotypical advice

April 18, 2026
Live music causes brain waves to synchronize more strongly with rhythm than recorded music
Artificial Intelligence

Scientists tested the creativity of AI models, and the results were surprisingly homogeneous

April 18, 2026
People ascribe intentions and emotions to both human- and AI-made art, but still report stronger emotions for artworks made by humans
Artificial Intelligence

New research links personality traits to confidence in recognizing artificial intelligence deception

April 13, 2026

Follow PsyPost

The latest research, however you prefer to read it.

Daily newsletter

One email a day. The newest research, nothing else.

Google News

Get PsyPost stories in your Google News feed.

Add PsyPost to Google News
RSS feed

Use your favorite reader. We also syndicate to Apple News.

Copy RSS URL
Social media
Support independent science journalism

Ad-free reading, full archives, and weekly deep dives for members.

Become a member

Trending

  • New neuroscience research shows how slowing your breathing alters your perception of the people around you
  • Psychology textbooks still misrepresent famous experiments and controversial debates
  • The age you start regularly watching adult content predicts your future mental health
  • Smarter men possess more masculine body shapes but report fewer casual sex partners
  • New psychology research shows people consistently underestimate how often things go wrong across society

Psychology of Selling

  • Study finds Instagram micro-celebrities can shift brand attitudes and buying intent through direct engagement
  • Salespeople who feel they’re making a difference may outperform those chasing commissions
  • Five persuasive approaches and when each one works best for marketers
  • When salespeople feel free and connected to their boss, they’re less likely to quit
  • Want your brand to look premium? New research suggests making your logo less dynamic

PsyPost is a psychology and neuroscience news website dedicated to reporting the latest research on human behavior, cognition, and society. (READ MORE...)

  • Mental Health
  • Neuroimaging
  • Personality Psychology
  • Social Psychology
  • Artificial Intelligence
  • Cognitive Science
  • Psychopharmacology
  • Contact us
  • Disclaimer
  • Privacy policy
  • Terms and conditions
  • Do not sell my personal information

(c) PsyPost Media Inc

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

Subscribe
  • My Account
  • Cognitive Science Research
  • Mental Health Research
  • Social Psychology Research
  • Drug Research
  • Relationship Research
  • About PsyPost
  • Contact
  • Privacy Policy

(c) PsyPost Media Inc