Subscribe
The latest psychology and neuroscience discoveries.
My Account
  • Mental Health
  • Social Psychology
  • Cognitive Science
  • Neuroscience
  • About
No Result
View All Result
PsyPost
PsyPost
No Result
View All Result
Home Exclusive Artificial Intelligence

People who use AI may pay a social price, according to new psychology research

by Eric W. Dolan
July 14, 2025
in Artificial Intelligence, Business
[Adobe Stock]

[Adobe Stock]

Share on TwitterShare on Facebook

A new study highlights a potential downside to using artificial intelligence at work: people who rely on AI tools to get their job done are often judged more harshly by others. Despite AI’s potential to boost productivity, workers may face negative social evaluations — especially around their competence, effort, and motivation — just for using it. The research, published in the Proceedings of the National Academy of Sciences, provides evidence that this perception can even affect hiring decisions.

Generative artificial intelligence refers to tools like ChatGPT and other systems that can produce humanlike text, images, or code. These tools are becoming increasingly common in workplaces across industries, offering assistance with writing, analysis, creative tasks, and more. While they can enhance performance, some people worry that relying on AI might make them look lazy or replaceable. That concern may not be unfounded.

The research team — social scientists studying how people evaluate one another in professional contexts — wanted to understand whether using AI would actually harm a person’s reputation at work. Drawing on attribution theory, which explores how people explain others’ actions, the researchers reasoned that observers might see AI use as a sign of lower ability or effort. Just as past studies have shown that people sometimes assume help-seeking reflects personal weakness rather than situational need, the team predicted that AI users might face similar biases.

“I became interested in this topic after conversations with employees at several different organizations who expressed hesitation about using generative AI at work not because of issues with its output, but because of the social dynamics associated with AI use,” explained study author Jessica Reif, a PhD candidate at Duke University Fuqua School of Business. “These discussions, coupled with industry surveys that suggested people hide their AI use at work, made me want to examine whether the social evaluation penalty employees seem to anticipate is real.”

The researchers conducted four experiments involving more than 4,400 participants. The studies were preregistered, meaning the team documented their methods and hypotheses in advance to reduce bias. Across all experiments, the researchers examined both people’s expectations about how they would be judged for using AI and how observers actually evaluated others who used AI tools.

In the first study, 497 participants imagined using either a generative AI tool or a traditional dashboard tool to complete a task at work. They then rated how they thought managers and coworkers would perceive them. People who imagined using AI expected to be judged as lazier, more replaceable, and less competent and diligent than those using non-AI tools. They also said they would be less likely to disclose their use of AI to others. These responses suggest that workers are aware of a possible stigma around AI use.

But are these fears justified? In the second study, the researchers asked 1,215 participants to read brief descriptions of hypothetical employees and rate them on various traits. The employees were described as receiving help from either AI, a human coworker, or no help at all. Participants consistently rated the employees who used AI as lazier, less competent, less diligent, and less independent than those in the other two conditions. The effect was not tied to the employee’s age, gender, or job type. These results show that people really do hold negative views of AI users, even when the help received is essentially the same.

“We were surprised that the social evaluation penalty in our studies occurred regardless of the age, occupation, or gender of the employee we described,” Reif told PsyPost. “In total, we compared 384 unique stimuli that systematically varied these attributes, in addition to whether the target was described as getting help from AI, getting help from another source, or getting no help at all. For instance, I thought it might be plausible that evaluations of AI users would vary depending on whether the employee described was early in his or her career (e.g., age 25) versus more experienced (e.g., age 40+). We did not find that to be the case.”

Google News Preferences Add PsyPost to your preferred sources

The third study tested whether these perceptions influence real decisions. In this case, one group of participants (801 people) completed a visual task and reported how often they used AI. A second group (1,718 people) played the role of hiring managers and chose between candidates who either used AI daily or not at all. They were told their pay would depend on the candidate’s performance, which gave them an incentive to choose carefully.

Overall, managers who did not use AI themselves tended to favor candidates who also did not use AI. In contrast, managers who used AI more often were more likely to prefer candidates who used AI daily. This suggests that people’s own experience with AI influences how they view others who use it. Those who are unfamiliar with the technology may be more skeptical or suspicious of its users.

The final study aimed to understand why these biases occur and whether they could be reduced. The researchers asked 1,006 participants to evaluate hypothetical candidates applying for tasks that were either manual (like handwriting notes) or digital (like sending personalized emails). Some candidates were described as using AI regularly, while others used traditional tools like Microsoft Office. Participants who did not use AI themselves were more likely to view AI users as lazy — and this perception translated into lower evaluations of job fit, especially for manual tasks.

However, when the task was digital and clearly suitable for AI assistance, the penalty disappeared. In fact, AI users were even seen as slightly better suited for the digital task than non-AI users. The researchers also found that frequent users of AI were less likely to judge AI users as lazy, suggesting that familiarity can help reduce the stigma.

“The top takeaway from this work is that AI use can carry a social cost,” Reif explained. “In our studies, employees described as using generative AI were evaluated as being lazier, less competent, and less diligent than employees using other tools or sources of help at work to perform the same task. The irony in our findings is that while some employees likely use AI because they are motivated to be more productive at work, its use may lead others to evaluate them as less motivated.”

One strength of the research is its experimental design, which allowed the team to isolate the effect of AI use from other variables. However, the authors caution that their work has some limitations. All the studies relied on online samples rather than real-world organizations, which may affect how well the results translate to actual workplaces. Also, the descriptions of AI tools were intentionally broad, which might not reflect the variety of tools and use cases that exist today.

“A key limitation of our studies is that the evaluators did not personally know the targets they were evaluating and were simply reporting their first impression upon reading about them,” Reif noted. “For example, we can’t say how perceptions about an employee who has a longstanding reputation for being a hard worker might change if he or she starts using AI. It could be the case that the effects we observe in our studies are weaker when the evaluator has more knowledge about the target or an existing work relationship.”

“One future direction I am excited about is unpacking why the social evaluation penalty we document occurs. For example, it may be the case that evaluators make assumptions about how those who use AI are spending the time they save. In addition to research on social evaluations, I am conducting research that investigates other social dynamics surrounding AI use in the workplace, such as how AI influences who seeks advice and information from whom.”

The study, “Evidence of a social evaluation penalty for using AI,” was authored by Jessica A. Reif, Richard P. Larrick, and Jack B. Soll.

Previous Post

Researchers identify a shared brain pattern behind feeling surprised

Next Post

Effort or fate? Sexual mindsets predict whether bedroom bliss spills over into broader wellbeing

RELATED

Why most people fail to spot AI-generated faces, while super-recognizers have a subtle advantage
Artificial Intelligence

Why most people fail to spot AI-generated faces, while super-recognizers have a subtle advantage

February 28, 2026
Major study reshapes our understanding of assortative mating and its generational impact
Business

A man’s psychological fit at work tends to increase when his financial values align with his partner’s

February 28, 2026
People with social anxiety more likely to become overdependent on conversational artificial intelligence agents
Artificial Intelligence

AI therapy is rated higher for empathy until people learn a machine wrote the text

February 26, 2026
New research: AI models tend to reflect the political ideologies of their creators
Artificial Intelligence

New research: AI models tend to reflect the political ideologies of their creators

February 26, 2026
Stress disrupts gut and brain barriers by reducing key microbial metabolites, study finds
Artificial Intelligence

AI and mental health: New research links use of ChatGPT to worsened psychiatric symptoms

February 24, 2026
Stanford scientist discovers that AI has developed an uncanny human-like ability
Artificial Intelligence

How personality and culture relate to our perceptions of artificial intelligence

February 23, 2026
Young children are more likely to trust information from robots over humans
Artificial Intelligence

The presence of robot eyes affects perception of mind

February 21, 2026
Psychology study reveals a fascinating fact about artwork
Artificial Intelligence

AI art fails to trigger the same empathy as human works

February 20, 2026

STAY CONNECTED

LATEST

Fascinating new research reveals your heart rate drops when your brain misperceives the world

Scientists discover psychedelic drug 5-MeO-DMT induces a state of “paradoxical wake”

Sexsomnia: How common is sleep sex?

Children of divorce develop stronger morals but face hidden emotional struggles

Altering gut bacteria with antibiotics reduces inflammation from traumatic brain injuries

Asexual women tend to prioritize different traits in a partner compared to heterosexual women

The psychological reason why dark humor isn’t for everyone

Broad claims about gender and behavior fall apart when studies include ethnically diverse samples

PsyPost is a psychology and neuroscience news website dedicated to reporting the latest research on human behavior, cognition, and society. (READ MORE...)

  • Mental Health
  • Neuroimaging
  • Personality Psychology
  • Social Psychology
  • Artificial Intelligence
  • Cognitive Science
  • Psychopharmacology
  • Contact us
  • Disclaimer
  • Privacy policy
  • Terms and conditions
  • Do not sell my personal information

(c) PsyPost Media Inc

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

Subscribe
  • My Account
  • Cognitive Science Research
  • Mental Health Research
  • Social Psychology Research
  • Drug Research
  • Relationship Research
  • About PsyPost
  • Contact
  • Privacy Policy

(c) PsyPost Media Inc