Subscribe
The latest psychology and neuroscience discoveries.
My Account
  • Mental Health
  • Social Psychology
  • Cognitive Science
  • Psychopharmacology
  • Neuroscience
  • About
No Result
View All Result
PsyPost
PsyPost
No Result
View All Result
Home Exclusive Artificial Intelligence

People who use AI may pay a social price, according to new psychology research

by Eric W. Dolan
July 14, 2025
in Artificial Intelligence, Business
[Adobe Stock]

[Adobe Stock]

Share on TwitterShare on Facebook

A new study highlights a potential downside to using artificial intelligence at work: people who rely on AI tools to get their job done are often judged more harshly by others. Despite AI’s potential to boost productivity, workers may face negative social evaluations — especially around their competence, effort, and motivation — just for using it. The research, published in the Proceedings of the National Academy of Sciences, provides evidence that this perception can even affect hiring decisions.

Generative artificial intelligence refers to tools like ChatGPT and other systems that can produce humanlike text, images, or code. These tools are becoming increasingly common in workplaces across industries, offering assistance with writing, analysis, creative tasks, and more. While they can enhance performance, some people worry that relying on AI might make them look lazy or replaceable. That concern may not be unfounded.

The research team — social scientists studying how people evaluate one another in professional contexts — wanted to understand whether using AI would actually harm a person’s reputation at work. Drawing on attribution theory, which explores how people explain others’ actions, the researchers reasoned that observers might see AI use as a sign of lower ability or effort. Just as past studies have shown that people sometimes assume help-seeking reflects personal weakness rather than situational need, the team predicted that AI users might face similar biases.

“I became interested in this topic after conversations with employees at several different organizations who expressed hesitation about using generative AI at work not because of issues with its output, but because of the social dynamics associated with AI use,” explained study author Jessica Reif, a PhD candidate at Duke University Fuqua School of Business. “These discussions, coupled with industry surveys that suggested people hide their AI use at work, made me want to examine whether the social evaluation penalty employees seem to anticipate is real.”

The researchers conducted four experiments involving more than 4,400 participants. The studies were preregistered, meaning the team documented their methods and hypotheses in advance to reduce bias. Across all experiments, the researchers examined both people’s expectations about how they would be judged for using AI and how observers actually evaluated others who used AI tools.

In the first study, 497 participants imagined using either a generative AI tool or a traditional dashboard tool to complete a task at work. They then rated how they thought managers and coworkers would perceive them. People who imagined using AI expected to be judged as lazier, more replaceable, and less competent and diligent than those using non-AI tools. They also said they would be less likely to disclose their use of AI to others. These responses suggest that workers are aware of a possible stigma around AI use.

But are these fears justified? In the second study, the researchers asked 1,215 participants to read brief descriptions of hypothetical employees and rate them on various traits. The employees were described as receiving help from either AI, a human coworker, or no help at all. Participants consistently rated the employees who used AI as lazier, less competent, less diligent, and less independent than those in the other two conditions. The effect was not tied to the employee’s age, gender, or job type. These results show that people really do hold negative views of AI users, even when the help received is essentially the same.

“We were surprised that the social evaluation penalty in our studies occurred regardless of the age, occupation, or gender of the employee we described,” Reif told PsyPost. “In total, we compared 384 unique stimuli that systematically varied these attributes, in addition to whether the target was described as getting help from AI, getting help from another source, or getting no help at all. For instance, I thought it might be plausible that evaluations of AI users would vary depending on whether the employee described was early in his or her career (e.g., age 25) versus more experienced (e.g., age 40+). We did not find that to be the case.”

The third study tested whether these perceptions influence real decisions. In this case, one group of participants (801 people) completed a visual task and reported how often they used AI. A second group (1,718 people) played the role of hiring managers and chose between candidates who either used AI daily or not at all. They were told their pay would depend on the candidate’s performance, which gave them an incentive to choose carefully.

Overall, managers who did not use AI themselves tended to favor candidates who also did not use AI. In contrast, managers who used AI more often were more likely to prefer candidates who used AI daily. This suggests that people’s own experience with AI influences how they view others who use it. Those who are unfamiliar with the technology may be more skeptical or suspicious of its users.

The final study aimed to understand why these biases occur and whether they could be reduced. The researchers asked 1,006 participants to evaluate hypothetical candidates applying for tasks that were either manual (like handwriting notes) or digital (like sending personalized emails). Some candidates were described as using AI regularly, while others used traditional tools like Microsoft Office. Participants who did not use AI themselves were more likely to view AI users as lazy — and this perception translated into lower evaluations of job fit, especially for manual tasks.

However, when the task was digital and clearly suitable for AI assistance, the penalty disappeared. In fact, AI users were even seen as slightly better suited for the digital task than non-AI users. The researchers also found that frequent users of AI were less likely to judge AI users as lazy, suggesting that familiarity can help reduce the stigma.

“The top takeaway from this work is that AI use can carry a social cost,” Reif explained. “In our studies, employees described as using generative AI were evaluated as being lazier, less competent, and less diligent than employees using other tools or sources of help at work to perform the same task. The irony in our findings is that while some employees likely use AI because they are motivated to be more productive at work, its use may lead others to evaluate them as less motivated.”

One strength of the research is its experimental design, which allowed the team to isolate the effect of AI use from other variables. However, the authors caution that their work has some limitations. All the studies relied on online samples rather than real-world organizations, which may affect how well the results translate to actual workplaces. Also, the descriptions of AI tools were intentionally broad, which might not reflect the variety of tools and use cases that exist today.

“A key limitation of our studies is that the evaluators did not personally know the targets they were evaluating and were simply reporting their first impression upon reading about them,” Reif noted. “For example, we can’t say how perceptions about an employee who has a longstanding reputation for being a hard worker might change if he or she starts using AI. It could be the case that the effects we observe in our studies are weaker when the evaluator has more knowledge about the target or an existing work relationship.”

“One future direction I am excited about is unpacking why the social evaluation penalty we document occurs. For example, it may be the case that evaluators make assumptions about how those who use AI are spending the time they save. In addition to research on social evaluations, I am conducting research that investigates other social dynamics surrounding AI use in the workplace, such as how AI influences who seeks advice and information from whom.”

The study, “Evidence of a social evaluation penalty for using AI,” was authored by Jessica A. Reif, Richard P. Larrick, and Jack B. Soll.

RELATED

Stanford scientist discovers that AI has developed an uncanny human-like ability
Artificial Intelligence

Artificial intelligence helps decode the neuroscience of dance

November 28, 2025
Study finds gratitude mediates the impact of support in long-term relationships
Business

New research links “dark triad” traits to the quiet quitting phenomenon

November 28, 2025
AI chatbots often misrepresent scientific studies — and newer models may be worse
Artificial Intelligence

One in eight US adolescents and young adults use AI chatbots for mental health advice

November 26, 2025
Scientists identify a fat-derived hormone that drives the mood benefits of exercise
Artificial Intelligence

A mathematical ceiling limits generative AI to amateur-level creativity

November 24, 2025
Study identifies creativity and resilience as positive aspects of ADHD diagnosis
Business

Large-scale trial finds four-day workweek improves employee well-being and physical health

November 22, 2025
Study reveals AI’s potential to detect loneliness by deciphering speech patterns
Artificial Intelligence

How generative AI could change how we think and speak

November 20, 2025
AI can already diagnose depression better than a doctor and tell you which treatment is best
Artificial Intelligence

Study finds nearly two-thirds of AI-generated citations are fabricated or contain errors

November 20, 2025
Humans create more novelty than ChatGPT when retelling stories
Artificial Intelligence

Social reasoning in AI traced to an extremely small set of parameters

November 18, 2025

PsyPost Merch

STAY CONNECTED

LATEST

How the brain transforms continuous sound into distinct words

Whom you observe in your daily life alters your willingness to tax the rich

Artificial intelligence helps decode the neuroscience of dance

Psychologists say climate anxiety is a form of pre-traumatic stress

Specific depression symptoms linked to distinct patterns of inflammation and cognitive deficit

Single gene mutation linked to increased alcohol tolerance and consumption

New research links “dark triad” traits to the quiet quitting phenomenon

A common amino acid reduces brain plaques in animal models of Alzheimer’s disease

RSS Psychology of Selling

  • Brain wiring predicts preference for emotional versus logical persuasion
  • What science reveals about the Black Friday shopping frenzy
  • Research reveals a hidden trade-off in employee-first leadership
  • The hidden power of sequence in business communication
  • What so-called “nightmare traits” can tell us about who gets promoted at work
         
       
  • Contact us
  • Privacy policy
  • Terms and Conditions
[Do not sell my information]

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

Subscribe
  • My Account
  • Cognitive Science Research
  • Mental Health Research
  • Social Psychology Research
  • Drug Research
  • Relationship Research
  • About PsyPost
  • Contact
  • Privacy Policy