Subscribe
The latest psychology and neuroscience discoveries.
My Account
  • Mental Health
  • Social Psychology
  • Cognitive Science
  • Psychopharmacology
  • Neuroscience
  • About
No Result
View All Result
PsyPost
PsyPost
No Result
View All Result
Home Exclusive Artificial Intelligence

Smash or pass? AI could soon predict your date’s interest via physiological cues

by Eric W. Dolan
February 19, 2024
in Artificial Intelligence, Social Psychology
[Adobe Stock]

[Adobe Stock]

Share on TwitterShare on Facebook

In a world where technology touches almost every aspect of our lives, it’s not surprising that researchers are now exploring how it can help us understand the nuances of human interaction, particularly in the realm of conversation. A recent study by engineers at the University of Cincinnati has made a fascinating leap in this direction, suggesting that soon, artificial intelligence could potentially tell you if your first date is really into you — or not — based on physiological responses alone.

By analyzing data from wearable technology that tracks heart rates, respiration, and perspiration, the study has managed to classify the type of conversation two people are having with notable accuracy. This research, published in the journal IEEE Transactions on Affective Computing, not only opens up new possibilities for enhancing interpersonal communication but also paves the way for applications in mental health counseling and education.

The inspiration for this study stemmed from a keen interest in physiological synchrony — a phenomenon where individuals’ physiological responses, such as heart rate and breathing patterns, become synchronized during conversation or collaboration. This synchrony is believed to be a marker of how engaged or in tune two people are with each other during an interaction.

Previous studies have shown that the degree of physiological synchrony can predict how well individuals collaborate on tasks, the empathy patients feel from their therapists, and the engagement students experience with their teachers.

Researchers embarked on this study to explore whether technology could leverage physiological synchrony to infer the dynamics of a conversation without needing verbal feedback or observation. In a world increasingly reliant on digital communication, understanding the underlying emotional connections between individuals could significantly enhance virtual interactions, making this research both timely and relevant.

Eighteen pairs of participants were initially recruited, with the final analysis focusing on sixteen dyads due to data quality issues.

These pairs, encompassing friends, roommates, coworkers, relationship partners, and strangers, engaged in four distinct conversation scenarios: two-sided positive, two-sided negative, one-sided with the person on the left talking, and one-sided with the person on the right talking. These scenarios were crafted to elicit a range of emotional and physiological responses, which were then recorded using wearable sensors measuring heart rate, skin conductance, and respiration.

Participants were also asked to report their feelings using the Self-Assessment Manikin (SAM) and the Interpersonal Interaction Questionnaire (IIQ) to provide a comprehensive picture of each conversation’s emotional landscape.

University of Cincinnati engineering students demonstrate how they taught a computer to distinguish types of conversations based only on physiological cues. (Photo credit: Andrew Higley/University of Cincinnati)

The researchers applied advanced data processing techniques to filter and analyze the signals, extracting features related to both individual responses and the synchrony between participants. This rich dataset served as the input for machine learning algorithms, which were tasked with classifying the type of conversation based on physiological data alone.

The researchers found that a two-stage classification approach, which first distinguished between one-sided and two-sided conversations before further classifying them, achieved the highest accuracy of 75%. This indicates a strong potential for technology to understand conversation dynamics through physiological signals.

However, when the researchers relied solely on physiological data without the additional context provided by synchrony features, the accuracy dropped to 65.6%, highlighting the importance of these features in capturing the nuances of human interaction.

Interestingly, incorporating personality traits into the analysis did not improve the classification accuracy, suggesting that the structured nature of the conversation scenarios may have limited the impact of individual personality differences. The most successful classifications came from using a combination of physiological signals, with heart rate variability, skin conductance responses, and respiration rates among the most informative features.

The potential applications of this research are varied. Lead author and University of Cincinnati doctoral student Iman Chatterjee suggested that the technology developed through their study could serve as a tool for honest feedback in social interactions, effectively playing a “smash or pass” role in evaluating the dynamics of a date or any conversation.

“The computer could tell if you’re a bore,” he explained, highlighting how a modified version of their system might measure a person’s interest level, compatibility, and engagement in a conversation.

But despite the study’s significant achievements, the researchers are quick to acknowledge its limitations and the need for further exploration. One major consideration is the high accuracy (96.9%) achieved using self-report data, raising questions about the added value of physiological measures in contexts where individuals are already aware of their interaction dynamics. This suggests that future research should focus on more spontaneous conversation scenarios that might offer new insights into the role of physiological synchrony.

Moreover, the study’s structured conversation scenarios, while useful for initial exploration, may not fully capture the complexity of natural human interactions, which are often more nuanced and less predictable. Future studies could benefit from incorporating additional data types, such as speech patterns, gestures, and even EEG signals, to enrich the analysis and potentially improve classification accuracy.

The study, “Automated Classification of Dyadic Conversation Scenarios using Autonomic Nervous System Responses,” was authored by Iman Chatterjee, Maja Goršič, Mohammad S. Hossain, Joshua D. Clapp, and Vesna D. Novak.

RELATED

Media coverage of artificial intelligence split along political lines, study finds
Artificial Intelligence

Survey reveals rapid adoption of AI tools in mental health care despite safety concerns

December 13, 2025
Harrowing case report details a psychotic “resurrection” delusion fueled by a sycophantic AI
Developmental Psychology

Social dominance orientation emerges in early childhood independent of parental socialization, new study suggests

December 13, 2025
Harrowing case report details a psychotic “resurrection” delusion fueled by a sycophantic AI
Artificial Intelligence

Harrowing case report details a psychotic “resurrection” delusion fueled by a sycophantic AI

December 13, 2025
Scientists say X (formerly Twitter) has lost its professional edge — and Bluesky is taking its place
Political Psychology

What are legislators hiding when they scrub their social media history?

December 12, 2025
Higher diet quality is associated with greater cognitive reserve in midlife
Social Psychology

Older adults who play pickleball report lower levels of loneliness

December 12, 2025
Psychosocial stress triggers an oxytocin response in women, study finds
Business

Oxytocin curbs men’s desire for luxury goods when partners are ovulating

December 12, 2025
Scientists just uncovered a major limitation in how AI models understand truth and belief
Parenting

Encouraging parents to plan sex leads to more frequent intimacy and higher desire

December 11, 2025
Scientists just uncovered a major limitation in how AI models understand truth and belief
Cognitive Science

New review challenges the idea that highly intelligent people are hyper-empathic

December 11, 2025

PsyPost Merch

STAY CONNECTED

LATEST

Amphetamine overrides brain signals associated with sexual rejection

Survey reveals rapid adoption of AI tools in mental health care despite safety concerns

New research maps how the brain processes different aspects of life satisfaction

Social dominance orientation emerges in early childhood independent of parental socialization, new study suggests

Researchers uncover a distinct narrative pattern in autistic people and their siblings

New study reveals how vulvar appearance influences personality judgments among women

Harrowing case report details a psychotic “resurrection” delusion fueled by a sycophantic AI

What are legislators hiding when they scrub their social media history?

RSS Psychology of Selling

  • Mental reconnection in the morning fuels workplace proactivity
  • The challenge of selling the connected home
  • Consumers prefer emotionally intelligent AI, but not for guilty pleasures
  • Active listening improves likability but does not enhance persuasion
  • New study maps the psychology behind the post-holiday return surge
         
       
  • Contact us
  • Privacy policy
  • Terms and Conditions
[Do not sell my information]

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

Subscribe
  • My Account
  • Cognitive Science Research
  • Mental Health Research
  • Social Psychology Research
  • Drug Research
  • Relationship Research
  • About PsyPost
  • Contact
  • Privacy Policy