When it comes to reading a person’s state of mind, visual context—as in background and action—is just as important as facial expressions and body language, according to a new study.
The findings challenge decades of research positing that emotional intelligence and recognition are based largely on the ability to read micro-expressions signaling happiness, sadness, anger, fear, surprise, disgust, contempt, and other positive and negative moods and sentiments.
“Our study reveals that emotion recognition is, at its heart, an issue of context as much as it is about faces,” says lead author Zhimin Chen, a doctoral student in psychology at the University of California, Berkeley.
Expressions and emotion
Researchers blurred the faces and bodies of actors in dozens of muted clips from Hollywood movies and home videos. Despite the characters’ virtual invisibility, hundreds of study participants were able to accurately read their emotions by examining the background and how they were interacting with their surroundings.
The “affective tracking” model that Chen created for the study allows researchers to track how people rate the moment-to-moment emotions of characters as they view videos.
“Face it, the face is not enough to perceive emotion.”
Chen’s method is capable of collecting large quantities of data in a short time, and could eventually gauge how people with disorders like autism and schizophrenia read emotions in real time, and help with their diagnoses.
“Some people might have deficits in recognizing facial expressions, but can recognize emotion from the context,” Chen says. “For others, it’s the opposite.”
The findings, based on statistical analyses of the ratings researchers collected, could also inform the development of facial recognition technology.
“Right now, companies are developing machine learning algorithms for recognizing emotions, but they only train their models on cropped faces and those models can only read emotions from faces,” Chen says.
“Our research shows that faces don’t reveal true emotions very accurately and that identifying a person’s frame of mind should take into account context as well.”
For the study, the researchers tested the emotion recognition abilities of nearly 400 young adults. The visual stimuli they used were video clips from various Hollywood movies as well as documentaries and home videos that showed emotional responses in more natural settings.
Study participants went online to view and rate the video clips. The researchers superimposed a rating over the video so they could track each study participant’s cursor as it moved around the screen, processing visual information and rating moment-to-moment emotions.
In the first of three experiments, 33 study participants viewed interactions in movie clips between two characters, one of which researchers blurred, and rated the perceived emotions of the blurred character. The results show that study participants inferred how the invisible character was feeling based not only on their interpersonal interactions, but also from what was happening in the background.
Next, approximately 200 study participants viewed video clips showing interactions under three different conditions: one in which everything was visible, another in which researchers blurred the characters, and another in which they blurred the context. The results showed that context was as important as facial recognition for decoding emotions.
In the final experiment, 75 study participants viewed clips from documentaries and home videos so that researchers could compare emotion recognition in more naturalistic settings. Again, context was as critical for inferring the emotions of the characters as were their facial expressions and gestures.
“Overall, the results suggest that context is not only sufficient to perceive emotion, but also necessary to perceive a person’s emotion,” says senior author David Whitney, a psychology professor. “Face it, the face is not enough to perceive emotion.”
The research appears in the Proceedings of the National Academy of Sciences.
Source: UC Berkeley