Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 14 August 2023
Sec. Emotion Science
This article is part of the Research Topic Current Trends on Emotional Processing: Contributions of Acquired Brain Injury, Neurodegenerative Diseases, and Ageing View all 5 articles

Facial affect recognition in context in adults with and without TBI

Lyn S. Turkstra
Lyn S. Turkstra1*Sam Hosseini-MoghaddamSam Hosseini-Moghaddam2Sophie WohltjenSophie Wohltjen3Sara V. NurreSara V. Nurre4Bilge MutluBilge Mutlu3Melissa C. DuffMelissa C. Duff5
  • 1Faculty of Health Sciences, McMaster University, Hamilton, ON, Canada
  • 2University of Toronto Schools, Toronto, ON, Canada
  • 3Department of Computer Sciences, University of Wisconsin-Madison, Madison, WI, United States
  • 4American Speech-Language-Hearing Association, Rockville, MD, United States
  • 5Department of Hearing and Speech Sciences, Vanderbilt University, Nashville, TN, United States

Introduction: Several studies have reported impaired emotion recognition in adults with traumatic brain injury (TBI), but studies have two major design features that limit application of results to real-world contexts: (1) participants choose from among lists of basic emotions, rather than generating emotion labels, and (2) images are typically presented in isolation rather than in context. To address these limitations, we created an open-labeling task with faces shown alone or in real-life scenes, to more closely approximate how adults with TBI label facial emotions beyond the lab.

Methods: Participants were 55 adults (29 female) with moderate to severe TBI and 55 uninjured comparison peers, individually matched for race, sex, and age. Participants viewed 60 photographs of faces, either alone or in the pictured person’s real-life context, and were asked what that person was feeling. We calculated the percent of responses that were standard forced-choice-task options, and also used sentiment intensity analysis to compare verbal responses between the two groups. We tracked eye movements for a subset of participants, to explore whether gaze duration or number of fixations helped explain any group differences in labels.

Results: Over 50% of responses in both groups were words other than basic emotions on standard affect tasks, highlighting the importance of eliciting open-ended responses. Valence of labels by participants with TBI was attenuated relative to valence of Comparison group labels, i.e., TBI group responses were less positive to positive images and the same was true for negative images, although the TBI group responses had higher lexical diversity. There were no significant differences in gaze duration or number of fixations between groups.

Discussion: Results revealed qualitative differences in affect labels between adults with and without TBI that would not have emerged on standard forced-choice tasks. Verbal differences did not appear to be attributable to differences in gaze patterns, leaving open the question of mechanisms of atypical affect processing in adults with TBI.

1. Introduction

Traumatic brain injury (TBI) is a leading cause of death and disability worldwide, with an estimated annual incidence of 69 million (Dewan et al., 2018). While individuals with milder injuries may recover pre-injury function, more than 20% of survivors will live with chronic impairments in sensorimotor and cognitive functions (Masel and DeWitt, 2010; Blennow et al., 2016).

Impairments in social functioning are among the most pervasive and disabling consequences of TBI, affecting up to 70% of all survivors (Benedictus et al., 2010). Indeed, adults with TBI report fewer regular social contacts than their uninjured peers, less social participation, and more isolation (Stocchetti and Zanier, 2016), all of which have negative effects on employment, health, and quality of life (McLean et al., 2014). Contributing to negative social outcomes are impairments is the ability to recognize emotions from others’ faces, voices, and body postures (Babbage et al., 2011; Zupan and Neumann, 2014; Byom et al., 2019; Neumann and Zupan, 2019; Turkstra et al., 2020). These impairments are particularly prevalent among those with moderate or severe TBI (Murphy et al., 2021), although they have been reported in some individuals with mild TBI (concussion) as well (Theadom et al., 2019). Impairments in emotion recognition can have a significant negative effect on broader social outcomes (Genova et al., 2017; Milders, 2018; Rigon et al., 2018; Rosenberg et al., 2018; Binder et al., 2019; Sherer et al., 2022), may play a role in rehabilitation of non-emotional functions (Spikman et al., 2013), and may be remediable (Cassel et al., 2019; Vallat-Azouvi et al., 2019). As a result, assessment of the ability to recognize emotions has been recommended for clinical management of patients with TBI (Togher et al., 2023).

While assessment of affect recognition is important, existing clinical and research assessment tools have several limitations. First, stimuli are typically faces presented in isolation. As Aviezer et al. (2017) observed, the notion that we can infer others’ emotions from their faces alone “is deeply ingrained in lay intuition, popular culture and scientific thought” (p. 47). Indeed, faces alone do not capture the rich visual, cultural, and social contexts in which we interpret affective displays in everyday life (Barrett et al., 2011; van Kleef et al., 2016). Studies over the past three decades have shown that recognition and interpretation of emotions is highly influenced by—and in many cases dependent on—context cues (Carroll and Russell, 1996; de Gelder et al., 2006; Righart and de Gelder, 2008a,b; Wieser and Brosch, 2012; Schwarz et al., 2013; Nelson and Mondloch, 2019). Context cues can be within the expresser (e.g., body posture) (Aviezer et al., 2012), within the scene (e.g., background cues to events) (de Gelder et al., 2006), or within the observer (e.g., the person’s mood) (Aviezer et al., 2017), and may override facial affect cues in some contexts (Kayyal et al., 2015). Context cues are dynamic over people, time and space, and might be particularly important when facial affect cues are ambiguous or the emotion is nuanced and complex (Aviezer et al., 2017).

When context effects have been studied, stimuli have been relatively primitive, such as superimposing a face on an unrelated complex scene (de Gelder et al., 2006), presenting an isolated face image after a written comment (Schwarz et al., 2013), or presenting a foreground face with other faces in the periphery that vary in gaze direction and affect (Mumenthaler and Sander, 2012). While these contextually embedded stimuli are an improvement over isolated faces, their ecological validity, and hence their clinical utility, is limited because artificially juxtaposed images are not something typically encountered in everyday life.

A second limitation of most tools is that response choices are derivatives of the stereotyped “basic” emotions proposed by Woodworth almost a century ago (Woodworth, 1938), i.e., happy, sad, afraid, surprised, disgusted, and angry. These basic emotions were popularized in research by Ekman and others, beginning in the 1960s (Ekman and Friesen, 1967). Ekman and colleagues argued that the basic emotions were “universal” (Ekman et al., 1969; Ekman, 1992), and thus should be used in emotion recognition research (Ekman et al., 1972). The Ekman and Friessen black-and-white photographs were the gold standard stimuli for decades of research, and basic emotions continue to dominate experimental stimulus sets. Studies using basic emotion stimuli have been fruitful, as results have shown impairments in many neurological populations, including TBI (Schmidt et al., 2010; Babbage et al., 2011), as well as multiple sclerosis (Henry et al., 2009; Charvet et al., 2014), brain tumors (Mu et al., 2012), stroke (Yuvaraj et al., 2013), Parkinson disease (Heller et al., 2014), Huntington’s Disease (Kipps et al., 2007), frontotemporal dementia (Kumfor and Piguet, 2012), and alcohol use disorders (Pabst et al., 2022). It has been increasingly recognized, however, that basic emotions represent a fraction of felt and displayed emotions in everyday life (Fernández-Dols and Crivelli, 2013, 2015; Aviezer et al., 2017). Despite evidence of their construct limitations, basic emotion categories continue to dominate clinical and experimental stimuli, even when using new technology such as 3D imaging (Lott et al., 2022) and virtual reality (Geraets et al., 2021).

One category of emotions that is common in everyday life but typically not included in test stimuli is social emotions. Social emotions can be defined as emotions that are interpretable only in the context of social information, including information in the social context and interactions with others, as well as inferences about others’ mental states (Buck, 1988; Adolphs, 2002). Social emotions can be negative, such as grief and despair, or positive, such as admiration and pride (Tamietto et al., 2007). As with basic emotions, there is strong evidence that recognition of social emotions is impaired in many neurological populations, including TBI (Turkstra et al., 2018), schizophrenia (Bora et al., 2006), epilepsy (Broicher et al., 2012), multiple sclerosis (Charvet et al., 2014), and Huntington’s Disease (Eddy et al., 2012). By definition, social emotions are understood in a social context, so stimuli to test recognition of these emotions should include features of social context.

The third limitation of most assessment tools is that the response modality is forced choice, where participants are asked to select the word that best describes the emotion shown by a person in a photograph or video. Choices typically are the same basic emotions listed above or a mix of basic and social emotions. A historical example of the latter is the Reading the Mind in the Eyes test (Baron-Cohen et al., 2001a), which includes variants of basic emotion words (e.g., sad vs. despondent, angry vs. annoyed, afraid vs. terrified) and social emotion words (e.g., disappointed, jealous), and also cognitive state terms (e.g., bored, preoccupied), mental action terms (e.g., fantasizing), personality characteristics (e.g., shy, decisive), and judgments about the person rather than what they are feeling (e.g., arrogant, friendly). Several studies have shown evidence of impairments on forced-choice like the Eyes Test in adults with TBI (e.g., Havet-Thomassin et al., 2006; Turkstra, 2008; Muller et al., 2009; Ubukata et al., 2014; Rodriguez-Rajo et al., 2022), but it is not clear that the tools assess what respondents think vs. how well they can map their thoughts to response choices.

Zupan et al. (2022) argued that forced-choice tasks using only basic emotions are unlikely to “capture nuances in how people think about and perceive emotion” (p. 3). Their argument was based on the potential mismatch between an individual’s emotional lexicon, which is likely to be idiosyncratic, and the standard response options on emotion recognition tasks. As an illustration, Zupan et al. asked typical adults to label emotions using an open-ended response format, then asked novel raters to categorize those responses according to the basic emotion categories happy, sad, angry, fearful, and neutral. Degree-of-fit ratings were calculated for each word, based on the frequency with which raters assigned that word to each category. Results showed that other than happy, which was the only positively valenced word, the degree of fit was low across categories, i.e., the words people generated in the open-labeling task did not fit neatly into the basic emotion categories.

In a precursor to the present study, Turkstra et al. (2017) presented photographs of faces in isolation or in real-life scenes to a sample of university students, and asked participants what they thought the person in the photograph was feeling. Only 28% of responses were basic emotion terms. More than one third of those were “happy” or “happiness,” consistent with the findings of Zupan et al. (2022) and other studies showing happy is the easiest emotion to label (Rosenberg et al., 2014; Hayes et al., 2020). Other studies have likewise shown content differences between open- vs. forced-choice affect labeling in typical children (e.g., Cassels and Birch, 2014) and adults (e.g., Winters, 2005). These results suggest that forced-choice formats test an individual’s ability to identify which response option is most like what they think the person in the photograph is feeling, rather than what they actually think the person is feeling. This limitation could be particularly problematic for individuals with TBI, who are known to have challenges with inference (Johnson and Turkstra, 2012) and decision making (Bonatti et al., 2008).

Cassels and Birch (2014) compared children’s responses on open- vs. forced-choice versions of the Eyes Test described above (Baron-Cohen et al., 2001b). Eyes Test scores were higher for the closed-ended than open-ended version, but scores on the latter had higher correlations with constructs like empathy that the Eyes Test should measure. The open-ended version was more sensitive to group differences between children with and without learning disabilities, and had a lower correlation with vocabulary test scores, i.e., scores were not confounded by vocabulary ability. Minimizing vocabulary confounds is particularly important in TBI research, given the verbal recall challenges often associated with TBI.

In summary, while existing emotion recognition tasks have shown differences between adults with and without TBI, characteristics of task structure limit our understanding of how people with TBI identify emotions “in the wild.” To address limitations of existing tasks, we created an open-labeling task using complex visual scenes from real-life photographs, and asked adults with and without TBI to state what the person in each photograph was feeling. Based on previous studies in TBI, we expected group differences in the content of emotion labels. As the images were from real-life situations, we also expected that both groups would use proportionately fewer basic emotion words and more social emotion and other words (e.g., cognitive-state terms and evaluative terms). For insight into any group differences, a subset of participants completed the task with eye-tracking equipment, so we could analyze where participants looked in the images. We hypothesized that if verbal responses of adults with TBI differed qualitatively from those of adults without TBI, that difference might be attributable to TBI group participants looking at the different places in the image, specifically looking either longer or more frequently at the face than at the context in which that face was situated.

2. Materials and methods

2.1. Participants

Participants were 55 adults (29 females) with moderate–severe TBI and a comparison group of 55 uninjured adults, matched individually for age ± 5 years, race, and sex. If the participant reported that their TBI occurred after they completed their formal education, they were matched for years of education. If the injury occurred when the participant with TBI was in school, participants were matched on educational trajectory, operationalized as typical grades in school and intentions to pursue further schooling. All participants were recruited from the Midwestern United States as part of a larger study of social cognition in adults with TBI.

TBI severity was defined according to standard injury criteria (Malec et al., 2007), i.e., a loss of consciousness of 30 min or more or post-traumatic amnesia of 24 h or more, or a lowest Glasgow Coma Scale (Teasdale and Jennett, 1974) score of less than 13 in the first 24 h; and evidence of cortical or brainstem damage. Other inclusion criteria were self-identification as a Native English speaker and no reported history of a diagnosis of language or learning disability or neurological disorder affecting the brain, other than the TBI. Exclusion criteria were failing a pure-tone hearing screening test at 20 dB HL at 500, 1,000, 2,000, and 4,000 Hz; failing standard screenings for far and near vision; or testing in the aphasic range on the Western Aphasia Battery Bedside Screening Test (Kertesz, 2006).

2.2. Tasks

2.2.1. Emotion-in-context task

The emotion-in-context (EIC) task was comprised of 60 photographs from the LIFE Magazine online archives, chosen because they appeared to be emotionally evocative, captured people in real-life scenarios, showed at least one individual with a clearly visible facial affect display, had visual contexts that could influence interpretation of affective displays, and appeared to depict a range of basic and social emotions (emotion types were not determined a priori, as that was the goal of the study). Each photograph was presented in one of two formats: as a full photograph in its original form, with a one-inch square box drawn around the face to be labeled (face-in-context items, FC), and as a cropped image of only the face (face only, FO). Fifty-seven of the 60 faces (boxed in the FC condition or alone in the FO condition) were sized at one-to-two inches per side. Three images of faces alone had one or two sides that were three inches, a technical error that will be discussed in the limitations. Sample FC and FO stimuli are shown in Figure 1. Photograph order was randomized then fixed, so that each photograph was randomly assigned to appear in either the FC or FO condition.

FIGURE 1
www.frontiersin.org

Figure 1. Sample stimuli with face only and face in context. Image source: Interim Archives / RV Spencer. Reproduced with permission from Getty Images.

The EIC task was administered individually via a laptop computer, in a quiet room. Faces in the FO condition and boxed faces in the FC condition were centered on the screen (i.e., scenes were displayed so that the boxed face was at the center of the screen). Participants were seated approximately 16 inches from the laptop display and fitted with the SMI eye-tracking glasses (2012). The table was fitted with a chin-rest and height was adjusted to be comfortable for each participant. Each participant completed a standard calibration protocol, then began the task. For each image the participant was asked, “What emotion is this person feeling?” Spoken responses were written down verbatim by a research assistant seated out of sight of the participant. Participants completed the task at their own pace.

2.2.2. Cognitive tests

To compare the present study to previous publications, participants completed a series of tasks recommended by the Common Data Elements Committee for TBI research (Wilde et al., 2010). These were the California Verbal Learning Test (CVLT) (Delis et al., 2000), Wechsler Adult Intelligence Scales Processing Speed Index tests (WAIS-PSI) (Wechsler, 2008), and Trail-Making Tests A and B (Tombaugh, 2004).

2.3. Procedure

Participants completed the tests to characterize cognitive functions in the first or second session of the larger study, in a randomized order. The EIC task was scheduled on the second day of testing, after a non-emotion task, to avoid priming of affective responses. Participants had, however, completed two forced-choice emotion recognition tasks in a previous testing session that occurred 1 day to two-and-a-half months prior to the EIC, depending on participant availability and scheduling constraints. Potential effects of that earlier testing will be considered in the Limitations section.

The relevant institutional review boards approved all procedures. The authors assert that all procedures contributing to this work comply with the ethical standards of the relevant national and institutional committees on human experimentation and with the Helsinki Declaration of 1975, as revised in 2008.

2.4. Analysis

2.4.1. Verbal responses

Prior to analysis, all multiword responses were reduced to single words to eliminate any potential response length effects before analysis with Python’s natural language toolkit (Bird et al., 2009). Our reduction rule was to take the first label in each response (e.g., “angry, afraid” = angry), to avoid making mental state inferences about participants’ intents (e.g., that they were self-correcting).

2.4.2. Sentiment analysis

We ran a sentiment intensity analysis using Python’s natural language toolkit (nltk) package to determine whether the valence and intensity of the emotion labels produced for each picture differed between participants in the TBI vs. comparison group. Nltk’s sentiment intensity analysis relies on the Valence-Aware Dictionary for Sentiment Reasoning model (Hutto and Gilbert, 2014), which uses a dictionary of lexical features and their corresponding human-rated emotional intensities to determine the sentiment of new text passages. For each input text passage, nltk outputs a “sentiment score” between −1 and 1, which is the normalized sum of the emotional intensities of all lexical features included in the input. Sentiment scores near −1 correspond to intense, negative sentiments, and sentiment scores near 1 correspond to intense, positive sentiments.

We hypothesized that the TBI group’s responses might differ from those of comparison participants in both valence directions—that is, we did not expect TBI responses to be only more positive or only more negative. To better understand how TBI responses differed, we calculated the “true” valence for each image (positive, negative, or neutral). These “true” valences were determined by two independent researchers individually rating their perceived valence for each image, and then comparing their ratings, ensuring they agreed on the valence for each image included. Disagreements were resolved by a third researcher.

We obtained sentiment scores for each participant’s response to each image and performed a linear mixed effects analysis predicting each participant’s sentiment score from their experimental group (TBI vs. control) using the lme4 package in R (Bates et al., 2014). Participants and images were entered into the model as random intercepts, and experimental group, true valence, and their interaction term were entered into the model as fixed effects. We included linear and quadratic contrasts for true valence in order to more clearly understand any directional differences we might find. Sex (male or female) was also entered into the model as a fixed effect, for an exploratory analysis based on mixed evidence of sex differences in emotion labeling (Turkstra et al., 2020). We noted that one image included in the original dataset was pixelated and difficult to see, potentially inhibiting their ability to see what emotion was depicted. Thus, participants’ responses to this image were removed from further analysis.

2.4.2.1. Proportion of basic vs. social emotion words

We calculated the number of responses in the six basic emotion categories—happy, sad, disgusted, angry, afraid, and neutral—using wildcards to capture spelling and morphological variations (e.g., happy/happiness/happiness), and calculated percent of total responses for each word or spelling variant. We included either afraid or fear because affect labeling tasks commonly include either. To ensure that any group differences were not due to injury-related effects on word-finding in participants with TBI, we also calculated type-token ratio as a measure of lexical diversity in both groups. Data were summarized descriptively.

2.5. Gaze data

Gaze data were available for a subset of 37 participants: 18 in the TBI group (female = 9), and 19 in the comparison group (female = 12). For this analysis we were interested in whether participants looked at context cues, and thus used only the 30 FC stimuli. Trained research assistants coded the eye-tracking data from SMI Software (2012). For each fixation in each image, coders labeled the location of the fixation as Face (F), Scene (S), or Other (O), and calculated total fixation time and number of fixations for each area. Both total fixations and the number of fixations were calculated as both have been used as measures as attention.

Total fixation time and total number of fixations were compared between groups using a multivariate analysis of variance with TBI status, sex, and area of interest (face or scene) as independent variables and total number and duration of fixations as dependent variables.

3. Results

3.1. Demographics and cognitive test scores

Participant characteristics are shown in Table 1, including scores from cognitive tests recommended for TBI research (Wilde et al., 2010). Participants were predominantly Caucasian (n = 104), with 2 participants who self-identified as African American and 2 who self-identified as of mixed race. Analysis of variance (ANOVA) revealed a significant between-groups difference on all neuropsychological measures (p’s < 0.001). There were no significant sex-based differences on any measure except Trails A, and no significant interaction of group and sex (all p’s > 0.05). For Trails A, scores for women were significantly higher than scores for men, F(1, 109) = 6.71, p < 0.05. As there was no significant interaction of sex by group, this difference was not considered further in analysis.

TABLE 1
www.frontiersin.org

Table 1. Demographic characteristics.

3.2. Verbal responses

3.2.1. Sentiment analysis

We found a significant main effect of true valence, t(75.1) = 14.78, β = 1.15, p < 0.001. This main effect was significant for the included linear contrast, suggesting that images with negative true valence were more likely to receive lower sentiment scores and images with positive true valence were more likely to receive higher sentiment scores, validating sentiment scores we calculated from participants’ responses.

We also found a significant interaction between the TBI and comparison groups, and true valence (positive, neutral, or negative), t(5,641) = −4.37, β = −0.21, p < 0.001. This interaction was significant for the included linear contrast, indicating that the positive relationship between true valence and sentiment score was slightly attenuated for the TBI group compared to the comparison group. That is, TBI group responses to positive images were less positive than those of the comparison group, and their responses to negative images were less negative than those of the comparison group. The main effect of true valence, as well as the interaction between true valence and experimental group, are shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Boxplot showing participants’ mean sentiment scores for negatively valenced, neutral, and positively valenced images. Sentiment score means are shown in dark gray for the comparison group and light gray for the TBI group. ***p <0.001.

There were eight positively-valenced images with particularly high sentiment scores (M = 0.51, SD = 0.17). We were concerned that these images might be skewing the results in our initial analysis, so we completed a follow-up analysis with these images excluded. Results showed the same main effect of true valence, t(64.57) = 9.72, β = 1.1, p < 0.001; and interaction between true valence and group, t(4,852) = −3.82, β = −0.26, p < 0.001; suggesting that our effects were not driven by these images alone.

We also found a significant interaction between sex (male or female), and true valence (positive, neutral, or negative), t(5,639) = 2.81, β = 0.13, p = 0.005. This interaction was significant for the included linear contrast, indicating that the positive correlation between true valence and sentiment score was higher for the females compared to males. This higher correlation was primarily driven by female participants’ responses to positive images, and post hoc t-tests comparing males and females within each of the three true valence conditions confirmed this relationship [negative valence: t(2303.4) = 1.64, p = 0.1; neutral valence: t(1936.4) = −0.15, p = 0.88; positive valence: t(1310.4) = −4.49, p < 0.001]. These relationships are illustrated in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Boxplot depicting participants’ mean sentiment scores for negatively valenced, neutral, and positive images. Sentiment score means are shown in dark gray for males and light gray for females. ***p <0.001.

When reducing multi-word responses to single words for analysis, we observed that participants with TBI appeared to give more multi-word responses than the comparison group. The total number of multi-word responses in the TBI group was 258, median = 2, mode = 0, range = 0–30; and the total in the comparison group was 153, median = 1, mode = 0, range = 0–15. A median test showed this difference was not significant, chi2(1) = 3.05, Pr = 0.08. Two participants in the comparison group and seven in the TBI group generated more than 10 multi-word answers.

Three individuals in the TBI group also showed patterns that were not observed in the comparison group: one participant responded with “angry” for 22/60 items, vs. a range of 0–11 items for the remaining 99 participants (mode = 2); one participant said “serious” for 34/60 items and another participant said the same word for 14/60 items, vs. 0–10 for the remaining 98 participants (mode = 0); and one participant responded to 15 items with a description of what the person was doing (e.g., looking up at something, posing for a picture, waiting for someone to take a picture), a pattern that was not seen in any other participant in either group.

3.2.1.1. Proportion of basic emotion words

The percent of possible responses for each group (n = 3,300 per group) in each of the five basic emotion categories is shown in Table 2. Overall, less than half of responses were basic emotion words. Type-token ratio was 0.15 in the comparison group and 0.20 in the TBI group, i.e., overall, participants with TBI generated more different words than their uninjured peers.

TABLE 2
www.frontiersin.org

Table 2. Percent of possible responses in each group in the five basic emotion categories.

3.2.2. Gaze data

For the subset of participants with gaze data, there was a main effect of area of interest on number and total time of fixations, F(2, 65) = 11.28, p < 0.001; no significant effect of group, F(2, 65) = 1.28, p = 0.28, or sex, F(2, 65) = 0.50, p = 0.61; and no significant interaction of group by area of interest, F(2, 71) = 0.41, p = 0.66. Follow-up univariate tests showed a significant effect of area of interest on both number of fixations, F(1, 75) = 10.80, p < 0.005; and total fixation time, F(1, 75) = 23.70, p < 0.0001, with more fixations and time in the face region than the scene. Summary data for total number and duration of fixations are shown in Table 3. As there was no main effect of sex, data from male and female participants are combined in the table.

TABLE 3
www.frontiersin.org

Table 3. Average total number and duration of fixations on the scene vs. face for FC stimuli (n = 30), for participants in the TBI and comparison groups.

4. Discussion

We showed photographs of faces alone or in real-life scenes to adults with or without TBI, and asked them to label what the pictured people were feeling. We chose an open-ended response format to capture what participants were thinking, rather than how well they could map what they were thinking to our researcher-generated categories of emotions. We predicted qualitative differences in verbal responses between adults with and without TBI, and that eye-tracking results would help explain those differences. Our predictions were partly supported: there were indeed qualitative differences in verbal responses between the two groups, but not in gaze patterns. We also found that fewer than half of responses were the “basic” six emotions traditionally tested in affect recognition studies, and about half of those were variants on the word “happy.” Our results add to the literature on affect recognition in adults with moderate or severe TBI, and raise questions about the ecological validity of forced-choice tasks for assessment of affect recognition after TBI.

4.1. Verbal responses

We hypothesized that adults with TBI would generate qualitatively different affect labels than their uninjured peers. Sentiment analysis of open-ended responses revealed that the valence of responses by adults with TBI was attenuated relative to that of their uninjured peers. For example, for one study image many comparison group responses were “confused,” “worried,” or “anxious.” For the same image, the most common response in the TBI group was “bored,” with rare responses of “curious” or even “interested.” It could be argued that the difference in valence was due to language limitations in the TBI group, i.e., that participants with TBI were less able to generate words for stronger emotions. This explanation is unlikely, as all participants passed a screening test for aphasia and participants were closely matched on education (or pre-injury educational trajectory if the injury occurred while the participant with TBI was in school), and perhaps most relevant, lexical diversity was actually higher in the TBI group than the comparison group. Also arguing against a lexical interpretation is what Cassels and Birch (2014) referred to as the “minimum verbal ability required to simply speak about emotions” (p. 16). Emotion labels tend to be high-frequency words, and among types of mental-state terms they are learned relatively early in life (Bretherton and Beeghly, 1982), thus it is unlikely that adults with TBI used attenuated words because they were unable to generate fewer words specifically for stronger emotions.

Attenuated valence is consistent with the emotional blunting and apathy that are commonly described in adults with moderate–severe TBI (Lane-Brown and Tate, 2011). Evidence includes a study using an online social game (Kelly et al., 2013), in which adults with TBI reported less hurt feelings than their uninjured peers when they were ostracized by other players. Similarly, in a study using evocative film clips (de Sousa et al., 2012), adults with TBI had less facial muscle activity, lower autonomic arousal, and less self-reported empathy than their uninjured peers. Emotional blunting has been linked to deficits in functions such as moral reasoning (Martins et al., 2012), and can have profound effects on family functioning (Worthington and Wood, 2018), so it merits further study. Not all adults with TBI show this reduction in intensity of feelings (Croker and McDonald, 2005; Amorapanth et al., 2018), however, and emotional blunting is only one component of the complex construct of apathy (Arnould et al., 2013; Quang et al., 2022), so future studies should consider individual differences and how detecting others’ feelings manifests in interpersonal interactions.

We hypothesized that because stimuli included complex visual scenes, participants in both groups would use fewer basic emotion words than social emotion or other words. This hypothesis was supported, as fewer than 50% of verbal responses in both groups were the traditional basic emotion words, and when happy responses were excluded, the number was less than 25%. The frequency of basic emotion words was higher than the previous study of young adults using this task (Turkstra et al., 2017), in which 28% of all responses were from the six basic emotion categories. One possible explanation, which we realized post-hoc, is that some participants had completed a force-choice emotion recognition task on the same day (i.e., their separate sessions were combined), and that task had the six basic emotion categories. Previous exposure to the six emotions could have biased participants’ responses toward the basic emotion categories. This bias likely explained some findings, such as several participants’ responses of “neutral,” a relatively low-frequency word for affect naming and one that is unlikely to be used by viewers naive to emotion recognition tasks. Effects of previous exposure also could explain why the percent of basic emotion responses was lower in the TBI group than the comparison group. Given their lower memory test scores, individuals with TBI might have been less likely than their peers to remember having seen the basic emotion words.

An unexpected finding was the number of times adults with TBI generated multiple words in response to an image. Although a median test showed no significant group difference, the number of multi-word answers was a gross measure and did not capture qualitative aspects of responses. Visual inspection of the data showed responses that could be classified as adding specificity or refining responses (e.g., “Waiting, attention directed,” “intense, focused”), changing the emotion (e.g., “Angry, afraid,” “Joking around, disgusted”), hesitating (e.g., “I do not know, hopeful”), and providing alternatives (e.g., “happy or satisfied”). It would be of interest to compare these qualitative features between groups. Based on evidence of impaired affect recognition in adults with TBI, participants in the TBI group might be more likely, for example, to give the wrong response first then self-correct. To avoid making inferences about participants’ intents (e.g., if they were self-correcting, refining answers, or disinhibited), we chose the first label they generated. Affect labeling must occur rapidly in everyday life, so the first response also was the most ecologically valid. Asking participants to explain their answers could provide further insights, with the caveat that this would be unnatural in everyday life. It also could be informative to look more closely at characteristics of participants who generated a large number of multi-word responses, and how these individual differences relate to other measures of social cognition, communication, and community outcome.

4.2. Gaze data

The second study hypothesis was that eye-tracking results would help explain any group differences between adults with and without TBI. That hypothesis was not supported. Participants in both groups looked more at the face, which is logical given that the task was to identify that person’s emotion, and there were no significant between-groups differences in total number or duration of fixations to the face vs. surrounding context. Two possible explanations for the discrepancy between verbal responses and gaze patterns are that (1) adults with TBI are looking at the same parts of the stimulus but perceiving or evaluating the information differently, or (2) adults with TBI are looking at different parts of the face, or fixating on parts of the stimulus in a different order than their uninjured peers (e.g., looking at the scene first rather than the face). The first explanation is consistent with the apathy studies cited earlier, and evidence of alexithymia after TBI (Fynn et al., 2021), and links between alexithymia and affect recognition (Neumann et al., 2014).

Results of a study by Greene et al. (2022) suggest the second explanation. These authors collected eye-tracking data while participants with TBI labeled static or dynamic images of facial expressions. They hypothesized that adults with TBI would have lower accuracy scores than their uninjured peers for both types of stimuli and would show “different eye scan patterns on static and dynamic tasks” (p. 3). Both the static and dynamic tasks had the typical basic emotion forced-choice options: anger, disgust, fear, surprise, happy, and sad (sic). Results replicated previous studies showing significant between-groups differences for recognizing anger, disgust, fear, and sadness; and no significant differences for happy, which as noted earlier is easiest for everyone, and surprise, which is most difficult for everyone (Jack et al., 2014). For both tasks, there was no main effect of group on fixation duration or number of fixations. The authors did find, however, differences in location of fixation on the static image task: participants with TBI looked first at the nose, which, as the authors stated, “is arguably the least informative part of the face” (p. 11), whereas controls looked first at the mouth. Greene et al. hypothesized that this finding was related to the lower accuracy scores in the TBI group but did not test that hypothesis directly. Fixation on regions within the face could be useful in future studies of context, as differences have been shown in populations with social similarities, such as adults with high-functioning autism (Setien-Ramos et al., 2022).

4.3. Limitations

A limitation of the study was that the use of news photographs was at the expense of experimental control. Photographs varied in image quality and in the number and nature of elements and people in each scene, so they are not directly comparable to each other or to other stimulus sets. As noted in the methods, for example, three of the FO images were sized differently than the others, and data for one image was discarded because the image was too pixelated to be labeled meaningfully. Future studies could use experimenter-generated images of scenes, to control factors such as exposure, focal length, and spatial frequency.

A second limitation of the study was that while the stimuli represented people of a wide variety of races and ethnicities, in a wide range of social, cultural, and economic contexts, the participants were almost all white European-heritage adults from the US Midwest. We employed several strategies to increase diversity of the sample but were unsuccessful. Literature on race and ethnicity effects has mostly focused on the race of the pictured person rather than the participant, other than the “in-group” effect (i.e., people are more accurate at identifying emotions on faces that look like their own race or ethnicity) (Elfenbein and Ambady, 2002; Dailey et al., 2010), and the TBI and comparison groups were matched, so it is not clear how the participants’ race would affect the study hypotheses. However, TBI occurs in all racial and ethnic communities, and studies that represent the TBI population are critical.

A third limitation is that for technical reasons, eye-tracking data were only available for a subset of participants. These participants were similar demographically to the full sample, and the TBI and comparison groups were well matched. Nevertheless, it is possible that we missed participants with TBI who had atypical gaze patterns.

5. Conclusion

In studies of affect recognition using de-contextualized stimuli and forced-choice responses, adults with TBI typically perform less accurately than their uninjured peers. It is not clear, however, if those studies capture the type of automatic, context-sensitive affect recognition that occurs in everyday life, vs. reflect the ability of adults with TBI to perform under specific task constraints. To better capture everyday affect recognition, we created stimuli with rich context cues and presented them in an open-response format. Results revealed qualitative differences in affect labels between adults with and without TBI, which did not appear to be attributable to differences in gaze patterns.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the University of Wisconsin-Madison Social and Behavioral Science Institutional Review Board, University of Iowa Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

LT was the senior researcher on this team, led much of the work cited, and led writing of the manuscript. SN created the original stimuli and co-developed the initial study concept with LT. BM and MD were co-investigators on the study and were responsible for developing the eye-tracking methods and oversaw implementation of all tasks and measures. SH-M updated the literature review, analyzed the gaze data, prepared the verbal response data for analysis, and contributed to manuscript preparation. SW analyzed the verbal response data and contributed to the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the NIH NICHD/NCMRR (award no. R01 HD071089) and NIH/NIGMS (award no. R25GMO83252).

Acknowledgments

We wish to thank the many students who administered study tasks to participants and contributed to data analysis. We also wish to thank Erica Richmond for leading participant recruitment and coordinating the study across sites; and Kristina Visscher for her contribution to the pilot gaze analysis that preceded this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abdallah, S. (2020). Social interactions in electronic networks of practice: a model for effective knowledge construction, problem solving, and learning. Int. J. E-Collab. 16, 27–44. doi: 10.4018/IJeC.2020040103

CrossRef Full Text | Google Scholar

Adolphs, R. (2002). Neural systems for recognizing emotion. Curr. Opin. Neurobiol. 12, 169–177. doi: 10.1016/S0959-4388(02)00301-X

CrossRef Full Text | Google Scholar

Amorapanth, P. X., Aluru, V., Stone, J., Yousefi, A., Tang, A., Cox, S., et al. (2018). Traumatic brain injury results in altered physiologic, but not subjective responses to emotional stimuli. Brain Inj. 32, 1712–1719. doi: 10.1080/02699052.2018.1519598

CrossRef Full Text | Google Scholar

Arnould, A., Rochat, L., Azouvi, P., and Van der Linden, M. (2013). A multidimensional approach to apathy after traumatic brain injury. Neuropsychol. Rev. 23, 210–233. doi: 10.1007/s11065-013-9236-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Aviezer, H., Ensenberg, N., and Hassin, R. R. (2017). The inherently contextualized nature of facial emotion perception. Curr. Opin. Psychol. 17, 47–54. doi: 10.1016/j.copsyc.2017.06.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Aviezer, H., Trope, Y., and Todorov, A. (2012). Holistic person processing: faces with bodies tell the whole story. J. Pers. Soc. Psychol. 103, 20–37. doi: 10.1037/a0027411

PubMed Abstract | CrossRef Full Text | Google Scholar

Babbage, D. R., Yim, J., Zupan, B., Neumann, D., Tomita, M. R., and Willer, B. (2011). Meta-analysis of facial affect recognition difficulties after traumatic brain injury. Neuropsychology 25, 277–285. doi: 10.1037/a0021908

PubMed Abstract | CrossRef Full Text | Google Scholar

Baron-Cohen, S., Wheelwright, S., Hill, J., Raste, Y., and Plumb, I. (2001a). The “Reading the mind in the eyes” test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism. J. Child Psychol. Psychiatry 42, 241–251. doi: 10.1111/1469-7610.00715

CrossRef Full Text | Google Scholar

Baron-Cohen, S., Wheelwright, S., Scahill, V., Lawson, J., and Spong, A. (2001b). Are intuitive physics and intuitive psychology independent? A test with children with Asperger syndrome. J. Dev. Learn. Disord. 5, 47–78.

Google Scholar

Barrett, L. F., Mesquita, B., and Gendron, M. (2011). Context in emotion perception. Curr. Dir. Psychol. Sci. 20, 286–290. doi: 10.1177/0963721411422522

CrossRef Full Text | Google Scholar

Bates, D., Mächler, M., Bolker, B., and Walker, S. (2014). Fitting linear mixed-effects models using lme4. arXiv preprint arXiv : 1406.5823. doi: 10.18637/jss.v067.i01

CrossRef Full Text | Google Scholar

Benedictus, M. R., Spikman, J. M., and van der Naalt, J. (2010). Cognitive and behavioral impairment in traumatic brain injury related to outcome and return to work. Arch. Phys. Med. Rehabil. 91, 1436–1441. doi: 10.1016/j.apmr.2010.06.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, A. S., Lancaster, K., Lengenfelder, J., Chiaravalloti, N. D., and Genova, H. M. (2019). Community integration in traumatic brain injury: the contributing factor of affect recognition deficits. J. Int. Neuropsychol. Soc. 25, 890–895. doi: 10.1017/S1355617719000559

CrossRef Full Text | Google Scholar

Bird, S., Klein, E., and Loper, E. (2009). Natural language processing with Python: analyzing text with the natural language toolkit. California: O’Reilly Media, Inc.

Google Scholar

Blennow, K., Brody, D. L., Kochanek, P. M., Levin, H., McKee, A., Ribbers, G. M., et al. (2016). Traumatic brain injuries. Nat. Rev. Dis. Prim. 2:16084. doi: 10.1038/nrdp.2016.84

CrossRef Full Text | Google Scholar

Bonatti, E., Zamarian, L., Wagner, M., Benke, T., Hollosi, P., Strubreither, W., et al. (2008). Making decisions and advising decisions in traumatic brain injury. Cogn. Behav. Neurol. 21, 164–175. doi: 10.1097/WNN.0b013e318184e688

PubMed Abstract | CrossRef Full Text | Google Scholar

Bora, E., Eryavuz, A., Kayahan, B., Sungu, G., and Veznedaroglu, B. (2006). Social functioning, theory of mind and neurocognition in outpatients with schizophrenia; mental state decoding may be a better predictor of social functioning than mental state reasoning. Psychiatry Res. 145, 95–103. doi: 10.1016/j.psychres.2005.11.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Bretherton, I., and Beeghly, M. (1982). Talking about internal states: the acquisition of an explicit theory of mind. Dev. Psychol. 18, 906–921. doi: 10.1037/0012-1649.18.6.906

CrossRef Full Text | Google Scholar

Broicher, S. D., Kuchukhidze, G., Grunwald, T., Kramer, G., Kurthen, M., and Jokeit, H. (2012). “Tell me how do I feel”--emotion recognition and theory of mind in symptomatic mesial temporal lobe epilepsy. Neuropsychologia 50, 118–128. doi: 10.1016/j.neuropsychologia.2011.11.005

CrossRef Full Text | Google Scholar

Buck, R. (1988). Human motivation and emotion. New York, NY: Wiley.

Google Scholar

Byom, L., Duff, M., Mutlu, B., and Turkstra, L. (2019). Facial emotion recognition of older adults with traumatic brain injury. Brain Inj. 33, 322–332. doi: 10.1080/02699052.2018.1553066

PubMed Abstract | CrossRef Full Text | Google Scholar

Carroll, J. M., and Russell, J. A. (1996). Do facial expressions signal specific emotions? Judging emotion from the face in context. J. Pers. Soc. Psychol. 70, 205–218. doi: 10.1037//0022-3514.70.2.205

PubMed Abstract | CrossRef Full Text | Google Scholar

Cassel, A., McDonald, S., Kelly, M., and Togher, L. (2019). Learning from the minds of others: a review of social cognition treatments and their relevance to traumatic brain injury. Neuropsychol. Rehabil. 29, 22–55. doi: 10.1080/09602011.2016.1257435

PubMed Abstract | CrossRef Full Text | Google Scholar

Cassels, T. G., and Birch, S. A. (2014). Comparisons of an open-ended vs. forced-choice ‘mind reading’ task: implications for measuring perspective-taking and emotion recognition. PLoS One 9:e93653. doi: 10.1371/journal.pone.0093653

PubMed Abstract | CrossRef Full Text | Google Scholar

Charvet, L. E., Cleary, R. E., Vazquez, K., Belman, A. L., and Krupp, L. B., US Network for Pediatric MS (2014). Social cognition in pediatric-onset multiple sclerosis (MS). Mult. Scler. 20, 1478–1484. doi: 10.1177/1352458514526942

PubMed Abstract | CrossRef Full Text | Google Scholar

Croker, V., and McDonald, S. (2005). Recognition of emotion from facial expression following traumatic brain injury. Brain Inj. 19, 787–799. doi: 10.1080/02699050500110033

CrossRef Full Text | Google Scholar

Dailey, M. N., Joyce, C., Lyons, M. J., Kamachi, M., Ishi, H., Gyoba, J., et al. (2010). Evidence and a computational explanation of cultural differences in facial expression recognition. Emotion 10, 874–893. doi: 10.1037/a0020019

CrossRef Full Text | Google Scholar

de Gelder, B., Meeren, H. K., Righart, R., van den Stock, J., van de Riet, W. A., and Tamietto, M. (2006). Beyond the face: exploring rapid influences of context on face processing. Prog. Brain Res. 155, 37–48. doi: 10.1016/S0079-6123(06)55003-4, S0079-6123(06)55003-4 [pii]

CrossRef Full Text | Google Scholar

de Sousa, A., McDonald, S., and Rushby, J. (2012). Changes in emotional empathy, affective responsivity, and behavior following severe traumatic brain injury. J. Clin. Exp. Neuropsychol. 34, 606–623. doi: 10.1080/13803395.2012.667067

PubMed Abstract | CrossRef Full Text | Google Scholar

Delis, D. C., Kramer, J. H., Kaplan, E., and Ober, B. A. (2000). California verbal learning test-adult version (CVLT-II). Austin, TX: The Psychological Corporation.

Google Scholar

Dewan, M. C., Rattani, A., Gupta, S., Baticulon, R. E., Hung, Y. C., Punchak, M., et al. (2018). Estimating the global incidence of traumatic brain injury. J. Neurosurg. 130, 1080–1097. doi: 10.3171/2017.10.JNS17352

CrossRef Full Text | Google Scholar

Eddy, C. M., Sira Mahalingappa, S., and Rickards, H. E. (2012). Is Huntington’s disease associated with deficits in theory of mind? Acta Neurol. Scand. 126, 376–383. doi: 10.1111/j.1600-0404.2012.01659.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ekman, P. (1992). Are there basic emotions? Psychol. Rev. 99, 550–553. doi: 10.1037/0033-295X.99.3.550

CrossRef Full Text | Google Scholar

Ekman, P., and Friesen, W. V. (1967). Head and body cues in the judgment of emotion: a reformulation. Percept. Mot. Skills 24, 711–724. doi: 10.2466/pms.1967.24.3.711

PubMed Abstract | CrossRef Full Text | Google Scholar

Ekman, P., Friesen, W. V., and Ellsworth, P. (1972). Emotion in the human face: guidelines for research and an integration of findings. Oxford: Pergamon Press.

Google Scholar

Ekman, P., Sorenson, E. R., and Friesen, W. V. (1969). Pan-cultural elements in facial displays of emotion. Science 164, 86–88. doi: 10.1126/science.164.3875.86

PubMed Abstract | CrossRef Full Text | Google Scholar

Elfenbein, H. A., and Ambady, N. (2002). On the universality and cultural specificity of emotion recognition: a meta-analysis. Psychol. Bull. 128, 203–235. doi: 10.1037/0033-2909.128.2.203

PubMed Abstract | CrossRef Full Text | Google Scholar

Fernández-Dols, J.-M., and Crivelli, C. (2013). Emotion and expression: naturalistic studies. Emot. Rev. 5, 24–29. doi: 10.1177/1754073912457229

CrossRef Full Text | Google Scholar

Fernández-Dols, J.-M., and Crivelli, C. (2015). “Recognition of facial expressions: past, present, and future challenges” in Understanding facial expressions in communication: cross-cultural and multidisciplinary perspectives. eds. M. Mandal and A. Awasthi (Washington, DC: Springer), 19–40.

Google Scholar

Fynn, D. M., Gignac, G. E., Becerra, R., Pestell, C. F., and Weinborn, M. (2021). The prevalence and characteristics of alexithymia in adults following brain injury: a meta-analysis. Neuropsychol. Rev. 31, 722–738. doi: 10.1007/s11065-021-09484-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Genova, H. M., Genualdi, A., Goverover, Y., Chiaravalloti, N. D., Marino, C., and Lengenfelder, J. (2017). An investigation of the impact of facial affect recognition impairments in moderate to severe TBI on fatigue, depression, and quality of life. Soc. Neurosci. 12, 303–307. doi: 10.1080/17470919.2016.1173584

PubMed Abstract | CrossRef Full Text | Google Scholar

Geraets, C. N. W., Klein Tuente, S., Lestestuiver, B. P., van Beilen, M., Nijman, S. A., Marsman, J. B. C., et al. (2021). Virtual reality facial emotion recognition in social environments: an eye-tracking study. Internet Interv. 25:100432. doi: 10.1016/j.invent.2021.100432

PubMed Abstract | CrossRef Full Text | Google Scholar

Greene, L., Barker, L. A., Reidy, J., Morton, N., and Atherton, A. (2022). Emotion recognition and eye tracking of static and dynamic facial affect: a comparison of individuals with and without traumatic brain injury. J. Clin. Exp. Neuropsychol. 44, 461–477. doi: 10.1080/13803395.2022.2128066

CrossRef Full Text | Google Scholar

Havet-Thomassin, V., Allain, P., Etcharry-Bouyx, F., and Le Gall, D. (2006). What about theory of mind after severe brain injury? Brain Inj. 20, 83–91. doi: 10.1080/02699050500340655

CrossRef Full Text | Google Scholar

Hayes, G. S., McLennan, S. N., Henry, J. D., Phillips, L. H., Terrett, G., Rendell, P. G., et al. (2020). Task characteristics influence facial emotion recognition age-effects: a meta-analytic review. Psychol. Aging 35, 295–315. doi: 10.1037/pag0000441

PubMed Abstract | CrossRef Full Text | Google Scholar

Heller, J., Dogan, I., Schulz, J. B., and Reetz, K. (2014). Evidence for gender differences in cognition, emotion and quality of life in Parkinson’s disease? Aging Dis. 5, 63–75. doi: 10.14366/AD.2014.050063

PubMed Abstract | CrossRef Full Text | Google Scholar

Henry, J. D., Phillips, L. H., Beatty, W. W., McDonald, S., Longley, W. A., Joscelyne, A., et al. (2009). Evidence for deficits in facial affect recognition and theory of mind in multiple sclerosis. J. Int. Neuropsychol. Soc. 15, 277–285. doi: 10.1017/S1355617709090195

PubMed Abstract | CrossRef Full Text | Google Scholar

Hutto, C., and Gilbert, E. (2014). “Vader: a parsimonious rule-based model for sentiment analysis of social media text” in Proceedings of the International AAAI Conference on Web and Social Media, 216–225.

Google Scholar

Jack, R. E., Garrod, O. G., and Schyns, P. G. (2014). Dynamic facial expressions of emotion transmit an evolving hierarchy of signals over time. Curr. Biol. 24, 187–192. doi: 10.1016/j.cub.2013.11.064

PubMed Abstract | CrossRef Full Text | Google Scholar

Johnson, J. E., and Turkstra, L. S. (2012). Inference in conversation of adults with traumatic brain injury. Brain Inj. 26, 1118–1126. doi: 10.3109/02699052.2012.666370

CrossRef Full Text | Google Scholar

Kayyal, M., Widen, S., and Russell, J. A. (2015). Context is more powerful than we think: contextual cues override facial cues even for valence. Emotion 15, 287–291. doi: 10.1037/emo0000032

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, M., McDonald, S., and Kellett, D. (2013). The psychological effects of ostracism following traumatic brain injury. Brain Inj. 27, 1676–1684. doi: 10.3109/02699052.2013.834381

PubMed Abstract | CrossRef Full Text | Google Scholar

Kertesz, A. (2006). Western aphasia battery. San Antonio, TX: Pearson Assessment.

Google Scholar

Kipps, C. M., Duggins, A. J., McCusker, E. A., and Calder, A. J. (2007). Disgust and happiness recognition correlate with anteroventral insula and amygdala volume respectively in preclinical Huntington’s disease. J. Cogn. Neurosci. 19, 1206–1217. doi: 10.1162/jocn.2007.19.7.1206

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumfor, F., and Piguet, O. (2012). Disturbance of emotion processing in frontotemporal dementia: a synthesis of cognitive and neuroimaging findings. Neuropsychol. Rev. 22, 280–297. doi: 10.1007/s11065-012-9201-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Lane-Brown, A. T., and Tate, R. L. (2011). Apathy after traumatic brain injury: an overview of the current state of play. Brain Impair 12, 43–53. doi: 10.1375/brim.12.1.43

CrossRef Full Text | Google Scholar

Lott, L. L., Spengler, F. B., Stachele, T., Schiller, B., and Heinrichs, M. (2022). EmBody/EmFace as a new open tool to assess emotion recognition from body and face expressions. Sci. Rep. 12:14165. doi: 10.1038/s41598-022-17866-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Malec, J. F., Brown, A. W., Leibson, C. L., Flaada, J. T., Mandrekar, J. N., Diehl, N. N., et al. (2007). The mayo classification system for traumatic brain injury severity. J. Neurotrauma 24, 1417–1424. doi: 10.1089/neu.2006.0245

CrossRef Full Text | Google Scholar

Martins, A., Faisca, L., Esteves, F., Muresan, A., and Reis, A. (2012). Atypical moral judgements following traumatic brain injury. Judgm. Decis. Mak. 7, 478–487. doi: 10.1017/S1930297500002813

CrossRef Full Text | Google Scholar

Masel, B. E., and DeWitt, D. S. (2010). Traumatic brain injury: a disease process, not an event. J. Neurotrauma 27, 1529–1540. doi: 10.1089/neu.2010.1358

CrossRef Full Text | Google Scholar

McLean, A. M., Jarus, T., Hubley, A. M., and Jongbloed, L. (2014). Associations between social participation and subjective quality of life for adults with moderate to severe traumatic brain injury. Disabil. Rehabil. 36, 1409–1418. doi: 10.3109/09638288.2013.834986

PubMed Abstract | CrossRef Full Text | Google Scholar

Milders, M. (2018). Relationship between social cognition and social behaviour following traumatic brain injury. Brain Inj. 33, 62–68. doi: 10.1080/02699052.2018.1531301

PubMed Abstract | CrossRef Full Text | Google Scholar

Mu, Y. G., Huang, L. J., Li, S. Y., Ke, C., Chen, Y., Jin, Y., et al. (2012). Working memory and the identification of facial expression in patients with left frontal glioma. Neurooncology 14:iv81. doi: 10.1093/neuonc/nos215

PubMed Abstract | CrossRef Full Text | Google Scholar

Muller, F., Simion, A., Reviriego, E., Galera, C., Mazaux, J. M., Barat, M., et al. (2009). Exploring theory of mind after severe traumatic brain injury. Cortex 46, 1088–1099. doi: 10.1016/j.cortex.2009.08.014

CrossRef Full Text | Google Scholar

Mumenthaler, C., and Sander, D. (2012). Social appraisal influences recognition of emotions. J. Pers. Soc. Psychol. 102, 1118–1135. doi: 10.1037/a0026885

PubMed Abstract | CrossRef Full Text | Google Scholar

Murphy, J. M., Bennett, J. M., de la Piedad Garcia, X., and Willis, M. L. (2021). Emotion recognition and traumatic brain injury: a systematic review and meta-analysis. Neuropsychol. Rev. 32, 520–536. doi: 10.1007/s11065-021-09510-7

CrossRef Full Text | Google Scholar

Nelson, N. L., and Mondloch, C. J. (2019). Children’s perception of emotions in the context of live interactions: eye movements and emotion judgements. Behav. Process. 164, 193–200. doi: 10.1016/j.beproc.2019.05.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Neumann, D., and Zupan, B. (2019). Empathic responses to affective film clips following brain injury and the association with emotion recognition accuracy. Arch. Phys. Med. Rehabil. 100, 458–463. doi: 10.1016/j.apmr.2018.07.431

PubMed Abstract | CrossRef Full Text | Google Scholar

Neumann, D., Zupan, B., Malec, J. F., and Hammond, F. (2014). Relationships between alexithymia, affect recognition, and empathy after traumatic brain injury. J. Head Trauma Rehabil. 29, E18–E27. doi: 10.1097/HTR.0b013e31827fb0b5

PubMed Abstract | CrossRef Full Text | Google Scholar

Pabst, A., Gautier, M., and Maurage, P. (2022). Tasks and investigated components in social cognition research among adults with alcohol use disorder: a critical scoping review. Psychol. Addict. Behav. 36, 999–1011. doi: 10.1037/adb0000874

PubMed Abstract | CrossRef Full Text | Google Scholar

Quang, H., McDonald, S., Huynh-Le, P., Nguyen, T.-V., Le, N.-A., Lam-Nguyen, N.-T., et al. (2022). Apathy in a high prevalence population of moderate to severe traumatic brain injury: an investigation in Vietnam. Neuropsychology 36, 94–102. doi: 10.1037/neu0000781

PubMed Abstract | CrossRef Full Text | Google Scholar

Righart, R., and de Gelder, B. (2008a). Rapid influence of emotional scenes on encoding of facial expressions: an ERP study. Soc. Cogn. Affect. Neurosci. 3, 270–278. doi: 10.1093/scan/nsn021

PubMed Abstract | CrossRef Full Text | Google Scholar

Righart, R., and de Gelder, B. (2008b). Recognition of facial expressions is influenced by emotional scene gist. Cogn. Affect. Behav. Neurosci. 8, 264–272. doi: 10.3758/CABN.8.3.264

CrossRef Full Text | Google Scholar

Rigon, A., Turkstra, L. S., Mutlu, B., and Duff, M. C. (2018). Facial-affect recognition deficit as a predictor of different aspects of social-communication impairment in traumatic brain injury. Neuropsychology 32, 476–483. doi: 10.1037/neu0000368

PubMed Abstract | CrossRef Full Text | Google Scholar

Rodriguez-Rajo, P., Garcia-Rudolph, A., Sanchez-Carrion, R., Aparicio-Lopez, C., Ensenat-Cantallops, A., and Garcia-Molina, A. (2022). Social and nonsocial cognition: are they linked? A study on patients with moderate-to-severe traumatic brain injury. Appl. Neuropsychol. Adult 29, 1039–1048. doi: 10.1080/23279095.2020.1845171

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenberg, H., McDonald, S., Dethier, M., Kessels, R. P., and Westbrook, R. F. (2014). Facial emotion recognition deficits following moderate-severe traumatic brain injury (TBI): re-examining the valence effect and the role of emotion intensity. J. Int. Neuropsychol. Soc. 20, 994–1003. doi: 10.1017/s1355617714000940

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenberg, H., McDonald, S., Rosenberg, J., and Frederick Westbrook, R. (2018). Amused, flirting or simply baffled? Is recognition of all emotions affected by traumatic brain injury? J. Neuropsychol. 12, 145–164. doi: 10.1111/jnp.12109

PubMed Abstract | CrossRef Full Text | Google Scholar

Schmidt, A. T., Hanten, G. R., Li, X., Orsten, K. D., and Levin, H. S. (2010). Emotion recognition following pediatric traumatic brain injury: longitudinal analysis of emotional prosody and facial emotion recognition. Neuropsychologia 48, 2869–2877. doi: 10.1016/j.neuropsychologia.2010.05.029

PubMed Abstract | CrossRef Full Text | Google Scholar

Schwarz, K. A., Wieser, M. J., Gerdes, A. B., Muhlberger, A., and Pauli, P. (2013). Why are you looking like that? How the context influences evaluation and processing of human faces. Soc. Cogn. Affect. Neurosci. 8, 438–445. doi: 10.1093/scan/nss013

PubMed Abstract | CrossRef Full Text | Google Scholar

Setien-Ramos, I., Lugo-Marín, J., Gisbert-Gustemps, L., Díez-Villoria, E., Magán-Maganto, M., Canal-Bedia, R., et al. (2022). Eye-tracking studies in adults with autism spectrum disorder: a systematic review and meta-analysis. J. Autism Dev. Disord. 53, 2430–2443. doi: 10.1007/s10803-022-05524-z

CrossRef Full Text | Google Scholar

Sherer, M., Clark, A. N., Sander, A. M., Struchen, M. A., Bogaards, J., Leon-Novelo, L., et al. (2022). Relationships of self-awareness and facial affect recognition to social communication ability in persons with traumatic brain injury. Neuropsychol. Rehabil. 32, 2013–2028. doi: 10.1080/09602011.2022.2084118

PubMed Abstract | CrossRef Full Text | Google Scholar

SMI Software (2012). SensoMotoric Instruments GmBH. Teltow.

Google Scholar

Spikman, J. M., Boelen, D. H., Pijnenborg, G. H., Timmerman, M. E., van der Naalt, J., and Fasotti, L. (2013). Who benefits from treatment for executive dysfunction after brain injury? Negative effects of emotion recognition deficits. Neuropsychol. Rehabil. 23, 824–845. doi: 10.1080/09602011.2013.826138

PubMed Abstract | CrossRef Full Text | Google Scholar

Stocchetti, N., and Zanier, E. R. (2016). Chronic impact of traumatic brain injury on outcome and quality of life: a narrative review. Crit. Care 20:148. doi: 10.1186/s13054-016-1318-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Tamietto, M., Adenzato, M., Geminiani, G., and de Gelder, B. (2007). Fast recognition of social emotions takes the whole brain: interhemispheric cooperation in the absence of cerebral asymmetry. Neuropsychologia 45, 836–843. doi: 10.1016/j.neuropsychologia.2006.08.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Teasdale, G., and Jennett, B. (1974). Assessment of coma and impaired consciousness: a practical scale. Lancet 2, 81–84.

PubMed Abstract | Google Scholar

Theadom, A., McDonald, S., Starkey, N., Barker-Collo, S., Jones, K. M., Ameratunga, S., et al. (2019). Social cognition four years after mild-TBI: an age-matched prospective longitudinal cohort study. Neuropsychology 33, 560–567. doi: 10.1037/neu0000516

PubMed Abstract | CrossRef Full Text | Google Scholar

Togher, L., Douglas, J., Turkstra, L. S., Welch-West, P., Janzen, S., Harnett, A., et al. (2023). INCOG 2.0 guidelines for cognitive rehabilitation following traumatic brain injury, part IV: cognitive-communication and social cognition disorders. J. Head Trauma Rehabil. 38, 65–82. doi: 10.1097/HTR.0000000000000835

CrossRef Full Text | Google Scholar

Tombaugh, T. N. (2004). Trail making test a and B: normative data stratified by age and education. Arch. Clin. Neuropsychol. 19, 203–214. doi: 10.1016/S0887-6177(03)00039-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Turkstra, L. S. (2008). Conversation-based assessment of social cognition in adults with traumatic brain injury. Brain Inj. 22, 397–409. doi: 10.1080/02699050802027059

CrossRef Full Text | Google Scholar

Turkstra, L. S., Kraning, S. G., Riedeman, S. K., Mutlu, B., Duff, M., and Van Den Heuvel, S. (2017). Labelling facial affect in context in adults with and without TBI. Brain Impair 18, 49–61. doi: 10.1017/BrImp.2016.29

PubMed Abstract | CrossRef Full Text | Google Scholar

Turkstra, L. S., Mutlu, B., Ryan, C. W., Despins Stafslien, E. H., Richmond, E. K., Hosokawa, E., et al. (2020). Sex and gender differences in emotion recognition and theory of mind after TBI: a narrative review and directions for future research. Front. Neurol. 11:59. doi: 10.3389/fneur.2020.00059

CrossRef Full Text | Google Scholar

Turkstra, L. S., Norman, R. S., Mutlu, B., and Duff, M. C. (2018). Impaired theory of mind in adults with traumatic brain injury: a replication and extension of findings. Neuropsychologia 111, 117–122. doi: 10.1016/j.neuropsychologia.2018.01.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Ubukata, S., Tanemura, R., Yoshizumi, M., Sugihara, G., Murai, T., and Ueda, K. (2014). Social cognition and its relationship to functional outcomes in patients with sustained acquired brain injury. Neuropsychiatr. Dis. Treat. 10, 2061–2068. doi: 10.2147/NDT.S68156

PubMed Abstract | CrossRef Full Text | Google Scholar

Vallat-Azouvi, C., Azouvi, P., Le-Bornec, G., and Brunet-Gouet, E. (2019). Treatment of social cognition impairments in patients with traumatic brain injury: a critical review. Brain Inj. 33, 87–93. doi: 10.1080/02699052.2018.1531309

PubMed Abstract | CrossRef Full Text | Google Scholar

van Kleef, G. A., Cheshin, A., Fischer, A. H., and Schneider, I. K. (2016). Editorial: the social nature of emotions. Front. Psychol. 7:896. doi: 10.3389/fpsyg.2016.00896

PubMed Abstract | CrossRef Full Text | Google Scholar

Wechsler, D. (2008). Wechsler adult intelligence scale. San Antonio, TX: Pearson.

Google Scholar

Wieser, M. J., and Brosch, T. (2012). Faces in context: a review and systematization of contextual influences on affective face processing. Front. Psychol. 3:471. doi: 10.3389/fpsyg.2012.00471

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilde, E. A., Whiteneck, G. G., Bogner, J., Bushnik, T., Cifu, D. X., Dikmen, S., et al. (2010). Recommendations for the use of common outcome measures in traumatic brain injury research. Arch. Phys. Med. Rehabil. 91, 1650–1660.e17. doi: 10.1016/j.apmr.2010.06.033

CrossRef Full Text | Google Scholar

Winters, A. (2005). Perceptions of body posture and emotion: a question of methodology. New Sch. Psychol. Bull. 3, 35–45. doi: 10.1037/e741582011-003

CrossRef Full Text | Google Scholar

Woodworth, R. S. (1938). Experimental psychology. New York, NY: Henry Holt.

Google Scholar

Worthington, A., and Wood, R. L. (2018). Apathy following traumatic brain injury: a review. Neuropsychologia 118, 40–47. doi: 10.1016/j.neuropsychologia.2018.04.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Yuvaraj, R., Murugappan, M., Norlinah, M. I., Sundaraj, K., and Khairiyah, M. (2013). Review of emotion recognition in stroke patients. Dement. Geriatr. Cogn. Disord. 36, 179–196. doi: 10.1159/000353440

PubMed Abstract | CrossRef Full Text | Google Scholar

Zupan, B., Dempsey, L., and Hartwell, K. (2022). Categorising emotion words: the influence of response options. Lang. Cogn. 15, 29–52. doi: 10.1017/langcog.2022.24

CrossRef Full Text | Google Scholar

Zupan, B., and Neumann, D. (2014). Affect recognition in traumatic brain injury: responses to unimodal and multimodal media. J. Head Trauma Rehabil. 29, E1–E12. doi: 10.1097/HTR.0b013e31829dded6

CrossRef Full Text | Google Scholar

Keywords: traumatic brain injury, speech-language pathology, emotion recognition, affect recognition, social cognition, facial expression, eye-tracking, communication

Citation: Turkstra LS, Hosseini-Moghaddam S, Wohltjen S, Nurre SV, Mutlu B and Duff MC (2023) Facial affect recognition in context in adults with and without TBI. Front. Psychol. 14:1111686. doi: 10.3389/fpsyg.2023.1111686

Received: 29 November 2022; Accepted: 31 July 2023;
Published: 14 August 2023.

Edited by:

Carolina Pena Martinez, University of South Florida, United States

Reviewed by:

Raul Espert-Tortajada, University of Valencia, Spain
Álvaro Ruiz-García, Research Institute Hospital 12 de Octubre, Spain

Copyright © 2023 Turkstra, Hosseini-Moghaddam, Wohltjen, Nurre, Mutlu and Duff. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Lyn S. Turkstra, turkstrl@mcmaster.ca

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.