Introduction

In his seminal report of 11 cases, Kanner (1943) described children with an ‘autistic disorder of affective contact’. Indeed, modern diagnostic criteria and the tools commonly used to make autism spectrum disorder (ASD) diagnoses include items related to difficulties identifying and processing emotional information (APA 1994; Lord et al. 2000). In clinical and laboratory settings, perhaps the most common way to assess emotional identification and perception is through facial emotion recognition (FER) paradigms. Many studies have examined FER in ASD, with varying results (for review, see Harms et al. 2010). In the context of intellectual disability, children with ASD have shown deficits in FER (e.g., Celani et al. 1999); however, low IQ impacts FER, complicating interpretation of findings and identification of ASD-specific deficits. More recent studies have included predominantly high functioning individuals on the autism spectrum in order to isolate ASD-specific deficits in FER. Several of these studies have shown general ASD-related difficulties with labeling full facial expressions (Bormann-Kischkel et al. 1995; Buitelaar et al. 1999; Hobson 1986; Gross 2004). However, some studies have found no deficits in identifying full emotional expressions (Adolphs et al. 2001; Capps et al. 1992; Castelli 2005; Jones et al. 2010; Prior et al. 1990; Robel et al. 2004), while others have found FER deficits restricted to particular emotions, most often emotions with a negative valence, such as fear (Howard et al. 2000; Pelphrey et al. 2002), disgust, sadness, and anger (Ashwin et al. 2006). Mixed findings may occur in part because most images used in previous studies depict fully expressed emotions with diminished complexity and subtlety unlike the constantly changing expressions encountered in real-world interactions and because many interventions tailored to ASD currently incorporate training of facial emotion identification often using images of fully expressive faces. Therefore, the relative difficulty individuals with ASD may encounter, or different strategies they may use, with real-life FER could be masked in studies that use only these stereotyped images.

Indeed, faces change from neutral to expressive and from one expression to another in everyday contexts. Such movement has been shown to facilitate perception of faces and facial expressions in neurotypical adults (Berry 1990) and individuals with intellectual impairments (Harwood et al. 1999). Gepner et al. (2001) found that dynamic (i.e., video and stroboscopic) presentations of facial emotion also facilitated emotional expression recognition among individuals with ASD. These clips, however, lasted only 2 s each. To examine fine grain perceptual sensitivity to facial emotional expression (i.e., the intensity of expression a previously neutral face needs to exhibit for successful recognition), greater exposure to the stimuli is needed.

Only a few studies to date have utilized “morphing faces” stimuli (in which intensity of emotional expressions is varied) to assess not only accurate identification of, but also perceptual sensitivity to, facial expressions of emotion among individuals with ASD. Of those that have, most used blends or morphs between emotions, rather than beginning with a neutral expression and progressing to full emotional expression. For example, in one study, Humphreys et al. (2007) found ASD-related difficulties with fear recognition from facial expression blends (e.g., fear-surprise). In another study using emotion blends, Teunisse and de Gelder (2001) found atypical responses to anger-sadness and anger-fear blends but intact responses to happy-sad blends among individuals with ASD. In contrast, Homer and Rutherford (2008) used numerous emotion blends and documented primarily intact performance among individuals with ASD. Only one study has examined FER perceptual sensitivity based on fine grain progressions from neutral to full expression (using the emotional multimorph task) among individuals with ASD. Participants were categorized into two groups of either high or low callous-unemotional trait ratings (associated with a form of antisocial behavior) (Rogers et al. 2006). Using planned comparisons, only sadness required more intense facial expressions to be accurately recognized by individuals with both ASD and high callous-unemotional trait ratings than those with ASD and low ratings. However, a typically developing (TD) control group was not included in this study, which means that the ASD-related (a)typicality of performance on this task (whether responding to faces depicting sadness or any of the other five basic emotions) relative to TD peers of the same age and functioning level was not documented. Further work is needed to examine whether perceptual sensitivity to not only sad facial expressions, but also those depicting the other five basic emotions, is intact or impaired among individuals with ASD.

Identifying the functional impact of FER difficulties provides external validity for deficits documented in the laboratory setting; nevertheless, assessing correlations between everyday behavior and lab FER performance has only just begun to be explored in studies of ASD. For example, Humphreys et al. (2007) found that difficulties identifying fearful faces among individuals with ASD were associated with communication symptoms rated during a standardized interaction with a clinician (i.e., the Autism Diagnostic Observation Schedule [ADOS]). Similarly, Boraston et al. (2007) found a significant correlation between sadness recognition (in the context of abstract Heider and Simmel-type animations (Heider and Simmel 1944) designed to elicit emotion-laden descriptions, though without explicit social perceptual cues) and degree of impairment on the ADOS reciprocal social interaction score. Again, the ability to identify fear and sadness in particular, correlates with ASD symptomatology; however, these studies have been limited in utilizing only symptom counts from the standard diagnostic measures as real-life behavioral correlates. In addition to symptomatology, adaptive functioning, including social, communication, and daily living skills, provides a targeted correlate for validating the real-life impact of FER difficulties, particularly in ASD, in which adaptive behavior difficulties are common even in the highest functioning individuals (Liss et al. 2001; Kenworthy et al. 2010).

The current study extends the investigation of perceptual sensitivity to facial expression in ASD by using the emotional multimorph task in which faces take 20 steps to morph from neutral to full expression to assess fine-tuned differences in perceptual sensitivity between high functioning adolescents with ASD (to avoid confounds associated with intellectual impairments) and TD adolescents. Moreover, correlational analyses with not only autism-related behavior ratings (including symptomatology) but also adaptive functioning will clarify relationships between FER in the laboratory and everyday behaviors observed by caregivers and clinicians. It is predicted that (1) TD adolescents will provide more accurate identification of and will be more perceptually sensitive to facial expressions of fear and sadness than will adolescents with ASD and (2) sensitivity to these two negative emotions will correlate with both autism-related behavior ratings and adaptive functioning as rated by clinicians and parents in the ASD group.

Method

Participants

Thirty-one TD adolescents (three female) and 42 high functioning adolescents (four female) with an ASD (11 with high functioning autism, 27 with Asperger’s syndrome, three with pervasive developmental disorder-not otherwise specified, and one with an ASD but exact diagnosis unknown because of sparse developmental data) completed the emotional multimorph task. ASD diagnoses were given by experienced clinicians according to DSM-IV criteria. Additionally, all participants met the Autism Diagnostic Interview-Revised (Lord et al. 1994) cutoff for autism in the social domain and at least one other domain and/or met the ADOS-Generic (Lord et al. 2000) ASD cutoff for combined social and communication symptoms (Lainhart et al. 2006). Exclusion criteria for the ASD group included any known co-morbid medical conditions, genetic disorder (e.g., fragile X syndrome), or head injury/neurological insult which may affect cognitive functioning. TD participants were recruited from the community and parents of all TD participants underwent telephone screenings. TD participants were excluded from participation if they had ever received mental health treatment for anxiety, depression, or any other psychiatric condition, taken psychiatric medications, required special services in school, or had trauma/injury that could potentially affect cognitive functioning. Additionally, all participants scored 80 or higher on a Wechsler IQ test: all TD participants received the Wechsler Abbreviated Scale of Intelligence (WASI; Wechsler 1999); ASD participants received the WASI [n = 34], the Wechsler Intelligence Scale for Children (WISC)-III [n = 2] (Wechsler 1991) or WISC-IV [n = 5] (Wechsler 2003), or the Wechsler Adult Intelligence Scale-III [n = 1] (Wechsler 1997). The TD and ASD participants were group-matched on and therefore did not differ significantly in age, IQ, or sex ratio (see Table 1). Written consent from parents (and participants when they were 18 or older) and verbal and written assent from participants under age 18 were obtained.

Table 1 Participant characteristics: Mean (SD)

Measures

Experimental Task

The Emotional MultiMorph Task: The stimuli in this task were taken from Ekman and Friesen’s Pictures of Facial Affect Series (Ekman and Friesen 1976) and consisted of faces depicting six emotional expressions—happiness, sadness, fear, anger, surprise, and disgust. The task was administered previously by Blair et al. (2001). Photo images were created for each emotion by morphing a full emotional expression with a neutral expression on the same face. Utilizing 179 key feature-points, the faces were delineated manually (Rowland and Perrett 1995). For each emotionally expressive face, spatial location differences were calculated for each pair of corresponding feature-points on the neutral and fully expressive (i.e., end-point) faces. Algorithms were then used to produce equally distributed morphs between these end-points (Benson and Perrett 1991). The result was 21 photo images or slides advancing in 5% increments from a 0% (neutral) emotional face to a 100% (fully expressive) emotional face (see Fig. 1 for a depiction of a 5% graded morph from neutral to full expression of sadness). These images were presented in order, beginning with the most neutral. Each facial image (i.e., 5% increment) was shown for 3 s. Two different faces depicting each of the six emotions were shown resulting in a total of 12 trials. These 12 trials were presented in one of four pseudo-randomized orders that were counterbalanced across subjects.

Fig. 1
figure 1

Morphing faces task: From neutral to full expression of sadness in 5% increments

Prior to beginning the task, participants were told that they would see facial expressions on the computer screen and that each face would start out neutral but would slowly change to show one of six emotions listed on their instruction sheet (placed in front of the participant for the duration of the experiment). Participants also were asked to describe each of the six emotions, prior to testing, to confirm their understanding of the emotion vocabulary. If participants had any confusion, the experimenter provided a definition of said emotion and confirmed that the participant understood before beginning the task. Participants were instructed to name verbally the emotion being shown as soon as they could confidently identify it (i.e., without making any random guesses). The experimenter warned participants that they would not be told whether their response was correct or not. Participants were also informed that they could change their response at any time during the presentation of the 21-slide sequence per face. Thus, during the study participants viewed and evaluated 252 faces (21 images for each of 12 trials).

Performance was assessed along two dimensions: (a) accuracy in identifying each emotion (0–100%) and (b) sensitivity to subtle expressions of emotion, as measured by the morph series/slide number at which correct emotional identification occurred (1–21). The slide score gauged a participant’s perceptual sensitivity and/or how quickly s/he correctly labeled each emotion. The final answer from each trial was used to evaluate accuracy. If a participant provided an incorrect final response—even if a previous response on that trial was correct—a slide score was not calculated so that including slide scores for emotions that were incorrectly identified would not skew the results. Participants may have provided considerably delayed or quick and impulsive incorrect responses. Either way, slide scores would be skewed if using inaccurate trials. The results of the two trials for each target emotion were collapsed into single average accuracy and average slide scores for each of the six emotions.

In order to maintain sample size for omnibus statistics, such as mixed-model analysis of variance, for individuals with no slide score (because they inaccurately labeled the emotion on both trials), the group (ASD vs. TD) mean was substituted (which did not change any results reported below). However, it is important to note that this group mean substitution was not completed for the correlational analyses.

Behavioral Correlates

Autism Diagnostic Observation Schedule (ADOS)—Generic (Lord et al. 2000): The ADOS is a semi-structured assessment used to evaluate individuals suspected of having ASD. It consists of various activities that allow one to observe social and communication behaviors related to ASD diagnoses. Either module 3 or 4, given the age and expressive language abilities of the participants here, was administered. Classification is determined by cutoffs for both the broader diagnosis of ASD, as well as a narrower conceptualization of autism. A trained researcher who had undergone reliability confirmation administered the ADOS.

Social Responsiveness Scale (SRS) (Constantino 2002): The SRS is a 65-item informant-based rating scale utilized to assess (mostly social-communication) traits associated with ASD across the full range of severity. It distinguishes individuals with ASD from controls (Constantino et al. 2000) and is highly correlated with one of the most widely used diagnostic instruments in the field, the Autism Diagnostic Inventory (Constantino et al. 2003). The SRS provides an overall summary score that was used here as a correlate with FER.

Adaptive Behavior Assessment System-II (ABAS-II) (Harrison and Oakland 2003): The ABAS-II is an informant-completed questionnaire designed to assess a participant’s adaptive skills. The parent-completed ABAS-II was used to evaluate an individual’s strengths and weaknesses in the domains of communication, community use, functional academics, home living, health/safety, leisure, self-direction, social, and (for participants 17 and up) work and has proven effective in assessing adaptive difficulties among high functioning adolescents with ASD. Here, the Global Adaptive Composite was used as an overall measure of adaptive functioning and therefore as a correlate with FER.

Data Analysis

To establish that accuracy and sensitivity were indeed relatively independent metrics of FER, correlations between each of these metrics, collapsing across emotions, were run. Accuracy in identifying the six basic emotions was analyzed nonparametrically (using the Wilcoxon-Mann–Whitney test) because of the categorical nature of the data and ceiling effects in the TD group. Perceptual sensitivity was assessed with a mixed-model analysis of variance (ANOVA) with group (ASD vs. TD) as the between subjects factor and emotion (six basic emotions) as the within subjects factor. Follow-up one-way ANOVAs were used to examine group differences in sensitivity to each emotion. Analysis of covariance (ANCOVA) was used to test group differences in perceptual sensitivity for each emotion, taking into account the group difference in overall perceptual sensitivity (collapsed across the six emotions). Pearson correlations between both age and FSIQ, and FER perceptual sensitivity for each of the six emotions as well as overall FER accuracy (across all six emotions) were run in the ASD group only in order to insure that these potential nuisance variables were not affecting performance. Subsequently, Pearson and partial correlations (whenever age or IQ was related to perceptual sensitivity) were run to assess the relationship between sensitivity to emotional expressions or overall FER accuracy and both autism symptomatology and adaptive functioning.

Results

Intriguingly, overall FER accuracy and perceptual sensitivity were not significantly correlated with one another in either the TD (r = .19, p = .31) or ASD (r = −.06, p = .72) group, suggesting that these are largely independent measures of FER. Collapsing across all six emotions, TD adolescents were more accurate overall in emotion identification than ASD adolescents (Z = 2.99, p = .003). Examining each emotion individually revealed that TD adolescents were more accurate in identifying only anger (Z = 3.30, p = .001; see Table 2). This effect, however, should be interpreted with caution because of a ceiling effect in the TD group; only one TD participant failed a trial of an expression of anger.

Table 2 Percent accuracy in emotion identification and percent intensity of facial expression needed to accurately identify each emotion by group: Mean (standard deviation)

A mixed-model ANOVA revealed main effects of group (F(1,58) = 14.08, p < .001), with the TD group showing greater perceptual sensitivity than the ASD group (see Table 2; Fig. 2), and emotion (F(5,54) = 12.51, p < .001). There was also a significant interaction of group by emotion (F(5,54) = 2.54, p = .03). Follow-up one-way ANOVAs revealed that the ASD group showed diminished perceptual sensitivity for each individual emotion (all Fs(1,68–1,71) > 6.37, ps < .02), except for fear, which demonstrated a trend in this direction (F(1,65) = 3.79, p = .06). If submitted to a Bonferroni correction, four (happiness, anger, sadness, disgust) of the six emotions remained significantly different between groups. After covarying for differences in overall perceptual sensitivity, only perceptual sensitivity to faces depicting sadness (of the six emotions) remained significantly different between the ASD and TD groups (F(1,70) = 4.75, p = .03). Adding FSIQ as a covariate did not alter the pattern of findings.

Fig. 2
figure 2

Cumulative percentage of adolescents with autism spectrum disorders (ASD) versus typically developing (TD) adolescents who accurately identified each of the six emotions at each intensity level

Results showed that perceptual sensitivity scores for faces depicting happiness (r = −.48, p = .002) and surprise (r = −.35, p = .03) were significantly correlated with FSIQ in the ASD group, whereas age was not significantly correlated with any perceptual sensitivity scores. Therefore, subsequent correlational analyses for happiness and surprise include both standard Pearson and partial (covarying the effects of FSIQ) values.

Pearson correlations revealed a consistent pattern in which clinician or parent ratings of everyday functioning (i.e., autism social-communication symptomatology and adaptive functioning) were associated with perceptual sensitivity to facial depictions of sadness (see Table 3). More specifically, diminished sensitivity to sadness was associated with a higher number of social and communication symptoms from the ADOS, greater endorsement of ASD traits from the SRS, and lower adaptive functioning scores on the ABAS-II Global Adaptive Composite. For autism-related behavior ratings as measured by the ADOS (social-communication symptoms only) and the SRS, sensitivity to faces depicting happiness also showed a positive correlation; however, once IQ was partialled only the association with the ADOS social-communication symptoms remained significant. Complementing these correlations with perceptual sensitivity, greater overall FER accuracy (across all six emotions) was found to be associated with increased adaptive functioning.

Table 3 Pearson (r) and partial (pr) correlations between behavior ratings and facial emotion recognition perceptual sensitivity and accuracy within the group of adolescents with autism spectrum disorders

A confusion matrix was constructed to examine whether any particular emotions were systematically mislabeled as certain other emotions and whether systematic mislabeling varied by group (Table 4). In general, ASD participants made similar errors to those made by TD participants, but in larger numbers. Participants most commonly labeled sad as angry or fear and labeled disgust as angry.

Table 4 Facial emotion recognition confusion matrix by group

Discussion

When asked to identify emotions using a morphing faces paradigm, adolescents with ASD showed overall diminished perceptual sensitivity to facial emotions relative to TD adolescents. More specifically, perceptual sensitivity to four of the six emotions was lower in ASD after correction for multiple comparisons. After accounting for the overall group differences in perceptual sensitivity to emotional faces, adolescents with ASD required higher intensity facial depictions of sadness, in particular, for accurate recognition. This finding is tempered in that it did not survive Bonferroni correction for multiple comparisons; nevertheless, it extends findings from a previous study showing ASD-related difficulty in accurately recognizing sadness both in faces and in animations without explicit social cues (Boraston et al. 2007). The significant correlation found here between diminished perceptual sensitivity to sad faces and the number of social-communication symptoms endorsed using the ADOS also expands findings from Boraston et al. (2007) and suggests the functional relevance of this deficit to symptomatology. Moreover, we extend these findings across raters and domains; not only clinician-rated social-communication symptoms (using the ADOS), but also parent-reported ASD traits (using the SRS) and adaptive functioning (using the ABAS-II) were associated with diminished perceptual sensitivity to sad faces. Though there was a main effect of group for FER accuracy (a measure largely independent of perceptual sensitivity based on correlational analyses), this finding was driven primarily by group differences in identifying faces depicting anger. The tentative finding of diminished accuracy (but not sensitivity) in identifying faces depicting anger has some precedence in the extant literature (Ashwin et al. 2006; Wallace et al. 2008). However, because there was a limited number of trials assessing FER accuracy for each emotion and there was a ceiling effect of accurately identifying angry faces among the group of TD adolescents, we remain cautious about this finding. Finally, overall FER accuracy (collapsing across all six emotions) was positively correlated with scores on a measure of adaptive skills, further linking FER with everyday functioning in ASD.

Difficulty processing sadness, a marker for emotional empathy, has been associated with a particular type of antisocial behavior, psychopathy (Woodworth and Waschbusch 2008), though not as strongly as diminished recognition of fear (Marsh and Blair 2008). Diminished perceptual sensitivity to sad faces in ASD also may relate to difficulties with empathy. Nonverbal processing of affect is a principal component of empathy (Elfenbein and Ambady 2002) and ratings of the intensity of facial sadness based on pupil size have been linked to empathic abilities (Harrison et al. 2007). Some studies have shown autism-related difficulties (even compared to clinical controls) in emotional empathy-related behaviors, such as responding to distress cues (Scambler et al. 2007; Sigman et al. 1992). Even more directly related to the FER results documented here, Rogers et al. (2006) found diminished sensitivity in recognizing sadness in a subset of ASD boys with high callous-unemotional trait ratings versus those with low ratings; all of whom were selected from a special school for children with behavioral difficulties (unlike the participants here who did not have externalizing behavior problems based on parent report [unpublished data]).

Although the current findings do not directly address emotional empathy in ASD, they likely reflect the enormous variability in the autism spectrum, including associated behavioral and emotional difficulties, particularly among higher functioning individuals (Kenworthy et al. 2005; Siegel et al. 1996). A model (Blair 2008) has been delineated in which those with psychopathy or high psychopathic traits demonstrate intact or better cognitive empathy (e.g., theory of mind) but low emotional empathy, as indicated by the processing of faces depicting sadness and fear. In contrast and perhaps inconsistent with our findings, according to this model, individuals with ASD provide a mirror image with intact emotional empathy and impaired cognitive empathy (Dziobek et al. 2008). Nevertheless, our findings add to a growing literature demonstrating emotional processing difficulties in ASD and suggest that emotional empathy (insofar as this is associated with perceptual sensitivity to sadness) is not intact in ASD.

Our results also indicate that FER deficits, particularly in sensitivity to perceiving sad facial expressions, are strongly linked to both symptomatology and real-world social functioning in this group. It is perhaps unsurprising to find these associations between FER and real-life behavior ratings made by clinicians and parents because everyday situations demand fast, efficient assessment of emotion during interactions. Failing to quickly and accurately perceive another’s emotional state can have dire consequences for an individual’s ability to communicate and interact with others. Furthermore, because people often do not encode a full, stereotypical emotional expression when communicating, sensitivity to subtle emotional expressions, even more than accuracy in identifying stereotypical expressions, may predict success in social interactions.

Alternatively, it could be that lower adaptive functioning and higher symptomatology ratings reflect the ASD group’s diminished experience engaging in social interactions (including identifying and processing emotions in others) which is in turn reflected in performance on the FER task, though this is unlikely to explain the specific findings regarding sadness. Regardless, the correlations between behavior ratings and the sensitivity score for sad facial expressions in the ASD group are not a function of overall poor performance across tasks; for example, IQ was not associated with sensitivity to faces depicting sadness. Similarly, the documented significant correlations were not due to capitalizing on greater variance in performance on this task, since slide scores to faces depicting surprise were associated with greater variability (i.e., a larger standard deviation) in performance than were slide scores to sad faces.

It could be that more subtle changes in facial movement associated with the neutral-sad morphs compared to the morphs of the other five emotions were more difficult to process for individuals with ASD and drove the correlations with everyday functioning. However, general perceptual discrimination (e.g., ‘spot the difference’) is usually viewed as a strength (or at least intact), not a weakness, in ASD (for review, see Happé and Frith 2006; Mottron et al. 2006). Furthermore, even after Bonferroni correction, the impairments in perceptual sensitivity of FER in ASD were not limited to sadness, but included other emotions with more pronounced facial movements, such as anger. Though the subtlety of change associated with neutral-sad morphs could have played a role in the findings, it does not completely explain them.

Also relevant to our findings are previous neuroimaging studies. For example, when asked to view morphing expressions of fear and anger, individuals with ASD demonstrated hypoactivation of the amygdala and the fusiform gyrus, and when contrasting morphed versus stereotyped depictions of these emotions, adults with ASD (as compared to control participants) irregularly modulated the amygdala, posterior superior temporal sulcus, and the fusiform gyrus (Pelphrey et al. 2007). Pelphrey et al.’s study is part of a larger literature implicating a crucial role for the amygdala in processing facial depictions of negative emotions in particular (Costafreda et al. 2008). In addition to its well-known role in fear perception (Adolphs et al. 1995), the amygdala appears to play a crucial role in perceiving sad facial expressions as well. Implicit processing of both masked (Killgore and Yurgelun-Todd 2007) and unmasked sad faces varying in intensity (Blair et al. 1999) has been shown to activate the amygdala in neurotypical children and adults, in contrast to other emotions (happy and angry) which did not activate the amygdala in these studies. Furthermore, Adolphs and Tranel (2004) found that patients with bilateral amygdala lesions rated sad morphs with less specificity than individuals with unilateral amygdala damage, other brain damage, or no brain damage. Finally, a recent study (Monk et al. 2010) provides evidence of atypical amygdala activation and functional connectivity (with cortical structures) among adults with ASD when viewing sad facial expressions (while completing a probe detection task). In addition to functional neuroimaging studies suggesting abnormal amygdalar activation patterns to emotionally expressive faces in ASD (Ashwin et al. 2007; Pelphrey et al. 2007; Wang et al. 2004), investigations employing structural neuroimaging techniques indicate an abnormal developmental growth trajectory of the amygdala in children and young adults with ASD (Schumann et al. 2004). Based on our findings here, future work should further explore the link between perceptual sensitivity to distinct facial emotion depictions and both brain structure and function among individuals with ASD.

In the current study, the similarity of error patterns to the various emotions in our ASD and TD groups suggests similar looking patterns. Although the ASD group was less accurate in recognizing angry faces, both groups most often mistook sad and disgusted faces for angry and confused fear and surprise. These results suggest that both ASD and TD individuals were led to misclassify emotions by looking at the same (misleading/difficult to process) parts of the faces. Future research could use eye-tracking technology to determine if aberrant scan paths might account for FER difficulties in the ASD group. Interestingly, in our own eye-tracking study (Snow et al., submitted) we found no differences in time spent focusing on the eye and mouth regions for ASD versus TD participants (a subset of whom participated in the present study) consistent with some recent reports (e.g., Anderson et al. 2006; Rutherford and Towns 2008), but not others (e.g., Pelphrey et al. 2002; Rutherford et al. 2007).

Though we document overall difficulties in FER perceptual sensitivity among adolescents with ASD relative to TD adolescents, it remains possible that response bias/caution could have been a contributory factor. In addition, it may be viewed as a weakness that we utilized a limited number of emotional face exemplars. Finally, using ANCOVA procedures in which the covariate (overall perceptual sensitivity in this case) contains variance contributed by the dependent variables of interest (perceptual sensitivity scores for each of the six emotions) is a potential weakness. This practice has been a point of considerable debate and discussion in neuroimaging work, where for example, regionally specific volumetric differences are presented uncorrected and after controlling for overall brain volume to account for effects of overall brain size differences. However, we remain confident in the validity and specificity of our finding of an ASD-related decrease in perceptual sensitivity to sad facial expressions, because, alone among the six emotions tested, the result for sadness (a) survived covariation accounting for overall differences in perceptual sensitivity across all six emotions (which could be considered an overly conservative correction) and (b) was the only index significantly associated with both autistic behavior ratings and adaptive skills. Moreover, using a limited number of emotional face exemplars allowed maximal perceptual sensitivity resolution (i.e., 5% increments) balanced against task length and demands on participants’ attentional resources. Finally, even when examining single trial data for each emotion (not presented here), the pattern of perceptual sensitivity results remained similar to those reported, suggesting a robust effect. Nevertheless, to address these concerns, future work should use a greater number of emotional face exemplars (most likely with greater incremental changes in emotional expressiveness) and utilize control (e.g., object) morphing tasks to account for potential response bias and to test general perceptual sensitivity to non-face and non-social stimuli.

Conclusion

The present study adds to a growing body of work showing that high functioning individuals with ASD have difficulty processing facial depictions of emotions, particularly in terms of perceptual sensitivity. Here we show that high functioning adolescents with ASD require greater intensity of sad facial expressions for accurate recognition than do TD adolescents. Furthermore, this diminished sensitivity to sad facial expressions was uniquely associated with not only autistic behavior ratings but also adaptive functioning. These findings may reflect purported empathy deficits associated with ASD.