Accurate facial-emotion recognition is fundamental in many contexts, and especially within reciprocal relationships. Accompanied by physiological responses, emotions are purposefully or passively conveyed through conduct, behaviour, and expression (Leman et al.,
2012). For neurotypically (NT) developed individuals, human faces provide unparalleled sources of socio-emotional data (Farah et al.,
1998; Kanwisher et al.,
1996). However, for individuals with a diagnosis of Autism Spectrum Disorder (ASD), differences in the fusiform face area (FFA) and amygdala influence gaze and memory, influencing the interpretation and response to non-verbal cues of others (e.g., Baron-Cohen et al.,
2000; Golarai et al.,
2006; Pelphrey & Carter,
2008; Schultz,
2005). Research suggests that emotion recognition differences are common within populations of individuals with an ASD diagnosis (Hobson,
1986). Challenges in distinguishing emotions can be affiliated with reductions in life-satisfaction and interpersonal difficulties (Carton et al.,
1999). The purpose of the current research was to explore how ASD-diagnosed adults and NT controls classify emoji representing the six basic emotions (Study 1), and whether the effects of emoji on the perceived emotionality of short narrative texts differ across participant groups and across the emoji used (Study 2).
Emotion Recognition Differences in ASD
Emotion recognition typically relies on holistic processing, utilizing spatial configuration of primary features (Diamond & Carey,
1986; Farrah et al.,
1998; Valentine,
1988). Fridlund’s (
1994) behavioural ecological perspective postulates that the human ability to interpret and express facial emotional cues serves as both an inference and predictor of behavioural intent (Adams et al.,
2006). Such cognitive and emotional proficiencies provide most humans with the ability to empathise (Baron-Cohen et al.,
2003). Ekman and Friesen (
1971) maintained that emotion directly reflects emotions
felt, as opposed to conveyance of behavioural intent (Fridlund,
1994). Following Ekman and Friesen (
1971), researchers delineated six basic cross-culturally identifiable expressions: happiness, disgust, fear, sadness, surprise, and anger. The communicative function of emotions has been established as integral within dynamic reciprocal relationships, and impairment of this ability—as proposed in ASD (Thye et al.,
2018)—has been scrutinised.
Described as a pervasive neuro-developmental condition, ASD affects 1 in 160 children worldwide (World Health Organisation,
2004; WHO). A genetically-inclined, neuro-developmental paradigm has been proposed to underpin autism (Folstein & Rosen-Sheidley,
2001), supported by the identification of unique patterns of brain development and activity in individuals with a diagnosis of ASD (e.g., Hill et al.,
2004). Many such individuals are prone to systemising—the rule-governed inductive process of data-gathering, quantifying differences and correlations to generate predictable results (Kidron et al.,
2018). Systemising is typically associated with males, object processing, and is applicable to phenomena which are lawful, finite, and deterministic (Baron-Cohen et al.,
2003). Lawson et al. (
2004) demonstrated that ASD was associated with systemising and divergence from empathising.
However, some of the ‘stereotypical’ findings in this area are problematic. A recent editorial by Fletcher-Watson and Bird (
2020) deconstructs relationship between autism and empathy. From the offset, the authors clearly identify that a major obstacle is that there is no single standardised unequivocal researcher definition of empathy/empathising. Young autistic children with concomitant intellectual disability have been found to be more likely to fail to detect another person’s emotional cues, due to differential orienting strategies in these children (Fletcher-Watson & Bird,
2020; Mundy,
2018). Chita-Tegmark (
2016) suggests that such differences might extend to adults, although this has been challenged (e.g., Johnson,
2014). Individuals not only have to perceive the emotional expressions/behaviours of another, they must be able to correctly identify this information correctly, and Harms et al. (
2010) have suggested that this is more difficult for autistic people. If emoji are used in interpersonal communication by senders to communicate their own emotional states, are these emoji: ‘recognised’ similarly by autistic and non-autistic individuals?, and are the effects of emoji on the perceived sentiment of written texts the same for autistic and non-autistic people?
Another part of the process, as described by Fletcher-Watson and Bird (
2020) is the embodiment of the emotional signals of another person—that is, experiencing the same emotions. Finally, an autistic person might be perceived as non-empathetic due to their responses to the emotional situation they are involved in. Fletcher-Watson and Bird (
2020) argue that this is not the outcome of a ‘lack of empathy’; rather, the autistic person is merely following a different “response-script” to that of an NT individual (Fletcher-Watson & Bird,
2020, p. 3). Milton (
2012) has suggested a
double empathy problem underlying patterns of research and real-world data. That is, challenges around communication and understanding between autistic and non-autistic people should not be seen as one-sided—rather, these complications are resultant from different perspectives of the communicators. For example, Edey et al. (
2016) and Sheppard et al. (
2016) demonstrate the non-autistic participants demonstrated difficulties when attempting to evaluate the emotional expressions of autistic persons (Fletcher-Watson & Bird,
2020).
Currently, a fully neurobiological model of ASD is lacking (Sivaratnam et al.,
2015); hence, ASD is predominantly explained via cognitive models. Historically, conversations around differences in Theory of Mind (ToM; Premack & Woodruff,
1978) abilities were predominant. ToM outlines one's capacity to predict mental states, and thus actions, intentions, and beliefs of those around them (Frith & Frith,
2003; Wang,
2015; Wellman,
1992). Difference in ToM abilities between ASD and NT samples have been suggested by studies of cognition-based emotion recognition (Baron-Cohen et al., 1993). However, such claims have been contested, suggesting that results may have arisen from experimenter bias and social conditioning (Fiene & Brownlow,
2015; Said et al.,
2011). Chevallier et al. (
2012) argued that children diagnosed with ASD perform poorly in ToM tasks administered by a researcher in a face-to-face context; such testing constructs a social situation, thus misrepresenting the performances of participants with ASD relative to NT controls. Chevallier et al. examined this by administering the false-belief test via computer, as opposed to in-person. Although NT individuals outperformed participants with ASD in traditional researcher-administered trials, no difference was found between-groups when administered via computer. This implies sensitivity differences to social situations only.
Indeed, researchers have recently begun to partial out the variability associated with alexithymia and autism. Alexithymia and autism are distinct but potentially co-morbid considerations (Fletcher-Watson & Bird,
2020). Alexithymia is characterised by difficulties in identifying emotional arousal and feelings (Nemiah et al.,
1976). Alexithymia affects approximately 50% of individuals with autism (Bird & Cook,
2013), as opposed to 10% of the general population (Salminen et al.,
1999). Previous research posits that alexithymia might underlie the stereotypical impairment of emotion recognition in ASD populations (e.g., Cook et al.,
2013; Grynberg et al.,
2012; Ketelaars et al.,
2016; Swart et al.,
2009) and has led to the formulation of the Alexithymia Hypothesis (Bird & Cook,
2013). Work by Brewer et al. (
2015) suggests that autism may be associated with non-typical ToM but not impaired empathy, whereas alexithymia may be associated with non-typical empathy but not atypical ToM (Brewer et al.,
2015; Fletcher-Watson & Bird,
2020).
Autism has been framed by an interest model—that is, characterised by monotropic attention strategies, repetitive behaviours and interests, and attentional ‘tunnelling’ (e.g., Lawson,
2010; Murray et al.,
2005). Monotropic theories posit that autism is defined by ‘single-minded’ attentional systems, with selects one information source at a time, which might result in certain social cues being neglected if another source of information is more-salient (Fletcher-Watson & Bird,
2020; Murray et al.,
2005).
Historically, it was believed that emotion recognition was unequivocally impaired in individuals with ASD, through failure to accurately comprehend others’ emotional states (Hobson,
1986). This has been countered by data from studies with well-matched pairs (e.g., Ozonoff et al.,
1990). Pelphrey et al. (
2002) utilised photographs representing the six basic emotions. Their study consisted of two phases: in the first, visual scan paths were examined whilst ASD and NT participants viewed images; in the second, emotion identification accuracy between-groups was compared. Five male, high-functioning ASD-diagnosed and five male NT participants were recruited. ASD-diagnosed participant scan-path analyses were consistent with highly variable viewing patterns of external facial features (e.g., ears, chin, hairline); NT controls showed consistent strategic paths over internal facial features (e.g., eyes, nose, mouth). Phase two demonstrated differences in emotion-recognition accuracy between-groups, with greater judgement-diversity evident in ASD-diagnosed participants. Although seemingly confirmatory of NT individuals outperforming their ASD-diagnosed peers, only fear recognition was significantly different—most-commonly mistaken for disgust or anger
. Significant differences were not observed for the remaining five emotions. These results were obtained from small samples, with a lack of matched-pairs, and all-male participant pool.
Uljarevic and Hamilton’s (
2012) meta-analysis encompassed 48 studies (
N = 980) involving ASD-diagnosed participants. Of these, 28 utilised Ekman and Friesen’s (
1976) facial affect stimuli. Studies incorporating measures of full-scale intelligence quotients (FSIQs) and a wide participant age range (6–41 years) were analysed. This meta-analysis found no significant difference among ASD-diagnosed participants in happiness recognition (applied as a baseline in the absence of neutral face data), sadness, surprise, disgust, or anger. Fear was acknowledged as less-accurately recognised than happiness, but with only marginal significance. Uljarevic and Hamilton’s (
2012) evaluation demonstrated surprise was no more misperceived than any other emotion; however, they acknowledge complexities in drawing comparisons between the emotions most- and least-accurately recognised, given that only eight studies compared all six emotions.
Uljarevic and Hamilton (
2012)’s meta-analysis found no effects of age or IQ on emotion recognition; hence, recognition differences are not necessarily subgroup-specific for ASD-diagnosed individuals (e.g., “lower-functioning” individuals). Studies which matched participants on IQ were, at best, indicative of ASD-diagnosed participants preforming at the expected level for their
mental age, as opposed to analogous with individuals of the same
chronological age. Consistent performance in happiness recognition appears to oppose a universality of atypicality; poorer fear recognition aligns with theories associating reduced eye-contact and poorer amygdalaic fear processing. Uljarevic and Hamilton (
2012) propose that previous findings may be mediated by stimulus timings—given ASD-diagnosed individuals’ divergent looking-patterns, results may reflect limited processing-time rather than recognition difficulties. Collectively, results indicate atypical facial processing in ASD-diagnosed samples, suggesting a mechanism which actualises social information processing differences in ASD.
Emotion Recognition and Online Communication
Mazurek (
2013) argues that a reduction in peer-engagement for ASD-diagnosed individuals is associated with decreased life-satisfaction, increased anxiety, depression, and low self-esteem. Social media provides opportunities for ASD-diagnosed individuals to interact with peers in environments void of non-verbal communicative cues, having less socially-regimented rules of engagement, lack of eye-contact, and reduced reliance on non-verbal cues of facial affect and emotional decoding (Burke et al.,
2010). Emoji are frequently used in online interactions and communications, as a proxy for face-to-face interactions.
Emoji are pictorial images which can mimic facial expressions and are considered a paralinguistic medium through which attitudes, emotions and narratives are shared, often in conjunction with written text (Rodrigues et al.,
2017). Kaye et al. (
2017) state that emoji serve two primary functions: (i) portray emotional or social intent, (ii) reduce potential discourse ambiguity. Social Information Processing theory (Walther,
1992) states that
cues within online communications (amongst which emoji can be considered) develop and maintain relationships (Rodrigues et al.,
2017). Skovholt et al. (
2014) highlight that emoji function as context cues, attitude markers and social relationship organisers (e.g., decreasing formality). Research has shown both face- and face-emoji-related activation of the occipital-temporal cortex (Churches et al., 2014), suggesting that via associative learning, emoji processing lies parallel to human facial processing, and the associated emotion represented (Bai et al.,
2019).
In contrast to human face-processing research, studies suggest that ASD-diagnosed individuals are adept at recognising cartoon faces (Rosset et al.,
2007; van der Geest et al.,
2001). Atherton and Cross (
2018) highlight that ASD-diagnosed participants showed increased engagement with anthropomorphic images. Attentional bias research has shown that ASD-diagnosed individuals demonstrate increased fixation on cartoon-style characters, relative to real objects (van der Geest et al.,
2001). Hence, it may be presumed the use of cartoon-type faces (i.e., emoji) influence emotion recognition abilities in ASD-diagnosed populations.
Emoji and Language Processing
Emoji are frequently used alongside written language. Similar to emotion recognition, ToM can be aligned with text valence processing, stating that text comprehension depends on readers’ capacity to attribute others’ cognitive and affective states (Abu-Akel & Shamay-Tsoory,
2011). Emoji might be of benefit when considering Milton’s (
2012) double empathy problem—that is, the emoji might simultaneously enable the sender to convey their emotional state/intention and serve as a cue to the recipient to aid their own perception of the message and select an appropriate response more clearly. Pictorial representations may aid text interpretation (Walther & D’Addario,
2001); Derks et al. (
2007) and Lo (2008) demonstrated that emoticons strengthened emotional sentiment of texts, biasing readers toward emoticon valence. González-Ibáñez et al. (
2011) and Muresan et al. (
2016) found that emoticons were influential in classifying sarcastic, non-sarcastic, positive, and negative tweets. Thompson and Filik (
2016) stated that emoticons can reduce negative responses typically experienced in response to ironic texts. Walther and D’Addario (
2001) used artificially-created emotive emails containing either positive or negative emoticons. However, results indicated valence perceptions were unaffected, implying emoticons’ emotional influence was overshadowed by text sentiment—except for negative text accompanied by negative emoticons.
Many studies have focused primarily on conversational formats involving response dialogues (Riordan & Kreuz,
2010; Rodrigues et al.,
2017). Questions remain as to whether emoji impact on different texts, i.e., narrative sentences composed from an external third-person perspective. Willoughby and Liu (
2018) compared narrative and non-narrative sentences via iMessage conversations, containing either three (high frequency), one (low frequency), or no emoji. Results suggest that iMessages without emoji elicited greater levels of credibility and elaboration, whereas a higher number of emoji drew greater attentional focus, regardless of narrative format presented.
Robus et al. (
2020) examined the effects of emoji position and expression in neutral narrative sentences on eye movements during reading and subjective ratings of sentence emotional valence. Pre-tested neutral sentence stimuli were used, allowing for a purer measurement of emoji effects. Two emoji were used, identical in colour and formatting, which differed in expression—slightly smiling (
) and slightly frowning (
). Emoji influence on text valence was predominantly non-significant; this may have been a result of the lack of ‘strength’ of the emoji used and the artificial laboratory eye-tracking set-up.