Frontotemporal dementia (FTD) is an umbrella term used to describe a group of heterogeneous, progressive neurodegenerative brain disorders, characterized by degeneration of the frontal and/or temporal lobes (Piguet and Kumfor, Reference Piguet and Kumfor2020). It is the second most common cause of young-onset dementia (i.e. before the age of 65 years) after Alzheimer’s disease (AD) with an estimated prevalence of 7–15/100,000 individuals (Coyle-Gilchrist et al., Reference Coyle-Gilchrist2016; Hogan et al., Reference Hogan2016). FTD is clinically classified into three subtypes based on the most prominent clinical features at the time of presentation. The most common form is the behavioral variant of FTD (bvFTD), which is characterized by altered social behavior and personality (Rascovsky et al., Reference Rascovsky2011) arising from atrophy primarily involving orbitofrontal, anterior cingulate and anterior temporal regions bilaterally, as well as the subcortical nuclei. The other two subtypes, semantic dementia (SD) and progressive non-fluent aphasia (PNFA) (also referred to as semantic variant and non-fluent variant of primary progressive aphasia), are characterized primarily by language disturbances (Gorno-Tempini et al., Reference Gorno-Tempini2011). Clinically, SD presents with a progressive loss of word comprehension and conceptual (word) knowledge against a background of fluent speech output. SD is associated with marked asymmetric atrophy of the anterior temporal lobe, which is generally more pronounced in the left than in the right hemisphere. In ∼30% of patients, the pattern of atrophy is reversed (i.e. right greater than left), hereon labeled SD-right. These patients tend to show additional behavioral changes, similar to those observed in bvFTD, as well as prosopagnosia (Kumfor et al., Reference Kumfor2016; Ulugut et al., Reference Ulugut2021). PNFA patients exhibit effortful and halted speech in the context of preserved comprehension. Speech output is markedly reduced and often distorted and can be accompanied by agrammatism. These deficits are associated with focal left hemispheric brain atrophy, involving primarily the left inferior frontal and insular regions.
Social cognition, of which a central aspect is the ability to recognize emotions in facial expressions, is essential to all social interactions (Frith, Reference Frith2009). Deficits in facial emotion recognition have a profound impact on interpersonal social interactions, where failure to recognize, identify, and respond to emotional stimuli can lead to misinterpretation of social cues that guide normal behavior. Facial emotion recognition is supported by a large network of interconnected brain regions that includes anterior (orbitofrontal, temporal, insula, cingulate) posterior (parietal, fusiform), and subcortical (amygdala, thalamus) brain regions (Hutchings, et al., Reference Hutchings, Palermo, Piguet and Kumfor2017; Marshall et al., Reference Marshall2019; Van den Stock et al., Reference Van den Stock, Vandenbulcke, Sinke, Goebel and de Gelder2014).
Disturbance of facial emotion processing is well established in bvFTD and SD (Kumfor and Piguet, Reference Kumfor and Piguet2012). These deficits have been consistently observed, regardless of the type of stimuli (e.g. static or dynamic; black and white, or color; with or without contextual information) and the complexity of the emotional information portrayed (i.e. basic vs. complex emotions) (Goodkind et al., Reference Goodkind2015). Unsurprisingly, emotion recognition failure is also associated with reduced quality of the relationship with carers and increased caregiver burden (Hsieh et al., Reference Hsieh, Irish, Daveson, Hodges and Piguet2013; Spitzer, et al., Reference Spitzer, Shafir, Lerman and Werner2019).
On testing using basic emotions (anger, disgust, fear, happiness, sadness, and surprise), bvFTD and SD patients experience greatest difficulty recognizing and differentiating between the negative emotions (Kumfor et al., Reference Kumfor2011; Kumfor and Piguet, Reference Kumfor and Piguet2012; Savage et al., Reference Savage, Lillo, Kumfor, Kiernan, Piguet and Hodges2014). This difficulty is observed when performance is compared against that of healthy controls or against patients diagnosed with AD (Bora et al., Reference Bora, Velakoulis and Walterfang2016)—where emotion recognition deficits in this group tend to be milder, occur later in the disease process, and tend to be secondary to other cognitive deficits (e.g. attention, language, memory) (Bertoux et al., Reference Bertoux, de Souza, Sarazin, Funkiewiez, Dubois and Hornberger2015). Emotion processing is thought to be mostly preserved in PNFA, although the limited evidence is mixed, and when present is comparatively mild (Couto et al., Reference Couto2013; Piguet et al., Reference Piguet, Leyton, Gleeson, Hoon and Hodges2015).
To date, no research has examined the type of errors made during facial emotion recognition tasks in FTD and whether specific error patterns exist across the different subtypes. Analysis of error patterns will help determine if the errors committed reflect systematic errors between specific emotions, and whether errors involve subtle discrimination (e.g. fear vs. surprise), or gross recognition failures (e.g. happiness vs. sadness). In addition to improving understanding of the clinical phenomenology of FTD subtypes and potentially aid with diagnosis, knowledge of emotion processing error profiles may assist with the management of the disease, reduce carer burden, and help improve the social interactions between patient and carer.
To gain insight into error patterns in facial emotion recognition across FTD subtypes, we analyzed the errors made during a task of emotional face identification involving the six basic emotions and neutral faces: the Facial Affect Selection Task (FAST) (Kumfor et al., Reference Kumfor, Sapey-Triomphe, Leyton, Burrell, Hodges and Piguet2014; Miller et al., Reference Miller, Hsieh, Lah, Savage, Hodges and Piguet2012). Performance of FTD patients was compared against healthy controls and patients diagnosed with AD, to identify whether patterns of errors reflect a global effect of dementia (i.e. that would be similar across groups) or whether errors are group specific.
The overall aim of the study was to (1) identify the patterns of deficits on a task of basic emotion recognition in FTD subtypes compared with typical AD and (2) determine whether the types of errors (i.e. which emotions are mostly likely to be wrongly identified) and error patterns (i.e. which incorrect responses are most likely to be selected) differ across groups. Based on previous research, we hypothesize that: (i) all patient groups—but particularly bvFTD and SD—will experience disturbance in facial emotion recognition compared with healthy controls, most pronounced for negative emotions; (ii) the profiles of errors will differ across the dementia groups due to their different patterns of brain atrophy; and (iii) within groups, the incorrectly selected emotions (i.e. errors) will be perceptually similar to the target emotion.
Methods
Participants
Two-hundred-and-two individuals diagnosed with dementia (bvFTD: 62, SD-left: 29, SD-right: 14, PNFA: 21, AD: 76) were recruited from FRONTIER, the frontotemporal dementia clinical research group at the University of Sydney, Australia. All dementia patients underwent a comprehensive clinical and cognitive examination, a structural brain magnetic resonance imaging (MRI), supplemented by a carer-based interview. All patients met relevant current diagnostic criteria at the time of testing (Gorno-Tempini et al., Reference Gorno-Tempini2011; McKhann et al., Reference McKhann2011; Rascovsky et al., Reference Rascovsky2011). Clinical diagnosis was established by a multidisciplinary team including a behavioral neurologist, neuropsychologist, and occupational therapist. As part of the cognitive evaluation, participants completed the Addenbrooke’s Cognitive Examination (ACE), either the Revised or third version (Mioshi et al., Reference Mioshi, Dawson, Mitchell, Arnold and Hodges2006, Reference Mioshi, Hsieh, Savage, Hornberger and Hodges2010; Hsieh et al., Reference Hsieh, Irish, Daveson, Hodges and Piguet2013; So et al., Reference So2018). The ACE is a cognitive screening instrument that includes measures of attention, memory, fluency, language, and visuospatial function. Where necessary in the current study, ACE-R Total, Attention, Language and Visuospatial subdomain scores were converted to equivalent ACE-III scores as previously described (So et al., Reference So2018). Disease severity was measured using the Frontotemporal Lobar Degeneration Modified Clinical Dementia Rating Scale Sums of Boxes (CDR-FTLD; Knopman et al., Reference Knopman2008), where higher CDR-FTLD scores denote greater disease severity.
Ninety age- and education-matched healthy controls were also included in the study. Controls were recruited from our panel of volunteers or from the community and scored >88/100 on the ACE. Exclusion criteria for all participants included the following: concurrent psychiatric diagnosis, presence of other dementia or neurological disorders affecting the central nervous system, traumatic brain injury with loss of consciousness >5 minutes, and history of alcohol or substance abuse.
All participants or their person responsible provided written informed consent in accordance with the Declaration of Helsinki. The South Eastern Sydney Local Health District and the University of New South Wales ethics committees approved the study.
Facial emotion recognition assessment
Facial emotion recognition was assessed using the FAST (Kumfor et al., Reference Kumfor, Sapey-Triomphe, Leyton, Burrell, Hodges and Piguet2014; Miller et al., Reference Miller, Hsieh, Lah, Savage, Hodges and Piguet2012). In this task, participants are shown arrays of seven faces from the same person, expressing the six basic emotions (happy, angry, sad, surprise, fear, and disgust) and a neutral expression and asked to point to a verbally cued target (e.g. “point to the angry face”). Stimuli for this task were from the NimStim database (http://www.macbrain.org) and were unfamiliar identities to the participant. All faces were cropped to remove extraneous features such as hair, converted to grayscale and presented using Microsoft PowerPoint. The task comprised 42 trials and performance was untimed. No feedback was given during the task. One point was given for each correct answer and zero for an incorrect answer. Accuracy was converted to a percentage correct score for statistical analysis. In addition to the overall score, responses for each individual item were recorded to enable error analysis.
Statistical analyses
Data were analyzed using SPSS version 24.0 (IBM). Shapiro–Wilk tests were conducted to check assumptions of normality. All variables were normally distributed. Chi-squared test was used to assess dichotomous variables (i.e. sex). ANOVAs were used to assess group differences on demographic measures and overall performance on the FAST, followed by Bonferroni post hoc tests where appropriate. Performance on individual emotions was first analyzed using a repeated-measures ANOVA. Given the presence of a significant emotion by group interaction, group performance on each emotion was examined individually with univariate ANOVAs followed by Bonferroni post hoc tests where appropriate. Finally, the patterns of errors for each emotion were investigated with multivariate ANOVAs followed by Bonferroni post hoc tests where relevant. For all analyses, statistical significance was set at p < .05.
Results
Demographic and cognitive performance
No significant group differences were found for age (F(5, 286) = 1.756, p = .122), sex distribution (χ2 (5) = 4.552, p = .473), or years of education (F(5, 286) = 2.129, p = .062) (Table 1). Disease duration was similar across patient groups (F(4,197) = 1.891, p = .113). Overall, group differences in general cognition were present as demonstrated by the ACE total score (F(5, 286) = 68.607, p < .001), with all clinical groups performing significantly worse than controls (all p values < .001). Among the clinical groups, bvFTD and PNFA performed better than AD and SD-left on the ACE, with PNFA also outperforming SD-right (all p values < .007) (Table 1). Disease severity, as indexed by the CDR-FTLD, was significantly different across groups (F(5, 214) = 32.925, p < .001), with all the clinical groups scoring significantly higher than controls, with the exception of PNFA (Table 1). In addition, the PNFA group scored lower than the AD, bvFTD, and SD-right groups, and the SD-left group scored lower than the SD-right group.
Note. Values are mean ± standard deviation. *p < .001; #Chi-square value; highlighted cells denote significant differences with Controls. aSignificant post hoc tests across clinical groups. ACE = Addenbrooke’s Cognitive Examination; AD = Alzheimer’s disease; bvFTD = behavioral-variant frontotemporal dementia; CDR-FTLD = Frontotemporal Lobar Degeneration Modified Clinical Dementia Rating Scale Sums of Boxes; F:M = female : male; m = months; PNFA = progressive non-fluent aphasia; SD-L = semantic dementia, left-sided; SD-R = semantic dementia, right-sided; y = years.
Facial emotion processing
The total FAST score and performance on the individual emotions are presented in Figure 1 and Table 2. A significant group difference was present on total FAST performance (F(5, 286) = 36.24, p < .001) (Figure 1). Post hoc tests revealed that all clinical groups were impaired compared with controls (all p values < .001), with the exception of PNFA (p = .051). Among the clinical groups, the SD-right group performed significantly lower than all the other groups (all p values < .027), and the bvFTD group also scored lower than the PNFA group (p < .001). No other statistical differences were present across the clinical groups.
Note. Scores are shown as means ± standard deviations. Highlighted cells denote significant differences with Controls. aSignificant post hoc tests across clinical groups. AD = Alzheimer’s disease. bvFTD = behavioral-variant frontotemporal dementia. SD-L = semantic dementia left SD-R = semantic dementia right. PNFA = progressive non-fluent aphasia. **p < .001.
Analyses of the individual emotions using a repeated-measures ANOVA revealed an emotion by group interaction (F(30, 1425) = 5.360, p < .001). This interaction was investigated for each emotion with separate ANOVAs, which uncovered, in each instance, a significant group difference. Post hoc tests showed that, compared with controls, all clinical groups were impaired for the detection of disgust (all p values < .001). Furthermore, with the exception of PNFA, all groups were impaired for all the other negative emotions (anger, fear, sadness) as well as for surprise compared with controls (all p values < .002). The AD and bvFTD groups were also impaired for the detection of happiness and neutral (both p values < .001), and the SD-right group was also impaired for the detection of neutral emotions (p = .029).
Error pattern analyses
We investigated the frequency and profiles of errors made for each emotion. Overall (i.e. regardless of group membership), the most frequent errors were in response to the target emotions disgust and fear; together, these incorrect responses accounted for almost half of all errors made on this task (26.1% and 22.9%, respectively). The next most frequent incorrect response was to the target emotion sadness (15.6% of all errors made).
The patterns of errors are displayed in the heat maps in Figure 2. Errors mostly arose in response to negative target emotions and were generally within the same valence dimension (i.e. negative emotions were mistaken for another negative emotion). For each target emotion, we identified the primary error response (PER); that is, the emotion most commonly mistaken for the target emotion in each group (represented by black circles in each heat map, Figure 2).
Group differences in PER were investigated using multivariate analyses of variance. These analyses revealed two distinct patterns: PER for happy, disgust, anger, surprise, and neutral were the same for all groups (Figure 3A), except for surprise and neutral in the SD-right group (Figure 3B). In contrast, multiple PERs were observed across groups for fear and sadness (Figure 3C).
The PER for happy targets was surprise in all groups. Post hoc tests showed that AD and bvFTD tended to endorse surprise more frequently than controls (both p values < .001).
For disgust targets, the PER was anger in all groups. The AD, bvFTD, and SD groups were more likely to select this PER compared to controls (all p values < .001), and SD-left also endorsed anger significantly more frequently than PNFA (p = .001).
Response to anger targets revealed the opposite pattern with disgust being the PER for all groups. Post hoc tests showed that all groups, with the exception of PNFA, endorsed disgust more frequently than controls (all p values < .008). Among the clinical groups, SD-right selected the PER more frequently than AD, SD-left and PNFA, while bvFTD chose this PER more often than AD and PNFA (all p values < .002).
For surprise targets, the PER was fear for all groups, except for SD-right where it was happy. Post hoc tests showed that AD and bvFTD endorsed fear more frequently than controls (both p values < .007), while SD-right, AD, and bvFTD endorsed happy more often than controls (all p values < .015).
The PER in response to neutral targets was happy in all groups except for SD-right, where it was fear. Post hoc analyses showed that AD and bvFTD groups selected happy more frequently than the other groups (all p values < .001). The SD-right group selected fear significantly more frequently compared to all the other groups (all p values < .001).
The last two target emotions, fear and sadness, were associated with multiple PERs across groups (Figure 3). For the target emotion fear, four PERs emerged: surprise in controls and SD-left, sadness in SD-left and PNFA, disgust in bvFTD and SD-right, and anger in AD. Post hoc tests showed that SD-right endorsed surprise more often than controls (p < .008), while SD-left endorsed sadness more frequently than controls (p < .018), AD, SD-left and SD-right selected disgust more often than controls (all p values < .003), while SD-right chose disgust more frequently than PNFA (p < .032). Finally, AD and bvFTD endorsed anger more frequently than controls (both p values < .001).
For the target emotion sadness, three PERs emerged: disgust in bvFTD, SD-left and SD-right, fear in PNFA, and neutral in controls and AD. Post hoc tests revealed that disgust was endorsed more frequently by SD-right than by any other groups (all p values < .003) and more frequently by bvFTD and SD-left compared with controls (both p values < .001). All clinical groups selected fear more often than controls, with the exception of AD (all p values < .039). No significant group differences were observed with regard to the selection of neutral as an incorrect response.
Discussion
To our knowledge, this is the first study to systematically investigate patterns of error responses in FTD subtypes and AD in a task of basic facial emotion recognition. Although errors were committed by participants in all clinical groups, error patterns differ across groups. Not surprisingly, errors were predominantly observed in response to negative emotions. These incorrect responses tended to be primarily between emotions within the same valence dimension (e.g. fear vs. anger) and appeared to involve subtle discrimination errors, rather than across valence dimensions (e.g. happiness vs. sadness). Findings for each basic emotion are discussed below, as well as their implication for our understanding of facial emotion processing disturbance in these younger-onset dementia syndromes and their diagnosis.
Overall, a significant deficit in facial emotion recognition was observed in all dementia groups when compared with the healthy controls, supporting our first hypothesis, a finding consistent with previous research (Bora, et al., Reference Bora, Velakoulis and Walterfang2016; Hutchings et al., Reference Hutchings, Palermo, Piguet and Kumfor2017; Kumfor and Piguet, Reference Kumfor and Piguet2012). The severity of the deficit, however, varied across groups with bvFTD and SD-right showing the lowest accuracy rate overall. In contrast, the PNFA group was only mildly impaired, with the difference in performance with the control group approaching statistical significance. When examining each basic emotion separately, difficulties recognizing the facial emotions were most pronounced for the negative emotions (anger, disgust, fear, sadness), in line with previous studies (Kumfor et al., Reference Kumfor2011; Kumfor and Piguet, Reference Kumfor and Piguet2012; Park et al., Reference Park2017; Savage et al., Reference Savage, Lillo, Kumfor, Kiernan, Piguet and Hodges2014), and for surprise. Most commonly, the negative target emotions were primarily mistaken for other negative emotions. The notable exception, however, was, for the emotions surprise (positive) and fear (negative), emotions which have different valences but are perceptually similar and share more muscle movements than they possess distinctive ones (Roy-Charland et al., Reference Roy-Charland, Perron, Beaudry and Eady2014). In addition, difficulties with neutral detection were also identified in SD-right, bvFTD and AD, as well as with happiness in bvFTD and AD. In contrast to the well-established reports of a facial emotion recognition disturbance in bvFTD, findings in SD-right have been mixed (Kamminga et al., Reference Kamminga, Kumfor, Burrell, Piguet, Hodges and Irish2015; Mendez et al., Reference Mendez, Ringman and Shapira2015). The present study demonstrates that SD-right patients are indeed severely impaired on facial emotion recognition, and this deficit extends across all emotion types.
Turning our attention to the error profiles, we demonstrated that the types of errors made to the target emotions were similar across the clinical groups, providing limited support for our second hypothesis. The same PER was observed across the groups for three out of the seven displayed emotions (happy, disgust, anger). For another two emotions, surprise and neutral, all groups had the same PER, with the exception of SD-right, which tended to confuse surprise for happy (rather than for fear in the other groups) and neutral for fear (rather than happy in the other groups). For the remaining two emotions (sadness, fear), PERs across groups were more variable. Nevertheless, a common pattern emerged for sadness where three subtypes of FTD, SD-left, SD-right, and bvFTD, mistook this facial emotion for disgust significantly more frequently.
Further, although syndrome-specific error patterns were not identified, the PERs in bvFTD and SD-right were more commonly disgust. This emotion was never endorsed as the PER by the other groups, with the exception of anger where it was the PER for all groups including controls. Whether this default error response in bvFTD and SD-right reflects, a shared cognitive or biological mechanism will require further investigations but would align with the overlapping clinical features between these two clinical populations previously documented (e.g. changes in eating habits, increased behavioral rigidity) (Sato et al., Reference Sato2021; Younes et al., Reference Younes2022).
Notably, profiles of errors were not random, providing support for our third hypothesis. Indeed, errors tended to occur with emotions that were perceptually similar and followed a predicted model of errors based on confusability (Young et al., Reference Young, Perrett, Calder, Sprengelmeyer and Ekman2002). Nevertheless, reciprocal error patterns were found for only two emotions: disgust and anger in all groups, including controls. This shared disturbance would suggest the presence of a common breakdown in high-level cognitive emotion categorization for these two emotions. The absence of reciprocity and increasing variable patterns of error responses for the other emotions would indicate the combination of low- (perceptual, configural) and high- (classification) cognitive mechanisms. This is best illustrated with fear and sadness, emotions which had different PERs across groups and also greater within-group variability in errors, as shown by the heat maps in Figure 2. For these two emotions, errors were spread across most other negative emotions.
Although the dementia groups could not be distinguished based on PERs alone, our findings have important clinical implications during the clinical workup for dementia. Our results indicate that the recording of error responses during a facial emotion recognition task has relevance, but only if the differential diagnosis under consideration includes bvFTD or one of the SD subtypes. For the other groups, these data do not help differentiate across groups. It would then be sufficient to only record accuracy (i.e. correct or incorrect response). Moreover, for the sake of clinical efficiency, it appears that testing negative facial emotion recognition during screening cognitive assessment is sufficient, as recognition of positive emotions seems to be generally preserved across dementia groups.
Arguably, one limitation of this study relates to the method of testing and the use of static, black and white, two-dimensional faces, which bear little resemblance to real-life situations. In recent years, a growing body of research has used dynamic stimuli or images that have contained contextual information (Goodkind et al., Reference Goodkind2015; Kumfor et al., Reference Kumfor, Ibanez, Hutchings, Hazelton, Hodges and Piguet2018). Regardless of the testing method used, however, facial emotion processing disturbance has been found in bvFTD. Whether the pattern of errors would differ across FTD subtypes using these novel approaches remain to be investigated. The cross-sectional design of the study may also be another potential limitation. Indeed, it is possible that the profiles of errors may change with disease progression, which may help distinguish across these dementia syndromes.
In sum, this study was the first to examine systematically patterns of error response in facial emotion recognition across all subtypes of FTD. Research involving the SD-right subtype has been limited to date, and this study demonstrates that this group differs from the more common SD-left group and the other FTD groups. Nevertheless, this study also demonstrates that FTD subtypes cannot be distinguished based on error response alone. These findings add to previous research showing impaired facial emotion recognition in FTD subtypes, with bvFTD and SD-right having the poorest performance. It remains a challenge to clinically differentiate different subtypes of FTD, and more research should be done to gain insight into facial emotion recognition in FTD. Understanding the emotion recognition deficits specific to each FTD subtype may assist the management of the disease, as well as help to improve carer burden.
Conflict of interest
None.
Description of authors’ roles
K Gressie planned the study, analyzed the data, and wrote the initial version of the manuscript. F Kumfor was involved in the analysis of the data and manuscript revision. H Teng was involved in data collection and manuscript preparation. D Foxe provided statistical advice and was involved in manuscript revision and figure preparation. E Devenney was involved in data collection and manuscript revision. R Ahmed was involved in data collection and manuscript revision. O Piguet was responsible for the study design, supervision of data collection and analysis, and overall manuscript preparation.
Acknowledgements
This work was supported in part by funding to ForeFront, a collaborative research group dedicated to the study of frontotemporal dementia and motor neuron disease, from the National Health and Medical Research Council (NHMRC) (GNT1037746). In addition, this study was supported by NHMRC Project Grant (GNT1121791). FK is supported by an NHMRC Career Development Fellowship (GNT1158762). OP is supported in part by a NHMRC Leadership Fellowship (GNT2008020).