Literature DB >> 29038663

Repetition Blindness for Faces: A Comparison of Face Identity, Expression, and Gender Judgments.

Karen Murphy1, Zoe Ward2.   

Abstract

Repetition blindness (RB) refers to the impairment in reporting two identical targets within a rapid serial visual presentation stream. While numerous studies have demonstrated RB for words and picture of objects, very few studies have examined RB for faces. This study extended this research by examining RB when the two faces were complete repeats (same emotion and identity), identity repeats (same individual, different emotion), and emotion repeats (different individual, same emotion) for identity, gender, and expression judgment tasks. Complete RB and identity RB effects were evident for all three judgment tasks. Emotion RB was only evident for the expression and gender judgments. Complete RB effects were larger than emotion or identity RB effects across all judgment tasks. For the expression judgments, there was more emotion than identity RB. The identity RB effect was larger than the emotion RB effect for the gender judgments. Cross task comparisons revealed larger complete RB effects for the expression and gender judgments than the identity decisions. There was a larger emotion RB effect for the expression than gender judgments and the identity RB effect was larger for the gender than for the identity and expression judgments. These results indicate that while faces are subject to RB, this is affected by the type of repeated information and relevance of the facial characteristic to the judgment decision. This study provides further support for the operation of separate processing mechanisms for face gender, emotion, and identity information within models of face recognition.

Entities:  

Keywords:  emotion processing; face identity; face recognition; repetition blindness

Year:  2017        PMID: 29038663      PMCID: PMC5636015          DOI: 10.5709/acp-0221-3

Source DB:  PubMed          Journal:  Adv Cogn Psychol        ISSN: 1895-1171


INTRODUCTION

Temporal attention is typically investigated using rapid serial visual presentation (RSVP) tasks, in which a series of items (e.g., words, pictures) are briefly presented and participants are required to locate targets (critical items) embedded within a stream of distractors (Forster, 1970; Potter & Levy, 1969). When the RSVP items are presented in the same spatial location at the rate of eight to 12 items per second and the two critical items (C1 and C2) are different, both items are typically reported (Coltheart, Mondy, & Coltheart, 2005; Kanwisher, 1987). In contrast, when C1 and C2 are identical, report of the second critical item is impaired and this is known as repetition blindness (RB) (Kanwisher, 1987). According to the token individuation hypothesis (Kanwisher, 1987), visual information activates an existing memory representation (type) which combines with the episodic representation (token) of the item (e.g., where and when the item occurs in time and space, Arnell & Jolicoeur, 1997; Kanwisher, 1987). This type and token integration allows each item to form an individual memory representation (token individuated) which is used to recognize or report the item (Arnell & Jolicoeur, 1997). In RSVP streams, when C1 and C2 are different, both are typed and tokenized, allowing recognition or identification. When C1 and C2 are identical both are typed, but token individuation does not occur for C2, thereby producing RB for C2. An extensive body of research shows that RB occurs for words and objects (e.g., Arnell & Jolicoeur, 1997; Bavelier, 1994; Coltheart et al., 2005; Harris & Dux, 2005; Kanwisher, 1987; Kanwisher & Yin, 1993; Kanwisher, Yin, & Wojciulik, 1999). However, very few studies have examined RB for faces (Buttle, 2010; Coltheart, Bornhofen, Mondy, & Stephenson, 2004; Mondy, Coltheart, & Stephenson, 2004; Mowszowski, McDonald, Wang, & Bornhofen, 2012). Coltheart and colleagues (Coltheart et al., 2004; Mondy et al., 2004) demonstrated that RB occurred for famous faces, unfamiliar faces, nonfamous faces with differing facial expressions, and for different photographs of the same person. Buttle (2010) investigated RB for facial expression and gender judgments using four types of stimulus repetition: complete repeat (same individual and emotion), emotion repeat (different individual, same emotion), identity repeat (same individual, different emotion) and no repeat (different individual, different emotion). Emotion repetition did not produce RB for either the expression or gender judgments. For both judgment tasks, there was RB when C1 and C2 were completely identical (repeated identity and emotion) and the same individual with different emotional expressions (repeated identity only), although RB was larger for the complete repeat condition. While these results might suggest that emotion is immune to RB and that sufficient attentional resources must be available to process emotional information (Holmes, Vuilleumier, & Eimer, 2003; Pessoa, McKenna, Gutierrez, & Ungerleider, 2002), they are at odds with research demonstrating that the processing of emotional information is automatic (e.g., Batty & Taylor, 2003; Mogg & Bradley, 1999a, 1999b). A number of methodological issues may have impacted on the results of Buttle’s study, thereby eliminating emotion RB. The current study will address these issues to further examine the occurrence of RB for faces. In relation to Buttle (2010), one issue was that the study used the same number of single and dual target trials. This differs from many other RB studies that have used a small number of single target trials (e.g., approximately 5%) to control for response bias (e.g., Coltheart et al., 2005; Harris & Dux, 2005). The 50% proportion of single trials may have biased participants to report only one target in the repeat trials, thereby magnifying RB for the identity and complete repeat conditions. Another issue was that Buttle only used four faces, which may have increased participants’ familiarity with them and improved accuracy overall. Further, the faces also had visible hairstyles, which might have allowed the gender judgments to be made based on distinguishing features (e.g., hairstyles) rather than via the face recognition mechanisms. Hence, it is possible that the gender judgments were made through feature processing occurring outside the face processing system, whilst the emotion judgments utilized face recognition processes. This study addressed these issues by reducing the number of single target trials, increasing the number of face stimuli, and removing the hairstyles from the face photographs. The aim of the study was to examine the role of different repeated facial features (complete, identity, emotion) in the production of RB for gender, expression, and identity judgments to assess the occurrence of RB for these different processing mechanisms within the face recognition system. Various models suggest that identity, gender, and expression information are processed by different cognitive and neural mechanisms within the face processing system (e.g., Bruce & Young, 1986; Duchaine & Yovel, 2015; Haxby, Hoffman, & Gobbini, 2000). According to Bruce and Young’s (1986) cognitive model of face recognition, initially, faces are processed via a view-centered representation of the face and information about expression and lip movement are processed via separate mechanisms. If the face is familiar, it will activate the face recognition units (memory for the structure of faces—structural codes), while unfamiliar faces are processed via the directed visual processing pathway using pictorial codes of the face representation (Bruce & Young, 1986; Duchaine & Yovel, 2015). Thus, facial expression and face identity are processed via interrelated but separate pathways and this separation occurs early on in the visual perceptual processing stages (Fisher, Towler, & Eimer, 2016). Neural models of face processing also propose that face characteristics are processed via different pathways within the human brain (e.g., Duchaine & Yovel, 2015; Haxby et al., 2000). According to these approaches, face invariant aspects such as age, identity, and gender are processed via the ventral visual stream within the brain, which includes the regions of the fusiform face area and the inferior occipital cortex. Imaging data indicates that both familiar and unfamiliar faces are processed within the fusiform face area (e.g., Andrews, Davies-Thompson, Kingstone, & Young, 2010; Gobbini & Haxby, 2007). Variable facial attributes, such as speech and expression, are processed via a pathway from the occipital cortex to the superior temporal sulcus. Recent evidence suggests that the fusiform face area may also process facial expression (Bernstein & Yovel, 2015). However, Duchaine and Yovel (2015) suggest that expression processing within the fusiform face area reflects a general sensitivity to shape information, while the face processing area in the superior temporal sulcus responds to faces showing emotion information. The operation of separate pathways for expression and face identity processing is supported by studies of patients with prosopagnosia, who have impaired face recognition but are able to accurately judge the emotional expressions on faces (e.g., Humphreys, Donnelly, & Riddoch, 1993). Facial expression judgments activate brain regions (superior temporal sulcus and limbic structures) beyond those associated with face identification (beyond the fusiform face region, e.g., Habel et al., 2007; Hasselmo, Rolls, & Baylis, 1989; Haxby et al., 2000; Narumoto et al., 2000; Todorov & Engell, 2008), and fMRI studies show different neural correlates for facial identity and expression decisions (Jehna et al., 2011). Brain damage can produce impairments in emotion processing, leaving face identification unaffected. For example, bilateral amygdala lesions impair fear recognition (Adolphs, 1999; Adolphs, Tranel, Damasio, & Damasio, 1994; Jehna et al., 2011), and disgust recognition is impaired with disturbed connections of the striatum and basal ganglia (Hennenlotter et al., 2004; Jacobs, Shuren, Bowers, & Heilman, 1995; Sprengelmeyer et al., 1996). In these cases, face recognition or identification skills remain intact and only emotion processing is impaired. Thus, both the neural and behavioral data are consistent with the independent processing of emotion and face identity information within the face recognition system. It has been suggested that face, gender, and identity information might be processed separately (e.g., Ellis, Young, & Flude, 1990; Hole & Bourne, 2010). Some studies have shown that gender decisions are not necessary for face identification, supporting the independent operation of these processes (Calder et al., 2000; Ellis, Flude, Young, & Burton, 1996; Ellis et al., 1990; Wild et al., 2000; see Zhao & Hayward, 2010, for a different perspective). Other studies have shown that when participants categorized faces by gender, accuracy was greater than when they were asked to recognize faces from an entire set of faces (Wild et al., 2000) and that gender decisions can occur more rapidly than face identity judgments (Bruyer, Galvez, & Prairial, 1993). These results support the idea that gender decisions require the discrimination between relevant gender specific attributes of a face, while identification uses configural information (Wild et al., 2000). Neuropsychological and laboratory studies indicate that gender and identity judgments involve separate brain regions (Bruyer et al., 1993; Mattson, Levin, & Grafman, 2000; Sergent, Ohta, & MacDonald, 1992). For example, prosopagnosia patients can report the gender of a face, although they cannot identify the person (Bruyer et al., 1993; Tranel, Damasio, & Damasio, 1988). An ERP study found that both age and gender categorization tasks activated the fronto-central areas but not the occipito-temporal facial recognition areas (Mouchetant-Rostaing, Giard, Bentin, Aguera, & Pernier, 2000). Thus, the empirical evidence supports the operation of different pathways for the gender, expression, and face identity information within the face processing system. To date, no study has investigated these three aspects of face processing in relation to RB. Thus, this study compared RB for three different face judgment tasks when different components of the faces were repeated (emotion, identity, identity and emotion together). Utilizing three different face judgment tasks would manipulate the attentional demands of each task and allow examination of the impact of this factor on RB for different repeated face characteristics. By varying the repeated aspect of the face (identity, emotion, or emotion and identity) this study examined the vulnerability of the various face processing pathways to RB. This will determine if overall face identity or only certain facial features (emotion, gender) are susceptible to RB.

Method

Participants

Thirty-one (24 females, seven males) first year psychology students, aged between 17 and 65 years (Mage = 23.42, SD = 10.87) completed this study in return for course credit. All participants had normal or corrected to normal vision. The study was approved by the University Human Ethics Committee and participants provided written consent before completing the experiment. The data of five participants were excluded from the analysis due to poor performance on the single target catch trials in at least one judgment task (less than 70% accuracy, typically repeat responses). This exclusion criterion ensured that the RB effects were not biased by participants responding repeat as their default decision when they were unsure of their response. The final sample of 26 participants (21 females, five males) who had their data included in the analysis were aged between 17 and 65 years (Mage = 24.46, SD = 11.60).

Materials and RSVP Task Stimuli

Positive and Negative Affect Schedule

Affect has been shown to influence attentional resources (Flaisch, Junghöfer, Bradley, Schupp, & Lang, 2007). Therefore, participants completed the Positive and Negative Affect Schedule (PANAS; Watson, Clark, & Tellegan, 1988) to ensure that their mood state was within the normal range at the start of the experiment. Participants rated their current feelings (e.g., interested, distressed) on a five-point scale (1 = not at all, 5 = extremely) for each of the 10 positive and 10 negative affect items. Higher scores indicated higher levels of positive and negative affect.

Face and filler stimuli for the Rapid Serial Visual Presentation task

Faces were taken from the NimStim facial set (Tottenham & Nelson, 2000; Tottenham et al., 2009). The hair, ears, and clothing were removed from the photos of the faces to ensure identification could not occur through a single distinguishing feature (Allison, 2010). Thus, participants could only use the face to make their decisions about the gender, expression, and identity of the faces. Black and white photographs (2.8° × 3.4° of visual angle) of 12 different individuals (six male, six female) were displayed on white backgrounds. Each face was displayed with a happy and an angry expression (24 targets in total). Angry and happy expressions were used, as they are regarded as being comparable (opposites) for identification purposes (Bruce & Young, 1986; Ekman, 1999). Twelve filler stimuli were generated for use in the RSVP task. The filler stimuli were grayscale ovals placed randomly within a larger grey oval that were presented at the centre of a 4.8° × 4.8° of visual angle white rectangular background. These filler items served as masks within each trial of the experiment.

Procedure

The experiment was run using the DMDX software (Forster & Forster, 2003) on a Pentium IV 2.66GHz PC with a CRT flat screen monitor (1,024 × 724 resolution). Participants were seated 60 cm from the computer screen in a dimly lit room and completed three different RSVP tasks (face identification, gender, and expression judgment tasks) in a one hour session. For all tasks, each trial began with a central fixation cross displayed for 500 ms, followed by three filler items, C1, another filler item, C2, and three more filler items. As a result, each RSVP stream (trial) presented a nine-item sequence to participants. Single target trials included eight fillers and the critical item was presented in either the C1 or C2 position. Presentation rates were 90 ms per critical and filler item with no interstimulus interval. Figure 1 presents an example trial sequence for a complete repeat trial for the expression, gender, and identity judgment tasks.
Figure 1.

Example of the sequence of stimuli presented on a complete repeat trial for the expression, gender and identity judgment tasks. Sample face is image 01 from Tottenham et al. (2009).

Example of the sequence of stimuli presented on a complete repeat trial for the expression, gender and identity judgment tasks. Sample face is image 01 from Tottenham et al. (2009). At the start of each of the three judgment tasks, participants were informed that either one face, two identical faces, or two different faces would be presented on each trial. Progression between trials was self-paced (initiated by pressing the spacebar). There were four practice trials prior to each RSVP task, and task order was counterbalanced across participants to minimize practice effects. In each of the identity, gender, and expression judgment tasks, there were 96 trials, with each of the 12 happy and 12 angry faces being presented once in each of the four repetition conditions. The four repeat conditions included a complete repetition (same face and emotion), an emotion repetition (different face, same emotion), an identity repetition (same face, different emotion) and no repetition (different face, different emotion). An additional 10 trials, with only one critical item, were also included in each of the three RSVP tasks. For the expression and identity judgment tasks, C1 and C2 were of the same gender to prevent decisions being made on gender differences alone, which would have negated the expression and identity judgments required by the task. As the same faces had to be used for the complete and identity repeat conditions, the same gender faces were also used for the emotion and no repeat trials to ensure consistency of C1 and C2 presentations across all four repeat conditions. Had faces of different genders been used for the emotion and no repeat trials, this would have highlighted the presentation of the two different faces and confounded the RB effects. In the gender judgments, if the same gender faces were shown on every trial, then every response for the task, except the single trials, would have been female-female or male-male, which would have undermined the true gender decisions required in this task. To overcome this issue, the same gender faces were shown on the complete and identity repeat trials and different gender faces were presented on the emotion and no repeat trials. For the expression and gender judgments, participants wrote down the facial emotional expression/expressions (happy, angry) of the faces and face’s gender/genders (male, female) they observed within each of the RSVP streams (one, two repeats, or two different expressions or genders). For the identity judgment task, two different faces (same gender) with happy and angry emotional expressions appeared at the end of each trial (four faces shown, numbered one to four). Participants wrote down the number of the face/faces (single face, two repeats of the same face, or two different faces) they had observed within the stream. Responses were noted in an answer sheet for each RSVP task.

Design

This study used a fully repeated-measures design. The independent variables were: Judgment Task (expression, identity, gender judgments) and Repeat Condition (complete, emotion, identity, no repeat). The dependent variables were the percentage of trials with C1 and C2 correct and the percentage correct for the single critical item trials.

Results

Reliability Analysis of the Positive and Negative Affect Schedule

The positive (Cronbach’s α = .83) and the negative (Cronbach’s α = .81) affect scales of the PANAS (Watson et al., 1988) were reliable (overall = .94). All participants had scores on these measures that fell within the population norm range (Crawford & Henry, 2004; M = 27.26, SD = 6.16 for positive affect; and M = 13.43, SD = 3.67 for negative affect). Thus, no participant’s current positive or negative mood would have affected their task performance.

Single Target Catch Trial Data

A repeated-measures analysis of variance (ANOVA) was used to compare the accuracy data for the single target trials across the three judgment tasks. See Table 1 for relevant descriptive statistics. The effect of judgment task, F(2, 50) = 6.97, p = .002, ηp2 = .22, revealed poorer single target accuracy for the identity judgments than the expression, t(25) = 2.86, p = .008, d = 0.79, and gender judgments, t(25) = 3.16, p = .004, d = 0.88. There was no difference in accuracy for the expression and gender judgments, t(25) = 0.00, p = 1.00. Therefore, the task difficulty was greater for the identification judgments than for the gender and expression judgments, which would be due to the use of unfamiliar faces. The impact of greater task difficulty would be to reduce the magnitude of RB for the identity judgments. Therefore, this difference in single target trial accuracy does not negate any RB effects observed in the identity judgments task data in the following analysis.
Table 1.

Mean Accuracy (%), SE and Repetition Blindness Effects for the Expression, Gender, and Identity Judgment Tasks for the Different Target Repeat Conditions.

Repeat Condition
Complete RepeatRB Complete RepeatEmotion RepeatRB Emotion RepeatIdentity RepeatRB Identity RepeatNo RepeatSingle Target Trials
JudgmentM (SE)M (SE)M (SE)M (SE)M (SE)M (SE)M (SE)M (SE)
Expression26.92 (6.20)59.46 (6.23)55.61 (5.74)30.77 (5.18)83.49 (3.30)2.89 (1.43)86.38 (2.63)97.31 (1.31)
Gender32.69 (7.65)55.77 (7.95)80.61 (3.28)7.85 (1.91)64.26 (5.83)24.20 (5.40)88.46 (2.64)97.31 (1.05)
Identity28.69 (7.60)26.66 (8.14)50.16 (5.33)5.13 (3.88)47.28 (4.56)8.01 (4.14)55.29 (4.33)90.77 (2.14)

Note. RB Complete Repeat is difference between accuracy in no repeat and complete repeat conditions. RB Emotion Repeat is the difference between accuracy in the no repeat and emotion repeat conditions. RB Identity Repeat is the difference between accuracy in the no repeat and the identity repeat conditions.

Note. RB Complete Repeat is difference between accuracy in no repeat and complete repeat conditions. RB Emotion Repeat is the difference between accuracy in the no repeat and emotion repeat conditions. RB Identity Repeat is the difference between accuracy in the no repeat and the identity repeat conditions.

Repetition Blindness for the Face Expression, Gender, and Identity Judgment Tasks

A 3 × 4 (Judgment Task [identity, gender, expression judgments], × Face Characteristic Repeat Condition [complete, emotion, identity, no repeat]), repeated-measures ANOVA was used to examine RB in the three face judgment tasks. The effects of face characteristic repeat condition, F(3, 75) = 39.26, p < .001, ηp2 = .61, and judgment task, F(2, 50) = 36.88, p < .001, ηp2 = .60, and the interaction between judgment task and face characteristic repeat condition, F(6, 150) = 14.82, p < .001, ηp2 = .37, were all significant. See Table 1 for relevant descriptive statistics. A series of specific planned comparisons (2-tailed) were used as follows to further examine this interaction effect. First, the occurrence of an RB effect for the complete repeat, emotion repeat, and identity repeat conditions was examined by comparing accuracy for each of these different repeat types to accuracy in the no repeat condition. This analysis was conducted separately for each of the judgment tasks. The second set of analyses compared the magnitude of the three different RB effects (complete, emotion, identity) to determine which type of repetition produced the largest RB effect within each judgment task. The third set of analyses compared the magnitude of the RB effect for each repeat type (complete, emotion, identity) across the three different judgment tasks to determine which decision produced the largest RB effect for the different repeat types.

Repetition Blindness for the complete, emotion, and identity repetitions

There were large complete RB effects (reduced accuracy for the complete repeat compared to the no repeat condition) for the expression, t(25) = 9.55, p < .001, d = 1.87, gender, t(25) = 7.02, p < .001, d = 1.37, and identity judgments, t(25) = 3.27, p < .001, d = 0.64. Accuracy for the emotion repeat condition (same emotion, different face) was compared to accuracy in the no repeat condition to examine emotion RB. There was emotion RB for the expression, t(25) = 5.94, p < .001, d = 1.16, and gender judgments, t(25) = 4.11, p < .001, d = 0.81, but not for the identity judgments, t(25) = 1.32, p = .198, d = 0.26. RB for repeated face identity was examined by comparing performance in the identity repeat condition (same face, different emotion) to performance in the no repeat condition. There was identity RB for the gender judgments, t(25) = 4.42, p < .001, d = 0.87, and the identity RB effect approached significance for the expression, t(25) = 2.00, p = .056, d = 0.39, and the identity judgments, t(25) = 1.94, p = .064, d = 0.38. Thus, there was complete RB and identity RB effects for all three judgment tasks and emotion RB occurred for the expression and gender judgments.

Comparison of complete, emotion, and identity Repetition Blindness effects within each judgment task

For the expression judgments, the complete RB effect was larger than the emotion RB effect, t(25) = 5.26, p < .001, d = 1.03, or the identity RB effect, t(25) = 9.43, p < .001, d = 1.85. The emotion RB effect was larger than the identity RB effect, t(25) = 5.80, p < .001, d = 1.14. For the gender judgments, the complete RB effect was larger than the emotion RB, t(25) = 6.21, p < .001, d = 1.22, and the identity RB effects, t(25) = 4.30, p < .001, d = 0.84. The identity RB effect was larger than the emotion RB effect, t(25) = −2.82, p = .009, d = 0.55. For the identity judgments, there was more complete RB than identity RB, t(25) = 2.80, p = .010, d = 0.55, and, as noted above, no emotion RB. Thus, for the expression judgments, the complete RB effect was larger than the emotion RB effect, which was larger than the identity RB effect. In the gender judgment task, the complete RB effect was larger than the identity RB effect and smallest for the emotion RB effect. For the identity judgments, the complete RB effect was larger than the identity RB effect.

Comparison of complete, emotion, and identity Repetition Blindness effects across the judgment tasks

The magnitude of the complete RB effect was larger in the expression, t(25) = 5.30, p < .001, d = 1.04, and gender judgments than the identity judgments, t(25) = 6.48, p < .001, d = 1.27. There was no difference in the size of the complete RB effect for the expression and gender judgment tasks, t(25) = 0.75, p = .460, d = 0.15. In comparing the magnitude of the emotion RB effect across the three tasks, the expression judgments showed a larger emotion RB effect than the gender judgments, t(25) = 4.07, p = .001, d = 0.77. As noted above, there was no emotion RB for the identity judgments. The magnitude of the identity RB effect was larger in the gender than the expression judgments, t(25) = 3.95, p = .001, d = 0.78, and larger in the gender than the identity judgments, t(25) = 2.52, p = .018, d = 0.50. The magnitude of the identity RB effect did not differ for the expression and identity judgments, t(25) = 1.37, p = .183, d = 0.27. Hence, the complete RB effect was larger for the expression and gender judgment tasks than the identity judgments. Emotion repetition produced more RB for the expression decisions than the gender judgments and the identity RB effect was larger for the gender than the identity and expression judgments.

Discussion

This study compared RB for expression, gender, and identity judgments about faces to determine if the type of decision or the type of repeated face aspect affected the magnitude of the RB effects. There was a complete RB effect for all three judgment tasks. This is consistent with previous research for expression (complete repeat; Buttle, 2010), gender judgments of faces (Buttle, 2010), and unfamiliar and famous face identification (Coltheart, et al. , 2004 ; Mondy, et al. , 2004). Identity repeats (same face, different emotion) also produced RB in the three judgment tasks. This is consistent with previous studies showing identity RB for expression judgments (Buttle, 2010; Mowszowski et al., 2012), gender judgments (Buttle, 2010), and for famous faces with various expressions in a face identification judgment (Coltheart et al., 2004). Across the three judgment tasks, the magnitude of the complete RB effect was larger than the identity RB effects and larger than the emotion RB effects for the expression and gender decisions. The larger complete RB effect relative to the identity RB effect for the expression and gender decisions is consistent with Buttle (2010) and could occur via repetition of all facial elements in the complete repeat condition and only identity repetition (not emotion) for the identity RB effects. This study found emotion RB for the gender and expression judgments indicating that RB is not solely due to repeated face identity information but that it also occurs with repeated facial feature information (e.g., emotion). The emotion RB for the gender and expression judgments observed in this study contrasts with the lack of emotion RB reported by Buttle (2010) for these same decisions. The different results across the two studies may be due to a range of design differences. For example, this study used more faces than Buttle (current study used 12 faces, Buttle’s, 2010, study—four faces) and all distinguishing features (e.g., hair) were removed from the face stimuli. These methodological modifications would have made the expression and gender judgments more difficult than in Buttle’s study, thereby producing larger emotion RB effects. This idea is further supported by the fact that the complete repeat RB effects reported in this study for the gender and expression judgments were larger than those reported by Buttle (average of 26% compared to the current study average of 57.25%). The happy and sad faces used by Buttle were drawn from a different database than the happy and angry faces used in this study. Hence, the use of different emotions and faces that may have expressed different levels of emotionality may also account for the different emotion RB results across the two studies. Further, Buttle manipulated task as a between-subjects factor, which was a within-subject variable in this study. Thus, any of these design differences may have contributed to the different outcomes observed here compared to Buttle. The size of the complete RB effect was larger for the expression and gender judgments than face identity judgments. Accuracy for complete repeat trials was similar across all three judgment tasks, and therefore, this difference in RB magnitude was due to lower accuracy in the no repeat condition for the identity judgments compared to the other two judgment tasks. This outcome is consistent with the poorer overall single target accuracy for the identity than expression or gender judgments, confirming that the identity task was the most difficult decision. Importantly, the accuracy for the face identification task was comparable with previous RB studies (e.g., accuracy of 56 to 60% for the no repeat trials and 30 to 43% for the repeat trials; Coltheart et al., 2005; Harris & Dux, 2005; Hayward et al., 2010). This indicates that the complete RB effect for the identity judgments is consistent with the RB effects observed in previous studies. These judgement task performance differences are also consistent with studies that have demonstrated slower processing speeds for making face identity decisions compared to gender decisions (Bruyer et al., 1993; Palermo & Rhodes, 2007) and reports that expression judgments are so rapid that they occur almost without conscious awareness of the observer (Esteves, Dimberg, & Öhman, 1994; Esteves, Parra, Dimberg, & Öhman 1994; Murphy, Monohan, & Zajonc, 1995; Murphy & Zajonc, 1993). There was more emotion RB for the expression than the gender judgments and no emotion RB for the identity judgments. While emotion RB for the expression judgments would be expected, as this was the relevant facial characteristic for the judgment, the emotion RB for the gender judgments is a new finding. Moreover, it indicates that within an RSVP task, the processing of emotional information is automatic (Batty & Taylor, 2003; Mogg & Bradley, 1999a, 1999b) when making decisions about gender. The lack of emotion RB for face identity judgments is consistent with models of face recognition that posit the independent processing of emotion and identity information (e.g., Bruce & Young, 1986; Duchaine & Yovel, 2015; Haxby et al., 2000). It is also consistent with neural evidence showing that the amygdala only responds to emotional expressions when there are sufficient attentional resources to process this aspect of the faces (Pessoa et al., 2002). For example, within the expression judgment task, the face emotion was central to the task, and therefore, participants would have allocated their attentional resources to this feature during face processing. Further, as the gender and expression judgments were easier than the identity judgments, this allowed the participants to have additional attentional resources to process the face emotion during these tasks. Hence, for both the expression and gender judgments, the emotion of the face was attended to and tokenized upon presentation of Face 1. Therefore, when the emotion was repeated with Face 2, it could not be token individuated, and emotion RB occurred. In contrast, for the identity judgments, emotion was not encoded for either face as it was not relevant to the task decision, and the greater task difficulty limited the attentional capacity to process this task-irrelevant emotional information (Pessoa et al., 2002). This idea is supported by the similar performance accuracy in the emotion repeat and no repeat conditions for the identity judgment task, and it is consistent with Buttle’s (2010) suggestion that for RB to occur participants need to encode the repeated level of information. The lack of emotion RB for the identity judgments provides behavioral data that supports Duchaine and Yovel’s (2015) suggestion that the role of the fusiform face area in expression processing reflects a general sensitivity to shape information, and it is not an additional area for processing facial expression, as some have suggested (e.g., Bernstein & Yovel, 2015). Buttle (2010) reported that identity RB effects were of the same magnitude for gender and expression judgments. However, in this study, the identity RB effects were larger for the gender than the identity judgments and very small for the expression judgments. These results are consistent with participants attending to and encoding the overall face structure to make the gender and identity judgments. In contrast, when making expression judgments, participants focused their attention on this facial feature and ignored the overall face representation. This reduced identity RB, as the only face feature encoded was the nonrepeated emotion on the expression judgment trials, and thus, repeated face identity did not impact on task performance. For the expression judgments, the size of the RB effect was larger for complete than emotion repeats, which was larger than for identity repeats. In contrast, for the gender judgments, the complete RB effect was larger than the identity RB effect, and the emotion RB effect was the smallest. The identity judgments showed a larger complete RB effect than the identity RB effect and no emotion RB. These results indicate that the repeated facial component is processed differently if that aspect is relevant to the attentional requirements of the judgment task. According to the token individuation hypothesis, each judgment task produced RB for complete repeat trials because, although both C1 and C2 were typed, only C1 was token individuated (Kanwisher, 1987). The smaller identity than complete RB effects across the three tasks indicate that changing one feature (emotion) seems to assist with token individuation of the identity of the second face, thereby reducing identity RB. Emotion RB for the gender and expression judgments but not the identity judgments indicates that repeated presentation of one facial feature is sufficient to produce RB (only allow C1 to be token individuated), but the feature must be directly relevant to the task decision (operational during stimulus encoding). This idea is further supported by the larger emotion than identity RB effect in the expression judgment task, the emotion RB effect being larger for the expression than gender decisions and the identity RB effect being larger for the gender decisions than the identity and expression judgments. The complete, identity, and emotion RB effects can also be explained in relation to seeing faces multiple times in quick succession during real world face-to-face interactions. For example, the large complete RB effects for all tasks suggest that when the same person with the same emotion is encountered, their face representation is not token individuated for a second time, as their unchanged representation is not informative during this time. The smaller identity than complete RB effects suggest that changes in emotion during interactions are sufficient to update the representation of that person on some occasions (token individuate the second face). This would be vital for assessing changes in their emotional state and updates on the progress of the social interaction. When assessing a person’s identity, the lack of emotion RB (same emotion, different person) would provide the face recognition system with a strong signal that a new person has been encountered within the environment, thereby facilitating face recognition. In this case, emotion RB might slow down the face recognition processes which could have serious implications for survival. This contrasts with the encoding of facial emotion when making judgments about a person’s expression and gender. As these more global facial assessments might be used in observing unfamiliar faces, it would be vital to encode their emotional state in the very initial stages of face encoding and only signal alterations in emotional states within a short time period. Hence, the complete, identity, and emotion RB effects for faces appear to play a functional role in the operation of our face recognition processes within real world environments. In summary, the results of this study showed that the magnitude of the RB effect varied across the type of repeated facial information (e.g., more RB for complete than identity or emotion repeat conditions). It was also shown that the occurrence of complete, emotion, or identity RB effects for faces were affected by the task instructions, which directed participants to focus on certain aspects of the facial stimuli. Thus, the results of this study provide further support for the operation of multiple independent pathways (gender, expression, identity) within the face recognition system and suggest that task relevant facial characteristics and attentional resources can mediate the magnitude of RB for faces.
  47 in total

1.  The functional correlates of face perception and recognition of emotional facial expressions as evidenced by fMRI.

Authors:  M Jehna; C Neuper; A Ischebeck; M Loitfelder; S Ropele; C Langkammer; F Ebner; S Fuchs; R Schmidt; F Fazekas; C Enzinger
Journal:  Brain Res       Date:  2011-04-09       Impact factor: 3.252

2.  DMDX: a windows display program with millisecond accuracy.

Authors:  Kenneth I Forster; Jonathan C Forster
Journal:  Behav Res Methods Instrum Comput       Date:  2003-02

3.  Holistic processing underlies gender judgments of faces.

Authors:  Mintao Zhao; William G Hayward
Journal:  Atten Percept Psychophys       Date:  2010-04       Impact factor: 2.199

4.  Functional neuroanatomy of face and object processing. A positron emission tomography study.

Authors:  J Sergent; S Ohta; B MacDonald
Journal:  Brain       Date:  1992-02       Impact factor: 13.501

Review 5.  Neural systems for recognition of familiar faces.

Authors:  M Ida Gobbini; James V Haxby
Journal:  Neuropsychologia       Date:  2006-06-22       Impact factor: 3.139

6.  The role of the amygdala in implicit evaluation of emotionally neutral faces.

Authors:  Alexander Todorov; Andrew D Engell
Journal:  Soc Cogn Affect Neurosci       Date:  2008-10-01       Impact factor: 3.436

7.  Repetition blindness for rotated objects.

Authors:  William G Hayward; Guomei Zhou; Wai-Fung Man; Irina M Harris
Journal:  J Exp Psychol Hum Percept Perform       Date:  2010-02       Impact factor: 3.332

8.  Repetition blindness: type recognition without token individuation.

Authors:  N G Kanwisher
Journal:  Cognition       Date:  1987-11

9.  Preferential processing of threatening facial expressions using the repetition blindness paradigm.

Authors:  Loren Mowszowski; Skye McDonald; Danielle Wang; Cristina Bornhofen
Journal:  Cogn Emot       Date:  2012-04-23

10.  Expression is computed separately from facial identity, and it is computed separately for moving and static faces: neuropsychological evidence.

Authors:  G W Humphreys; N Donnelly; M J Riddoch
Journal:  Neuropsychologia       Date:  1993-02       Impact factor: 3.139

View more
  1 in total

1.  The role of discriminability in face perception: Interference processing of expression, gender, and gaze.

Authors:  Enguang Chen; Bingbing Xia; Yujing Lian; Qi Zhang; Xuexian Yang; Hailing Wang
Journal:  Atten Percept Psychophys       Date:  2022-09-08       Impact factor: 2.157

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.