Literature DB >> 35542977

Face learning via brief real-world social interactions induces changes in face-selective brain areas and hippocampus.

Magdalena W Sliwinska1,2, Lydia R Searle2, Megan Earl2, Daniel O'Gorman2, Giusi Pollicina2, A Mike Burton2, David Pitcher2.   

Abstract

Making new acquaintances requires learning to recognise previously unfamiliar faces. In the current study, we investigated this process by staging real-world social interactions between actors and the participants. Participants completed a face-matching behavioural task in which they matched photographs of the actors (whom they had yet to meet), or faces similar to the actors (henceforth called foils). Participants were then scanned using functional magnetic resonance imaging (fMRI) while viewing photographs of actors and foils. Immediately after exiting the scanner, participants met the actors for the first time and interacted with them for 10 min. On subsequent days, participants completed a second behavioural experiment and then a second fMRI scan. Prior to each session, actors again interacted with the participants for 10 min. Behavioural results showed that social interactions improved performance accuracy when matching actor photographs, but not foil photographs. The fMRI analysis revealed a difference in the neural response to actor photographs and foil photographs across all regions of interest (ROIs) only after social interactions had occurred. Our results demonstrate that short social interactions were sufficient to learn and discriminate previously unfamiliar individuals. Moreover, these learning effects were present in brain areas involved in face processing and memory.

Entities:  

Keywords:  fMRI; face perception; perception; perceptual learning

Mesh:

Year:  2022        PMID: 35542977      PMCID: PMC9396469          DOI: 10.1177/03010066221098728

Source DB:  PubMed          Journal:  Perception        ISSN: 0301-0066            Impact factor:   1.695


An extensive behavioural literature has demonstrated large differences between the recognition of familiar and unfamiliar faces. Studies demonstrate that participants perform significantly better in familiar compared to unfamiliar face recognition tasks (Ellis et al., 1979; Yarmey & Barker, 1971). Familiar face recognition is also less affected by changes in pose, lighting conditions, or picture characteristics than unfamiliar face recognition (Hancock et al., 2000; Hill & Bruce, 1996). Other factors also differentiate familiar and unfamiliar face recognition including image degradation (Burton et al., 1999), removal of external features (Ellis et al., 1979), viewpoint (Bruce, 1982), and context (Dalton, 1993). Matching unfamiliar faces also correlates more strongly with object matching than it does with familiar face matching (Megreya & Burton, 2006). Taken together, these findings demonstrate that familiar and unfamiliar faces are processed, to some extent differently, and that the mental representation of familiar faces is more robust to a wide range of environmental factors (Burton et al., 2011). Face processing is a complex cognitive operation that is processed across multiple brain areas (Haxby et al., 2000). These include the fusiform face area (FFA) (Kanwisher et al., 1997; McCarthy et al., 1997; Parvizi et al., 2012), the occipital face area (OFA) (Gauthier et al., 2000; Pitcher et al., 2009), the posterior superior temporal sulcus (pSTS) (Campbell et al., 1990; Pitcher & Ungerleider, 2021; Puce et al., 1997), and face-selective voxels in the amygdala (Morris et al., 1996; Pitcher et al., 2019). Many prior functional magnetic resonance imaging (fMRI) studies have investigated how familiar faces are processed in the brain (Gobbini & Haxby, 2007; Natu & O’Toole, 2011). The majority of these studies have focused on the role of the FFA. FFA activity was correlated with behavioural performance when participants correctly identified target faces (Grill-Spector et al., 2004) and the FFA is sensitive to identity changes of famous individuals (Rotshtein et al., 2005). Numerous studies have also reported greater activity in the FFA for familiar than unfamiliar faces (Elfgren et al., 2006; Gobbini et al., 2004; Pierce et al., 2004; Ramon et al., 2015; Sergent et al., 1992; Weibert & Andrews, 2015), although it is also important to note that some studies have reported no difference (Gorno-Tempini & Price, 2001; Leveroni et al., 2000). It is unclear why the existing literature is inconsistent regarding the role of the FFA in the recognition of familiar faces. One possible explanation is that the nature of the personal familiarity used across different fMRI studies can vary (e.g., personally familiar, visually familiar, or famous faces). Natu and O’Toole (2011) have suggested that this variation may account for the differences in the results reported across fMRI studies of familiar and unfamiliar face recognition. Neuroimaging studies have identified brain areas other than the FFA that also exhibit different neural responses to familiar and unfamiliar faces. For example, in one study, participants viewed photographs of unfamiliar, famous, and emotional faces (Ishai et al., 2005). Results showed famous faces and emotional faces elicited a greater neural response across a network of face-selective areas including the FFA, OFA, pSTS, amygdala, and hippocampus. A greater response to recently learned faces compared to unlearned faces in the hippocampus has also been reported in a study in which Caucasian participants learned unfamiliar South Korean faces (Ishai & Yago, 2006). Subsequent research by the same group further demonstrated that familiar and unfamiliar faces are associated with unique functional connectivity patterns between face areas in the occipitotemporal cortex and the orbitofrontal cortex (Fairhall & Ishai, 2007). Prior electroencephalography (EEG) studies of familiar and unfamiliar face processing have also demonstrated some neural sensitivity to familiarity. For example, the N250r event-related potential (ERP) component, recorded over the inferior temporal regions, is modulated by repetitions of face stimuli which is greater for familiar (famous) than unfamiliar faces (Schweinberger et al., 2004; Schweinberger et al., 2002). A more recent EEG study also demonstrated a robust dissociation in the neural response to personally familiar, versus unfamiliar faces that ranged from 200 to 600 msec after stimulus onset (Wiese et al., 2019). Studies have also shown that faces can become familiar over the course of an experiment resulting in differences between familiar and unfamiliar faces in ERP components (especially the N250). These studies have used computer-based face learning (Tanaka et al., 2006; Tanaka & Pierce, 2009), videos (Kaufmann et al., 2009), and photorealistic caricatures (Limbach et al., 2018), demonstrating that previously unfamiliar faces can acquire some level or familiarity from a computer screen in a lab environment. One of the problems with lab-based investigations of face learning is that they typically lack the context surrounding the way we meet new people day to day, effectively “controlling away” critical factors in learning (Burton, 2013; Jenkins & Burton, 2011). Of course, it is extremely difficult to study naturalistic face learning without sacrificing the levels of control available in a lab-based study. In this study, we investigated whether brief social interactions would be sufficient to bring about learning that could be measured by changes in both behaviour and neural response. To do so, we used a setting that allowed participants to meet previously unknown people, multiple times, interspersed with multiple imaging events. In doing so, we aimed to complement existing lab-based studies by providing evidence from more naturalistic learning. While this inevitably relinquishes some lab control, we hold that converging evidence from artificial and natural learning studies gives the best chance of providing an understanding of this complex and poorly understood phenomenon. In addition to face-selective areas, another brain area implicated in the recognition of familiar faces is the hippocampus. While it is not a visual area of the brain, the hippocampus is heavily implicated in memory, suggesting that it supports parallel judgments for familiarity and recollection (Squire et al., 2007). Prior fMRI studies have also demonstrated that the neural response to familiar and unfamiliar faces can be dissociated in the hippocampus (Elfgren et al., 2006; Ishai et al., 2002; O’Neil et al., 2013; Platek & Kemp, 2009; Ramon et al., 2015). In addition, neuropsychological studies have shown that cells in the hippocampus increase firing rates in response to familiar faces (Fried et al., 1997) and that removal of the amygdala and the hippocampus can impair face learning (Crane & Milner, 2002). Based on this prior evidence, we included the hippocampus in our study. Extensive prior evidence has demonstrated that face processing is right lateralised (Kanwisher et al., 1997; Landis et al., 1986; Pitcher et al., 2007; Yovel et al., 2003) so we focused our analysis on regions of interest (ROIs) in the right hemisphere. Face-selective areas (e.g., FFA, OFA, pSTS, and amygdala) were identified in each participant using an independent localiser and the right hippocampus was identified using a subcortical structural atlas (Desikan et al., 2006). To support real-world learning in our design, we used ourselves (the experimenters) as the actors for the social interactions. To achieve this, four of the experimenters became the experimental stimuli. The experimenters (henceforth called actors) each identified an individual whose face resembled their own, to act as their control. We then collected forty photographs of each actor and each foil from their own personal collection. This was done to give a wide sampling of the lighting, picture quality, pose, expression, angle, and context that can vary when learning a new face in the real world. Different sets of photographs were used in the fMRI and behavioural sessions. Participants completed an initial behavioural session during which it was confirmed that they did not know, and had never seen any of the actors or foils. Then, participants returned on a subsequent day and were scanned with fMRI while viewing photographs of actors and foils. Immediately after this scan, they interacted with the actors for 10 min. Then on subsequent days, participants repeated a second behavioural testing session, and then a second fMRI scan. Prior to each of these sessions, they again interacted with the actors for 10 min only.

Materials and Methods

Participants

Twenty-five volunteers participated in this study. Data from two participants were excluded from all analyses after we realised it was possible that they may have seen or met one of the actors before the study commenced. In addition, one participant was excluded from the main analysis because we were unable to identify the right OFA from their face localiser results. The remaining 22 participants (14 women and 8 men; age range: 18–35 years, mean age: 21 and SD = 4) were right-handed, neurologically healthy with normal or corrected-to-normal vision. Each participant provided informed consent and was paid for their time. The study was approved by the York Neuroimaging Centre (YNiC) Research Ethics Committee at the University of York.

Stimulus Sets

Behavioural and Main fMRI Tasks

A set of 320 photographs was used. The photographs presented faces of eight identities (four actors and four foils). Four of the authors (L.S., M.E., D.O’G.. and G.P.) were selected as the actors for this experiment. Each actor was asked to identify their own foil, an individual whose face resembled the actor's face. Actors were invited to consult their friends to find an individual they resembled. Similarity to the actor was confirmed by inspection of all authors. None of the foils were related to the actors. There were 40 photographs per identity. Actors selected their own photographs from their pre-existing personal collection. Photographs of three foils were taken from the social media (e.g., Facebook or Instagram) after obtaining permission from the foils for using their photographs in the experiment and in published figures (see Figure 1, e.g., photographs). One foil was an Australian celebrity who is not well known in the UK. Her photographs were taken from the Google Images. Photographs were “ambient images” (Burton et al., 2011; Sutherland et al., 2013), deliberately incorporating the range typically encountered as we recognise faces. Photographs showed faces expressing mostly positive or neutral emotions. The individuals also wore different accessories (e.g., make-up, jewellery, headwear, and glasses) and hairstyles across the photographs. Faces were shown from different angles as long as more than 75% of the face was visible to make the face recognition possible. Faces were also at different distances from the camera and were presented in grayscale.
Figure 1.

Examples of the photographs used in the study. The four photographs on the left show one of the actors. The four photographs on the right show the foil for that actor. Note that distinguishing these two people is a difficult task for unfamiliar viewers (Hancock et al., 2000; Young & Burton, 2017).

Examples of the photographs used in the study. The four photographs on the left show one of the actors. The four photographs on the right show the foil for that actor. Note that distinguishing these two people is a difficult task for unfamiliar viewers (Hancock et al., 2000; Young & Burton, 2017). To create behavioural matching tests, a subset of 80 photographs, 10 photographs per identity, were used. Each photograph was presented twice, once it was present in the “same” trial and once in the “different” trial. The same set of 80 photographs was used across the two blocks but they were always paired with a different stimulus. Performance on pairwise face-matching tests is used as a measure of degree of familiarity (Clutterbuck & Johnston, 2004; Kramer et al., 2018). The remaining 240 photographs were divided into six different stimuli sets, containing five photographs per identity, to create a total of six fMRI runs (three runs per scanning session). Each of the 240 photographs was presented only once across the scan runs. This was done to ensure that participants learnt the faces of the actors rather than the photographs themselves. The order of the six fMRI runs was randomised across subjects.

fMRI Functional Localisation

The stimuli consisted of 3-second movie clips of faces, bodies, scenes, objects, and scrambled objects. Movies of bodies, scenes, and scrambled objects were not relevant to this study hence their data are not presented. These stimuli have been successfully used for functional localisation of face-responsive areas in prior studies (Handwerker et al., 2020; Sliwinska, Bearpark, et al., 2020; Sliwinska, Elson, et al., 2020; Sliwinska & Pitcher, 2018). There were 60 movie clips for each category in which distinct exemplars appeared multiple times. Movies of faces and bodies were filmed on a black background and framed close-up to reveal only the faces or bodies of seven children as they danced or played with toys or adults (who were out of frame). Movies of scenes included 15 different locations which were mostly pastoral scenes filmed from a car window while driving slowly through leafy suburbs, along with some other films taken while flying through canyons or walking through tunnels that were included for variety. Movies of objects used 15 different moving objects that were selected in a way that minimises any suggestion of animacy of the object itself or of a hidden actor pushing the object. Those included mobiles, windup toys, toy planes and tractors, and balls rolling down sloped inclines. Movies of scrambled objects were constructed by dividing each object movie clip into a 15 × 15 box grid and spatially rearranging the location of each of the resulting movie frames. Within each block, stimuli were randomly selected from within the entire set for that stimulus category. This meant that the same movie clip could appear within the same block but given the number of stimuli this did not occur frequently.

Procedure

Each participant completed four testing sessions that took place on separate days (Figure 2). Sessions one and three were behavioural sessions and sessions two and four were fMRI scanning sessions. Each behavioural session lasted approximately 15 min and each fMRI session lasted for approximately 1 h. After the first behavioural session, an experimenter not performing as an actor (M.S.) ensured that the participant did not know and had never seen the actors or foils. If participants reported they recognised any of the actor or foil photographs, they were excluded from the study.
Figure 2.

The experimental timeline. Testing took place on four separate days. During session one, participants completed the first behavioural task. During session two, participants were scanned using functional magnetic resonance imaging (fMRI). When the scan was complete, participants interacted with actors for 10 min immediately after leaving the scanner. During session three, participants interacted with the actors for 10 min and then completed the second behavioural task. During session four, participants interacted with the actors for 10 min and then completed the second fMRI scan.

The experimental timeline. Testing took place on four separate days. During session one, participants completed the first behavioural task. During session two, participants were scanned using functional magnetic resonance imaging (fMRI). When the scan was complete, participants interacted with actors for 10 min immediately after leaving the scanner. During session three, participants interacted with the actors for 10 min and then completed the second behavioural task. During session four, participants interacted with the actors for 10 min and then completed the second fMRI scan. Social interactions lasted 10 min each and occurred immediately after session two (the first scan) and immediately prior to sessions three (the second behavioural test) and four (the second scan). The gap between sessions varied across participants due to logistical reasons but all participants completed all sessions within 2 weeks.

Real-World Social Interactions

All interactions included a planned set of activities involving each participant with all four actors. These were designed to resemble real-world interactions and forced participants to pay attention to the faces of the actors. At the first interaction, participants met the actors for the first time. The participant was greeted by the actors immediately after leaving the MRI scanner and taken by them to a separate room. For the first few minutes, the actors involved the participant in a general conversation. Then, each actor played the Rock Paper Scissors game 10 times with the participant maintaining the eye contact with the participant. The second interaction occurred prior to the second behavioural testing session. Actors met participants in the testing room and engaged them in a semi-structured conversation. Each actor was assigned five questions that were designed to elicit a short conversation (e.g., What qualities do you tend to look for in a friend?; What movies have you seen recently?; What would your dream job be?). After the conversation, each actor played Rock Paper Scissors 10 times with the participant. Actors then reminded participants about the instructions for the behavioural task and left the room. All testing occurred in the absence of the actors. After the completion of the task actors re-entered the testing room and finished the interaction with a very brief conversation. The third interaction occurred prior to the second fMRI scan. Actors engaged the participant in a short general conversation, explained the procedures of the second fMRI session, and went through the MRI safety questionnaire with the participant. Each actor always asked the same set of questions. Participants were then placed in the MRI scanner by operators (not the actors).

Behavioural Sessions

Participants performed a computer-based face-matching task in which they judged whether two photographs, appearing simultaneously side-by-side on the screen, presented the same identity or two different identities (Figure 3). Participants used their right or middle finger to respond “yes” or “no”, respectively, by pressing appropriate keys on the keyboard. The response time was unlimited, but participants were asked to provide their answers as quickly and accurately as possible. No performance feedback was provided.
Figure 3.

The timeline of the behavioural task procedure (left). Participants had to judge whether the two simultaneously presented photographs depicted the same identity or not. The presented trials represent “no”, “yes”, “no”, and “yes” trials. The timeline of the functional magnetic resonance imaging (fMRI) task procedure (right). Participants were scanned while photographs of actors or foils were presented for 3 s each. Participants were not given any instructions about the photographs. To ensure they pay attention to the stimuli, they were asked to press a response button every time the black fixation cross turned red (this occurred 30% of the time).

The timeline of the behavioural task procedure (left). Participants had to judge whether the two simultaneously presented photographs depicted the same identity or not. The presented trials represent “no”, “yes”, “no”, and “yes” trials. The timeline of the functional magnetic resonance imaging (fMRI) task procedure (right). Participants were scanned while photographs of actors or foils were presented for 3 s each. Participants were not given any instructions about the photographs. To ensure they pay attention to the stimuli, they were asked to press a response button every time the black fixation cross turned red (this occurred 30% of the time). The face-matching task was written using E-Prime v2.0 (Psychology Software Tools, Inc.) Each trial began with a fixation cross displayed for 500 msec and followed by the stimuli presentation that was shown until the participant responded. The response triggered the next trial. Each stimuli pair was presented on a white background. Stimuli were displayed on the 22-inch CRT monitor set at 1024 × 768 resolution and refresh rate of 85 Hz. The first behavioural session was led by an experimenter who was not one of the actors, while the second behavioural session was led by all the actors (see Real-world Social Interactions for more information). During each session, participants completed one block of the face-matching task. Each block consisted of 80 trials, with three different face-matching conditions. These were actor-actor trials (20 trials), foil-foil trials (20 trials), and actor-foil trials (40 trials). In the actor-actor and foil-foil trials, there were five trials per identity. The actor-foil trials presented a picture of an actor paired with a picture of her/his foil (10 trials per an actor). The first two conditions (actor-actor, foil-foil) constituted “same” trials while the third condition (actor-foil) constituted “different” trials. Half of the trials in the third condition had a photograph of the actor on the left side of the screen and a photograph of the foil on the right side of the screen while the reverse order of the presentation sides was applied in the second half of the trials. This was done to avoid learning the identities based on a particular side of the screen. The same stimuli pairings were presented to all the participants but the trial order was randomised across participants.

fMRI Sessions

In fMRI sessions, participants completed three functional runs of the main experimental tasks, followed by two runs of a face localiser task designed to identify face-selective ROIs. During the first fMRI session a T1-weighted structural brain scan was also collected to anatomically localise the functional data for each participant. In addition, during both sessions, a T1-fluid-attenuated inversion recovery (FLAIR) scan was acquired to improve co-registration of the functional and structural scans across the two scanning sessions. The main fMRI task was written using PsychoPy v12.0, an open-source software package. Each trial lasted 10 s and it commenced with a grey blank screen displayed for 6.75 s, followed by the black or red fixation cross displayed for 0.25 s, and then by a stimulus presentation for 3 s Each stimulus and fixation cross were presented against a grey background. Stimuli were presented using a ProPixx LED projector (VPixx Technologies, Quebec, Canada) set at 1920 × 1080 resolution with a refresh rate of 120 Hz. At the beginning and end of each run, there was a 10 s rest period. Each run lasted 7 min. Stimuli were presented in a slow event-related design with relatively long interstimulus intervals to avoid the overlaps of hemodynamic response function (HRF) across trials. In the main task, participants passively viewed photographs of the faces of actors and foils that were presented sequentially at the centre of the screen (5 photographs of each identity per run, 15 in total per scan). Each run contained 40 different photographs (5 photographs per identity). Each scan session contained three runs of the task. Run order was randomised across participants and across the fMRI sessions. Participants were instructed to pay attention to the screen but were not given any specific instructions regarding the faces. To ensure participants attended the screen, they were asked to press a response button when the fixation cross preceding each photograph was red instead of black. The red fixation cross occurred randomly in 3 out of 10 trials. Face-selective ROIs were identified using a dynamic face localiser task (Pitcher et al., 2011). Data were acquired using block-design runs, lasting 234 s each. During those runs, participants were instructed to watch videos of faces, bodies, scenes, objects, or scrambled objects, without performing any overt task. Each run contained two sets of five consecutive stimulus blocks to form two blocks per stimulus category per run. Each block lasted 18 s and contained stimuli from one of the five stimulus categories. Each functional run also contained three 18 s rest blocks, which occurred at the beginning, middle, and end of the run. During the rest blocks, a series of six uniform colour fields were presented for 3 s each. The order of stimulus category blocks in each run was palindromic (e.g., rest, faces, objects, scenes, bodies, scrambled objects, rest, scrambled objects, bodies, scenes, objects, faces, and rest) and randomised across runs.

Brain Imaging Acquisition and Analysis

Imaging data were acquired using a 3T Siemens Magnetom Prisma MRI scanner (Siemens Healthcare, Erlangen, Germany) at YNiC utilising a 20-channel phased array head coil tuned to 123.3 MHz. Functional images for the main and localisation tasks were recorded using a gradient-echo EPI sequence (35 interleaved slices, repetition time [TR] = 2,000 msec, echo time [TE] = 30 msec, flip angle = 80°; voxel size = 3 × 3 × 3 mm; matrix size = 64 × 64; field of view [FOV] = 192 × 192 mm) providing whole brain coverage. Slices were aligned with the anterior to posterior commissure line. Structural images were acquired using a high-resolution T-1-weighted 3D fast spoilt gradient (SPGR) sequence (176 interleaved slicesTR = 2300 msec, TE = 2.26 msec, flip angle = 8°; voxel size = 1 × 1 × 1 mm; matrix size = 256 × 256; FOV = 256 × 256). T1-FLAIR scans (35 interleaved slices, TR = 3000 msec, TE = 8.6 msec, flip angle = 150°; voxel size = 0.8 × 0.8 × 3.0 mm; matrix size = 256 × 256; FOV = 192 × 192) were acquired with the same orientation as the functional scans. Functional MRI data were analysed using fMRI Expert Analysis Tool (FEAT) included in the FMRIB (v6.0) Software Library (www.fmrib.ox.ac.uk/fsl). Data from the first four TRs from each run were discarded. The remaining images were slice time corrected and realigned to the first volume of each functional run and to the corresponding anatomical scan. The volume-registered data were spatially smoothed with a 5-mm full-width-half-maximum Gaussian kernel. Signal intensity was normalised to the mean signal value within each run and multiplied by 100 so that the data represented percent signal change from the mean signal value before analysis. The data from the main experimental task was entered into a general linear model (GLM) by convolving the standard HRF with the regressors of interest. These were either actors (data from all four actors was averaged together) or foils (data from all four foils was averaged together). This created four conditions, namely actor pre-interaction, actor post-interaction, foil pre-interaction, foil post-interaction. The model was convolved using a double-gamma HRF to generate the main regressors. Temporal derivatives for each condition were included. First-level functional results for each participant were registered to their anatomical scan and then to the Montreal Neurological Institute (MNI) 152-mean brain using a 12 degree-of-freedom affine registration. The T1-FLAIR scan was added as an expanded functional image to help aid registration. ROIs were identified for each participant using a contrast of greater activation evoked by dynamic faces than that evoked by dynamic objects, calculating significance maps of the brain using an uncorrected statistical threshold (p = .01). We identified four face-selective areas using anatomical landmarks, these were; the FFA, OFA, pSTS, and face-selective voxels in the amygdala. We were able to identify all four ROIs in the right hemisphere of 22 participants. In the left hemisphere, the ROIs were not identified across all participants (left FFA was present in 17 participants, left OFA in 19 participants, pSTS in 15 participants, and the left amygdala in 11 participants). For this reason, we focused our fMRI data analysis on the right hemisphere ROIs only as we have in prior fMRI studies of the face network using these stimuli (Pitcher et al., 2019; Pitcher et al., 2017). The peak voxel of activation was identified for each area and a 5 mm sphere was individually drawn around this point for each participant. The mean of the peak MNI coordinates for the ROIs across participants were right FFA—40, −52, −20, right OFA—41, −79, −15, right pSTS—52, −38, 3, right amygdala—19, −5, −15. A table of the peak MNI coordinates for all ROIs across all participants is shown in online Supplemental Materials. The right hippocampus was identified using the FSL mask from Harvard-Oxford Cortical and Subcortical Structural Atlas (Desikan et al., 2006). A point in the middle of the mask was located and a 5 mm sphere was individually drawn around this point for each participant. Within each functionally defined ROI, we then calculated the magnitude of response (percent signal change from a fixation baseline) for the four experimental conditions (e.g., actor pre-interaction, actor post-interaction, foil pre-interaction, and foil post-interaction). We also conducted a whole-brain group analysis using a two-way mixed model analysis of variance (ANOVA) (session by actor/foil) for all 22 participants. This analysis did not yield any activation that passed an appropriate level for a statistically significant threshold. However, at the request of the reviewers, we have included the subthreshold clusters we observed in online Supplemental Materials for information only.

Results

Behavioural Task

Behavioural results clearly demonstrated that the real-world social interactions were sufficient to learn the faces of the previously unfamiliar actors (Figure 4). In post-interaction behavioural test, accuracy for the actor-actor trials matching condition increased, while accuracy for the foil-foil trials condition was unchanged. Performance accuracy for the actor-foil trials was also not improved by social interactions. This relatively brief encounter with a live actor appears to improve subsequent recognition (as measured in a matching test) but this improvement is evident only in the actor-actor matching trials. This is consistent with studies in which familiarity effects are carried by improvements in viewers’ ability to cohere different images of the same face, or “telling faces together” (Jenkins & Burton, 2011; Ritchie & Burton, 2017) and recent theoretical developments in face learning offer potential mechanisms for this process (Kramer et al., 2018) but see (Blauch et al., 2021) for an alternate view.
Figure 4.

Mean accuracy data for the behavioural face-matching task before and after social interactions had occurred. Results showed that social interactions improved performance accuracy for the actor/actor trials only. There was a significant two-way interaction between session and trial type (p < .0001). Error bars show SE of the mean across participants. *Denotes a significant difference (p < .001).

Mean accuracy data for the behavioural face-matching task before and after social interactions had occurred. Results showed that social interactions improved performance accuracy for the actor/actor trials only. There was a significant two-way interaction between session and trial type (p < .0001). Error bars show SE of the mean across participants. *Denotes a significant difference (p < .001). Accuracy data were entered in a two (session: pre-interaction, post-interaction) by three (trial type: actor-actor, foil-foil, actor-foil) repeated measures ANOVA. Results showed a significant main effect of trial type (F(2, 42) = 23; p < .0001; partial η2 = 0.5) but not of session (F(1, 21) = 3.8; p = .065; partial η2 = 0.16). Crucially, there was a significant two-way interaction (F(2, 42) = 9.25; p < .0001; partial η2 = 0.30). Consistent with our hypothesis, we found a significant two-way interaction between session and trial type demonstrating that social interactions improved performance accuracy when matching actor-actor photographs. To understand what factors were driving this effect, we calculated the simple main effects. Results showed a significant difference before and after the interactions for the actor-actor trials (F(1,21) = 19, p < .001; partial η2 = 0.24) but not the foil-foil trials (F(1,21) = 0.1, p = .9; partial η2 = 0.00) or the actor-foil trials (F(1,21) = 0.1, p = .7; partial η2 = 0.00). Tukey's honestly significant difference (HSD) (honestly significant difference) tests showed that there was no significant difference between the actor-actor and foil-foil trials pre-interaction (p = .28) but there was a significant difference post-interaction (p < .0001). There were also significant differences between actor-actor and actor-foil trials pre-interaction (p = .002) and post-interaction (p < .0001). The same pattern was also apparent for foil-foil trials and actor-foil trials both pre-interaction (p = .04) and post-interaction (p = 0.014). We also analysed the reaction time (RT) data (Figure 5) in a two (session: pre-interaction, post-interaction) by three (trial type: actor-actor, foil-foil, actor-foil) repeated measures ANOVA. While they showed the same pattern as the accuracy data the interaction did not reach significance. The main effects of session (F(1, 21) = 14.48; p = .001; partial η2 = 0.40) and trial type (F(2, 42) = 7.86; p = .001; partial η2 = 0.26) were significant. The two-way interaction between these two conditions was not significant (F(2, 42) = 3.05; p = .06; partial η2 = 0.12).
Figure 5.

Mean accuracy data for the behavioural face-matching task before and after social interactions had occurred.

Mean accuracy data for the behavioural face-matching task before and after social interactions had occurred.

fMRI Results

The neural response in the right FFA, right OFA, right pSTS, right amygdala, and right hippocampus to actor photographs and foil photographs only differed in scan two, after the real-world social interactions had taken place (Figure 6). The pattern across all ROIs was similar, namely the response to foil photographs was lower in scan two (after social interactions had occurred) than in scan one. In contrast, the response to the actor photographs was unchanged between scan one and scan two. These results demonstrate that the learning effect we observed for the actor photographs in the behavioural task (Figure 4) was matched by a sustained neural response. The lack of a learning effect for the foil faces was matched with a weaker neural response. This result is consistent with fMRI studies of learning in humans and non-human primates reporting a drop in the neural response to the unlearned stimulus (Kaskan et al., 2017; Op de Beeck et al., 2006).
Figure 6.

Percent signal change data for the actor and foil photographs before and after social interaction in the face-selective areas and the right hippocampus. We observed a significant two-way interaction (p = .012; partial η2 = 0.26) between photograph type (actor and foil) and session (pre-interaction and post-interaction). This was driven by a reduction in the neural signal to foil photographs across all regions of interest (ROIs) in scan two after the real-world social interactions had occurred.

Percent signal change data for the actor and foil photographs before and after social interaction in the face-selective areas and the right hippocampus. We observed a significant two-way interaction (p = .012; partial η2 = 0.26) between photograph type (actor and foil) and session (pre-interaction and post-interaction). This was driven by a reduction in the neural signal to foil photographs across all regions of interest (ROIs) in scan two after the real-world social interactions had occurred. Face-selective ROIs were identified with data from the independent functional localiser using contrast of face movies greater than object movies. We were able to localise the four face-selective ROIs in 22 of the 23 participants (one participant did not have the right OFA). The hippocampus was identified using the Harvard-Oxford Cortical and Subcortical Structural Atlas. Percent signal change data were entered into a five (ROI: rOFA, rFFA, rpSTS, right amygdala, and right hippocampus) by two (photograph: actor, foil) by two (session: pre-interaction, post-interaction) repeated measures ANOVA. We found main effects of ROI (F(1,21) = 58, p < .0001; partial η2 = 0.73) but not of photograph (F(1,21) = 1.5, p = .23; partial η2 = 0.07) or session (F(1,21) = 1.6, p = .22; partial η2 = 0.07). There was no significant two-way interaction between ROI and photograph (F(4,84) = 0.5, p = .72; partial η2 = 0.02), or between ROI and session (F(4,84) = 0.4, p = .8; partial η2 = 0.02), but crucially there was a significant two-way interaction between photograph and session (F(1,21) = 7.6, p = .012; partial η2 = 0.26). There was no significant three-way interaction (F(4,84) = 0.7, p = .68; partial η2 = 0.03). Consistent with our hypothesis, we found a significant two-way interaction between photograph and session demonstrating that social interactions differentially affected the neural response to actor and foil photographs. To understand what factors were driving this effect, we calculated the simple main effects between the four conditions across all ROIs: actor pre-interaction (0.56, SE = 0.05), foil pre-interaction (0.55, SE = 0.04), actor post-interaction (0.57, SE = 0.05) and foil post-interaction (0.51, SE = 0.04). There was no significant difference between actor and foil photographs pre-interaction (F(1,42) = 0.8, p > .5; partial η2 = 0.02) but there was a significant difference post-interaction (F(1,42) = 7.5, p = .009; partial η2 = 0.15). In addition, there was no significant difference between pre-interaction actor photographs and post-interaction actor photographs (F(1,42) = 0.05, p > .5; partial η2 = 0.03) but there was a significant difference between pre-interaction foil photographs and post-interaction photographs (F(1,42) = 4.7, p = .04; partial η2 = 0.03).

Discussion

In the current study, we investigated the behavioural and neural basis of face learning using real-world social interactions. Participants were scanned on two separate days while viewing photographs of four actors or each actor's foil (the control stimuli). Between these scanning sessions, the participants met and interacted with the actors for 10 min on three separate days. In addition to the scanning sessions, participants completed a behavioural face-matching task before and after two of the real-world interactions. Behavioural results showed participants were significantly better at matching actor photographs than foil photographs but only after the social interactions had occurred (Figure 4). ROI analysis of the neuroimaging data focused on the FFA, OFA, pSTS, amygdala, and hippocampus in the right hemisphere. There was no difference between the neural response to actor and foil photographs in all ROIs in the first scan, prior to the social interactions. In contrast, the response to foil photographs was significantly lower than the response to actor photographs during the second scan (Figure 5). These results demonstrate that short real-world social interactions are sufficient to learn the faces of previously unfamiliar individuals, and that this learning process can be detected in the nodes of the face processing network and in the hippocampus. Prior behavioural studies have highlighted numerous differences between familiar and unfamiliar faces. It has been known for many years that memory for unfamiliar faces is severely affected by small changes in the image, for example, due to lighting, expression, or pose, whereas such changes barely affect familiar face recognition (Bruce, 1986; Ellis et al., 1979; Klatzky & Forrest, 1984). Simultaneous face matching is also difficult for unfamiliar faces, but trivially easy for familiar faces (Bruce et al., 1999; Burton et al., 1999). This has led some authors to argue that unfamiliar face processing (for identity) is primarily image-based (Hancock et al., 2000), whereas familiar face recognition is robust to variation because it relies on computation of the within-identity variability specific to each individual (Burton, 2013 ; Jenkins et al., 2011). In the current study, we sought to incorporate within-individual variability by collecting forty photographs of each actor and their foil for use in the behavioural and fMRI tasks. These photos were taken with different cameras, from different angles, and in different lighting conditions. Despite having a range of different variations in both the actor and foil photographs we only observed a difference between the actor and foil conditions after the social interactions had occurred. When meeting a previously unfamiliar individual, we are visually exposed to their face while we form opinions and gather semantic information from them (e.g., their name, where they are from, etc.). This wealth of information is used and encoded when learning to recognise their face. In the current study, we structured the social interactions (e.g., the actors asked pre-planned questions and played short games with the participants) but there was no attempt to specifically guide the face learning. In fact, we did not even instruct the participants that the aim of the study was to learn the faces of the actors. Instead, our intention was to incorporate natural, encounter-based learning into the design, rather than use the typical lab-based controlled exposure. Perhaps surprisingly, rather little is known about the process of face learning, but it seems likely that many different factors will have contributed to the learning effect we observed after the social interactions had taken place. Prior studies suggest that these factors could include semantic information (Heisz & Shedden, 2009), multisensory information (von Kriegstein & Giraud, 2006; von Kriegstein et al., 2005), and dynamic information (Kaufmann et al., 2009) have on behavioural performance accuracy. We did not control for any of these factors in our experimental design and it seems likely that all will have contributed to our results to some degree. However, most previous research has used a lab environment rather than in real-world social interactions. One of the challenges for future work will be to preserve naturalistic learning while also retaining the ability to manipulate potential contributions to learning in a systematic way. The ROI analysis revealed a significant difference between actor and foil faces only after social interactions had occurred (Figure 5). This is consistent with prior neuroimaging studies that reported a higher response to familiar than unfamiliar faces in the FFA (Elfgren et al., 2006; Gobbini et al., 2004; Pierce et al., 2004; Sergent et al., 1992; Weibert & Andrews, 2015) as well as in other face-selective areas (Ishai et al., 2005; Ishai & Yago, 2006). The present study differed from these prior studies in that we scanned participants before and after real-world face learning had occurred during a brief face-to-face conversation. While there was no significant main effect of social interaction, it is worth noting that the response to actor faces in scan two was not significantly greater than the response to actor faces in scan one. This decrease in the response to the unlearned stimulus is consistent with fMRI studies of learning in humans and non-human primates reporting a drop in the neural response to the unlearned stimulus (Kaskan et al., 2017; Op de Beeck et al., 2006). It is also possible that the reduction in the BOLD response to the foil faces in scan two was due to repetition suppression (also called adaptation) (Grill-Spector et al., 2006). The increased neural activation we observed after the real-life interactions with the actors may have reduced or eliminated the priming effect usually observed with repeated exposure to the same stimuli. This would be consistent with evidence demonstrating that that learning may bias repetition suppression effects towards repetition enhancement (Segaert et al., 2013). Finally, another possible account is the general attention of participants which may have been lower overall during the second session. However, it is important to note that neither of these factors can account for the difference in the response to actor and foil faces that we observed after the social interactions had taken place. The behavioural results showed the same pattern as the fMRI results, namely we observed a difference in matching actors and foils only after the social interactions had occurred (Figure 4). There were only two 10-minute interactions between the behavioural testing sessions demonstrating that 20 min were sufficient to learn the faces of the actors. This demonstrates the effectiveness and utility of these real-world interactions in the study of how face learning and person recognition occur. Interestingly, we did not observe a significant effect of social interaction for the different trials. In fact, across the literature on face matching, there is inconsistency about whether effects of familiarity or expertise are driven by “hits” (same person matching performance) or “correct rejections” (different person matching performance) (Bobak et al., 2019; Matthews & Mondloch, 2018; Ritchie et al., 2021). While studies of learning typically show improved accuracy with increased exposure (and some only report this), a breakdown of these effects shows they are typically driven by only one component of performance and it has (so far) been impossible to isolate a satisfactory explanation for this. In the present study, improvement is driven by performance in same-item trials, and we suggest that this is largely consistent with previous literature using more naturalistic exposure to faces. Further, our results are consistent with studies demonstrating that telling people apart is not the same process as telling them together (Jenkins et al., 2011). These studies employed a sorting paradigm to demonstrate that unfamiliar face recognition is less tolerant of within-person variation than familiar face recognition. Our design was clearly robust enough for participants to learn the actor’s face but not the foil face, leading to an increased level of uncertainty when matching actor-foil photographs, alongside an improvement in the ability to match different instances of the learned faces. Future studies could test this hypothesis in a two-stage real-world design in which participants are exposed to one group of actors and then the second group of actors who physically match the first group.

Conclusion

Our study has demonstrated that brief real-world social interactions are sufficient to learn the faces of previously unfamiliar individuals. Importantly, we were also able to detect the difference between learned and unlearned faces in the nodes of the face processing network and in the hippocampus. Finally, given the complexity of the social environments that we navigate on a daily basis, we propose that such real-world interactions can be adopted in future neuroimaging studies. Click here for additional data file. Supplemental material, sj-docx-1-pec-10.1177_03010066221098728 for Face learning via brief real-world social interactions induces changes in face-selective brain areas and hippocampus by Magdalena W. Sliwinska, Lydia R. Searle, Megan Earl, Daniel O’Gorman, Giusi Pollicina, A. Mike Burton and David Pitcher in Perception Click here for additional data file. Supplemental material, sj-xlsx-2-pec-10.1177_03010066221098728 for Face learning via brief real-world social interactions includes changes in face-selective brain areas and hippocampus by Magdalena W. Sliwinska, Lydia R. Searle, Megan Earl, Daniel O’Gorman, Giusi Pollicina, A. Mike Burton and David Pitcher in Perception
  76 in total

1.  Event-related brain potential evidence for a response of inferior temporal cortex to familiar face repetitions.

Authors:  Stefan R Schweinberger; Esther C Pickering; Ines Jentzsch; A Mike Burton; Jürgen M Kaufmann
Journal:  Brain Res Cogn Brain Res       Date:  2002-11

Review 2.  Mental representations of familiar faces.

Authors:  A Mike Burton; Rob Jenkins; Stefan R Schweinberger
Journal:  Br J Psychol       Date:  2011-06-14

3.  Unfamiliar faces are not faces: evidence from a matching task.

Authors:  Ahmed M Megreya; A Mike Burton
Journal:  Mem Cognit       Date:  2006-06

4.  TMS evidence for the involvement of the right occipital face area in early face processing.

Authors:  David Pitcher; Vincent Walsh; Galit Yovel; Bradley Duchaine
Journal:  Curr Biol       Date:  2007-08-30       Impact factor: 10.834

5.  Finding an unfamiliar face in a line-up: Viewing multiple images of the target is beneficial on target-present trials but costly on target-absent trials.

Authors:  Claire M Matthews; Catherine J Mondloch
Journal:  Br J Psychol       Date:  2018-04-16

6.  Learned Value Shapes Responses to Objects in Frontal and Ventral Stream Networks in Macaque Monkeys.

Authors:  Peter M Kaskan; Vincent D Costa; Hana P Eaton; Julie A Zemskova; Andrew R Mitz; David A Leopold; Leslie G Ungerleider; Elisabeth A Murray
Journal:  Cereb Cortex       Date:  2017-05-01       Impact factor: 5.357

7.  Effects of lighting on the perception of facial surfaces.

Authors:  H Hill; V Bruce
Journal:  J Exp Psychol Hum Percept Perform       Date:  1996-08       Impact factor: 3.332

Review 8.  The suppression of repetition enhancement: a review of fMRI studies.

Authors:  Katrien Segaert; Kirsten Weber; Floris P de Lange; Karl Magnus Petersson; Peter Hagoort
Journal:  Neuropsychologia       Date:  2012-11-14       Impact factor: 3.139

9.  Representational demands modulate involvement of perirhinal cortex in face processing.

Authors:  Edward B O'Neil; Victoria A Barkley; Stefan Köhler
Journal:  Hippocampus       Date:  2013-04-12       Impact factor: 3.899

10.  Influences of familiarity on the processing of faces.

Authors:  V Bruce
Journal:  Perception       Date:  1986       Impact factor: 1.490

View more
  1 in total

1.  Measuring the response to visually presented faces in the human lateral prefrontal cortex.

Authors:  Lara Nikel; Magdalena W Sliwinska; Emel Kucuk; Leslie G Ungerleider; David Pitcher
Journal:  Cereb Cortex Commun       Date:  2022-08-18
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.