Olesya Blazhenkova1. 1. Faculty of Arts and Social Sciences, Sabancı University, Istanbul, Turkey.
Abstract
Boundary extension is a common false memory error, in which people confidently remember seeing a wider angle view of the scene than was viewed. Previous research found that boundary extension is scene-specific and did not examine this phenomenon in nonscenes. The present research explored boundary extension in cropped face images. Participants completed either a short-term or a long-term condition of the task. During the encoding, they observed photographs of faces, cropped either in a forehead or in a chin area, and subsequently performed face recognition through a forced-choice selection. The recognition options represented different degrees of boundary extension and boundary restriction errors. Eye-tracking and performance data were collected. The results demonstrated boundary extension in both memory conditions. Furthermore, previous literature reported the asymmetry in amounts of expansion at different sides of an image. The present work provides the evidence of asymmetry in boundary extension. In the short-term condition, boundary extension errors were more pronounced for forehead, than for chin face areas. Finally, this research examined the relationships between the measures of boundary extension, imagery, and emotion. The results suggest that individual differences in emotional ability and object, but not spatial, imagery could be associated with boundary extension in face processing.
Boundary extension is a common false memory error, in which people confidently remember seeing a wider angle view of the scene than was viewed. Previous research found that boundary extension is scene-specific and did not examine this phenomenon in nonscenes. The present research explored boundary extension in cropped face images. Participants completed either a short-term or a long-term condition of the task. During the encoding, they observed photographs of faces, cropped either in a forehead or in a chin area, and subsequently performed face recognition through a forced-choice selection. The recognition options represented different degrees of boundary extension and boundary restriction errors. Eye-tracking and performance data were collected. The results demonstrated boundary extension in both memory conditions. Furthermore, previous literature reported the asymmetry in amounts of expansion at different sides of an image. The present work provides the evidence of asymmetry in boundary extension. In the short-term condition, boundary extension errors were more pronounced for forehead, than for chin face areas. Finally, this research examined the relationships between the measures of boundary extension, imagery, and emotion. The results suggest that individual differences in emotional ability and object, but not spatial, imagery could be associated with boundary extension in face processing.
Boundary extension (BE) is a common false memory error, in which people
confidently remember seeing a wider angle view of the scene than was actually viewed. They
recall the surrounding regions of the scene, which were not visible during the encoding. BE
is a constrained phenomenon, and scene extrapolation occurs just beyond the edges of a view
(Gottesman & Intraub, 1999;
Intraub & Richardson,
1989). This phenomenon is thought to comprise a two-stage process. First, a
spontaneous extrapolation beyond the visible boundaries occurs during the initial encoding
of an encountered scene. This happens due to a constructive nature of perception. It was
suggested that the limited view of a scene activates a perceptual schema or a constructed
visual-spatial representation of the expected layout outside the view (Chadwick, Mullally, & Maguire, 2013; Intraub, Bender, & Mangels, 1992;
Intraub, Gottesman, & Bills,
1998). Subsequently, during the retrieval, BE manifests as a false memory error. In
contrast to the traditional model of scene perception, which postulates that visual input is
a single source of information in scene representation, the multisource model of scene
perception proposes that scene perception involves processing in multiple modalities, one of
which is vision (Intraub, 2010,
2012). According to the
multisource model, scene perception is based on spatial egocentric framework. Multiple
sources of input (e.g., visual sensory, amodal, conceptual, and contextual) ‘fill-in’ this
framework with expectations about the surrounding visual field, which is only partially
revealed in a picture. Thus, mental scene extrapolation does not happen after the stimulus
is gone but already occurs during the scene presentation. In this view, BE is a source
monitoring error (Johnson, Hashtroudi,
& Lindsay, 1993) that arises due to the confusion between original sensory
input and internally generated detail (Seamon, Schlegel, Hiester, Landau, & Blumenthal, 2002) and the attempts to
determine which portion of the scene matches to the visual source. BE is considered to be a
highly adaptive process that contributes to the representation of continuity and coherence
of the world that exists beyond the limited visual input (Chadwick et al., 2013; Mullally, Intraub, & Maguire, 2012).This phenomenon is remarkably robust. A variety of different methods, such as drawing from
memory, recognition/rating of images that contain more or less of the original scene, or
border-adjustment, consistently demonstrated BE effect (Hubbard, Hutchison, & Courtney, 2010; Intraub & Bodamer, 1993; Intraub, Hoffman, Wetherhold, & Stoehs,
2006). It was found in people of various ages, including 3- to 7-month infants and
adults up to 84 years of age (Quinn
& Intraub, 2007; Seamon
et al., 2002). BE was observed for a wide range of target durations (250 ms to
15 s, Gottesman & Intraub,
2002; Intraub, Gottesman,
Willey, & Zuk, 1996) and retention intervals (42 ms to 8 days, Intraub & Dickinson, 2008; Intraub & Richardson, 1989; Safer, Christianson, Autry, &
Österlund, 1998). It is not limited to rectangular images but occurs for different
shapes such as circular or irregular (Daniels & Intraub, 2006). BE is so robust that it is not eliminated even by
the awareness of this phenomenon, though, test-informed instruction may lead to a reduction
in BE (Gagnier, Dickinson, &
Intraub, 2013; Intraub &
Bodamer, 1993).However, despite its robustness, BE does not occur for all kinds of image contents. It was
found primarily for images containing a background surface, but not for images with
depictions on blank backgrounds (Intraub
et al., 1998). This was explained by the role of mental schema anticipating the
continuous layout beyond the frame and not activating for images that do not depict partial
view of a scene (Intraub et al.,
1998). The hypothesis that BE occurs due to object completion was ruled out because
this phenomenon was observed even for scenes that did not contain incomplete cropped objects
(Intraub et al., 1992; Intraub & Bodamer, 1993).
Besides, BE did not differ for scenes containing whole objects and cropped objects (Gagnier et al., 2013). BE was
suggested to occur only for limited boundaries of the view, but not in response to isolated
objects removed from the context background (Gottesman & Intraub, 2002, 2003). Though, there is contrary evidence that BE may
occur for abstract scenes containing figures on a white blank background (McDunn, Siddiqui, & Brown, 2014).
BE was suggested to be specific to representation of scenes or images that convey scene
structure (Intraub et al., 1998).
A scene is generally defined as a continuous spatial layout, which extends beyond the
visible boundaries, while a nonscene refers to an isolated object, which is not embedded in
a spatial context (Hubbard et al.,
2010). More generally, a scene is defined as a depiction of a truncated view of a
continuous world (Gottesman &
Intraub, 2002). In this sense, cropped faces may be perceived, similarly to scenes,
as truncated views of a full body extending beyond the edges of the perceived view (Intraub & Richardson, 1989).Indeed, BE phenomenon was observed for a variety scene contents, including scenes with
single or multiple objects, scenes with cropped and noncropped objects, and even abstract
representations (Intraub & Bodamer,
1993; Intraub & Richardson,
1989; McDunn et al.,
2014). However, the majority of studies did not focus on examination of BE in
nonscenes. Possibly, BE may occur in response to some categories of nonscene objects, and in
particular, to a very special kind of images, such as faces (Farah, Wilson, Drain, & Tanaka, 1998). Although
human bodies and faces were sometimes included in the studied scenes (Gallagher, Balas, Matheny, & Sinha, 2005; Intraub et al., 2006; Ménétrier, Didierjean, & Vieillard,
2013; Munger & Multhaup,
2016), no research specifically focused on BE in face representations. There is
some evidence that presence of humans and human faces may affect BE. For example, Gallagher et al. (2005) found an
intriguing interaction between the background complexity and the presence of a person. When
scenes contained a human, the amount of BE linearly increased with the complexity of the
background, whereas for scenes without a human, the simplest and the most complex scenes
yielded the highest amounts of BE. Overall, investigation of faces seems to be an overlooked
dimension in the research on BE. Thus, the main purpose of the present study was to extend
the existing knowledge on BE for nonscene images by examining this phenomenon in face
processing.Although a large body of evidence supports cognitive and neurological dissociation in
processing faces and scenes, there is also research
suggesting that processing of faces and scenes shares a lot of similarities. Functional
magnetic resonance imaging research (Haxby et al., 2001; Kanwisher,
McDermott, & Chun, 1997; McCarthy, Puce, Gore, & Allison, 1997) indicated that face processing is
underpinned by the inferior occipital gyrus and lateral fusiform gyrus (also known as
“Fusiform Face Area” or FFA). Lateral Occipital Cortex and FFA were found to have selective
response to images of faces, but not those of houses or places (Kanwisher et al., 1997; Levy, Hasson, Avidan, Hendler, & Malach, 2001).
In contrast, so-called Parahippocampal Place Area (PPA) was found to respond strongly to
depictions of places, including indoor and outdoor scenes, and to respond weaker to
buildings, but not to faces (Epstein
& Kanwisher, 1998). Previously published research indicated the crucial role of
hippocampus in scene viewing and processing spatial locations (Bird & Burgess, 2008). Neuropsychological
literature documented cases of hippocampal damage followed by a selective impairment of
scene recognition and intact face memory (Carlesimo, Fadda, Turriziani, Tomaiuolo, & Caltagirone,
2001; Taylor, Henson, &
Graham, 2007). In addition, it was found that face processing relied on a detailed
central scrutiny and was more strongly associated with processing of central information,
whereas representations of buildings or scenes involved more peripheral information (Kanwisher, 2001; Levy et al., 2001). Neuroimaging
research confirmed that BE is underpinned by the scene-selective regions of the brain.Chadwick et al. (2013) indicated
the crucial role of hippocampus in anticipation and construction of scenes, as well as in
extrapolation of scenes beyond their physical borders. Research emphasized the role of
hippocampus in construction of scenes both in memory and the imagination (Zeidman, Mullally, & Maguire,
2014). Mullally et al.
(2012) found that patients with selective bilateral damage to hippocampus
demonstrated less BE errors than control participants, supporting the role of hippocampus in
BE. Inconsistent with this finding, Kim,
Dede, Hopkins, and Squire (2015) found that amnestic patients with hippocampal
damage exhibited BE similarly to healthy controls. Furthermore, Park, Intraub, Yi, Widders, and Chun (2007)
demonstrated that BE task causes selective activation in the PPA, a region associated with
processing scenes such as landscapes or buildings (Epstein & Kanwisher, 1998), but not in the
Lateral Occipital Cortex, typically associated with object recognition (Grill-Spector et al., 1999).Although many studies found the dissociation between face and place processing (Epstein & Kanwisher, 1998; Kanwisher et al., 1997; Levy et al., 2001; Taylor et al., 2007), there is also
evidence that faces and places share neural underpinnings, for example, involvement of the
ventral temporal cortex (Haxby et al.,
2001). In addition, prosopagnosia (inability to recognize the faces) often
co-occurs with topographical prosopagnosia (Landis, Cummings, Benson, & Palmer, 1986).
Overall, given the earlier distinction between face and scene processing as well as the
evidence of BE underpinnings by scene-selective brain areas, it could be possible that BE
may not occur in response to faces. However, a cropped face image may elicit the sense of
continuation beyond the picture boundaries (Intraub et al., 1992), similar to how a close-up
portrait (disembodied head) may be interpreted as a continuous scene that includes the lower
parts of the body (Intraub &
Richardson, 1989). Possibly, a close-up face image may perform as a scene, and when
being cropped, it may enforce the perception of continuity and mental reconstruction of a
coherent representation in the same way as the truncated scene does. The present study aimed
to examine whether BE can be observed for cropped face images.
Particularly, to examine BE for faces rather than scenes, face images were presented on
black backgrounds.Furthermore, the present work examined whether there is an asymmetry of BE
in face processing. Although BE typically occurs for all boundaries of an image, for
example, four edges of a photograph (Intraub & Richardson, 1989), the asymmetry in BE, for example, different
amounts of expansion at different sides of an image, was demonstrated by previous studies.
In particular, BE was found to enlarge in the direction of the implied/anticipated motion,
but not in the opposite direction (Courtney & Hubbard, 2004). Furthermore, a greater BE was found for the objects
that typically move faster than for the objects that typically move slower (e.g., airplane
vs. automobile). These findings may be explained by the role of BE in facilitation of the
subsequent scene recognition (Dickinson
& Intraub, 2008). For instance, anticipating a movement may increase BE in the
expected direction and may aid spatial integration of the successive views (Hubbard et al., 2010). Similarly,
implied motion in frozen-motion pictures and abrupt disappearance of a moving target cause
displacement in memory toward the direction of motion (Freyd, 1983; Freyd & Finke, 1984; Futterweit & Beilin, 1994). Hubbard (1996) highlighted the similarity between BE
and representational momentum, that is, displacement in memory of a moving target beyond the
true final location. However, Munger,
Owens, and Conway (2005) showed that BE and representational momentum are separate
processes. They found that establishing the spatial layout occurs before the continuation of
movement within a scene and concluded that BE cannot be due to a displacement in depth.
Another evidence of asymmetry in BE comes from experiments using attentional cueing. As it
was demonstrated by Intraub et al.
(2006), BE can be affected by planned eye fixations: while BE appeared on the cued
(to-be-fixated) side of the image, it was inhibited on the uncued side. Inconsistent with
this finding, research demonstrated that focal and increased attention may constrain BE
error (Dickinson & Intraub, 2009; Intraub, Daniel, Horowitz, & Wolfe, 2008).
Overall, these findings implied that asymmetry in scene representation may be caused by the
anticipatory processing in a certain direction.The present work tested the hypothesis that asymmetry in BE may exist for different
facial parts. This prediction was made based on the evidence for asymmetry in
attention to different regions of a face. In particular, previous research demonstrated that
upper face regions attract more attention than lower face parts (James, Huh, & Kim, 2010; McKelvie, 1976). Goldstein and Mackenberg (1966) examined the
recognition of isolated portions of faces and demonstrated that upper portions of the face
are more important for identification than lower portions. Most probably, this asymmetry in
attention occurs because the eyes are the most salient and socially important part of the
face (Janik, Wellens, Goldberg, &
Dell’Osso, 1978). Our gaze naturally focuses on eyes and mouth face regions (Janik et al., 1978; Mertens, Siegmund, & Grusser,
1993; Walker-Smith, Gale,
& Findlay, 1977; Yarbus,
1967), which are crucial for face identification and memorization and play
important role in social communication (Ellis, 1975). Mouth region is also important but less salient than “eyes” region
(McKelvie, 1976; Pellicano, Rhodes, & Peters,
2006). Markedly, the consistent looking preference for the upper part versus lower
face part was not found for scenes (James et al., 2010). It was proposed that common configuration for faces
determines top-focused pattern of exploration, whereas no such stereotypic patterns exist
for scene exploration. Thus, due to asymmetry in attention in processing top versus lower
facial parts, it was expected in the present research to find the corresponding top versus
bottom asymmetry in BE of cropped face images. In addition, the present hypothesis about
asymmetry of BE in face processing was inspired by the portrait composition “rules” from
photography and visual arts, which put certain limitations on framing faces. There are
common recommendations suggesting “good” places to crop and those to avoid. In particular, a
crop is acceptable in a forehead area; however, it is advised to be avoided in a chin area.
For example, according to Peterson
(2016), the creator of Digital Photo Secrets website, “We are used to seeing
pictures of people with the top of their head cropped off. It typically looks fine. The same
cannot be said for the bottom of the face – do not remove someone’s chin!” The present study
aimed to explore whether the location of a crop in face representation (forehead vs. chin)
affects BE. Based on Intraub et al.
(2006) findings, which indicate that BE can be amplified on the attended part of an
image, it can be predicted that BE would be greater for the top (forehead) than for the
lower (chin) part of the head. Alternatively, consistent with Intraub et al. (2008) study, which observed that
focal attention may constrain BE, a greater attention to the upper part of the head may
result in a reduction of BE. Gagnier
et al. (2013) recorded oculomotor activity to study the mechanisms underlying the
reduction in BE and to examine whether BE reflects a lack of eye fixations near the edges of
a picture. They contrasted cropped and whole-objects stimuli and assumed that the cropped
area creates a salient marker of boundary placement. The reduction in BE was expected at the
side where the object was cropped with a picture boundary due to increased attention to the
crop area. However, Gagnier et al. found that there was no difference in fixations to the
boundary and to the cropped region. BE occurred in spite of multiple fixations to the
boundary region and regardless whether the objects were cropped or not. In the present
study, the oculomotor behavior was recorded to examine the distribution of attention.
Consistent with findings of greater attention to upper face regions (James et al., 2010; McKelvie, 1976), it was expected to observe a higher
oculomotor activity in the top part of the face, including eyes and cropped boundary
regions. In addition, the present study explored the possible asymmetry in BE for lower
versus upper parts of the face.Finally, the present research intended to explore face BE in relation to
imagery. Mental imagery seems to have a special importance in understanding
mechanisms of BE. It was suggested that scene continuity activates schematic expectations
representing the visual world beyond the picture boundaries, and such mental schema
underlies not only perception and memory but also imagination of scenes (Intraub, 1997). Intraub et al. (1998) proposed that BE may occur
regardless whether perception or imagination activated the perceptual schema. Indeed, a
great body of evidence demonstrated the similarity between perception and imagery, which
share common representations, cognitive mechanisms, and neural underpinnings (Farah, 1988; Ganis, Thompson, & Kosslyn, 2004; Kosslyn et al., 1999; Shepard & Metzler, 1971). In
particular, O'Craven and Kanwisher
(2000) found that imagery of faces and places activated the same stimulus-specific
brain regions (FFA and PPA) as in perception, but the magnitude of activation was lower for
imagery than for perception. Although visual imagery and perception showed significant
overlap, research highlighted that imagery involves more top-down processing and more
prefrontal cortex involvement than perception (Ganis et al., 2004; Mechelli, Price, Friston, & Ishai, 2004). Hubbard et al. (2010) showed that
early visual processing areas were not involved in BE and proposed that BE is underpinned by
high-level processes. These findings highlight the possible role of mental imagery in BE.
However, previous studies provided only limited evidence about the relationship between
imagery and BE. In particular, research examined how imagery instructions
may influence BE. Although BE was found primarily for images containing a background surface
but not for those with blank backgrounds (Intraub et al., 1998), imagery instructions altered
these findings. In particular, BE was observed when participants were deliberately imagining
extended backgrounds around objects on blank backgrounds (Gottesman & Intraub, 2002). Imagery may make
people believe that they have experienced some events that they have not experienced (Loftus, 2003). Yet, it was proposed
that the act of imagination per se does not always increase source memory errors. Foley, Foy, Schlemmer, and Belser-Ehrlich
(2010) demonstrated that, in contrast to spontaneous imagery instructions,
deliberate imagery instructions may even decrease source memory errors. Foley et al. claimed
that the awareness about the act of generating the images might distinguish these imagined
items from the actually seen items, thus leading to reduced source memory errors. Using a
variety of imagery instructions, Munger
and Multhaup (2016) found that explicit imagination of sensory details in the scene
does not result in increased BE. They concluded that there is no imagination effect on
BE.While the majority of studies on imagery and BE primarily focused on the effect of imagery
instructions on BE, there were only few attempts to explore the relationship between BE and
individual differences in imagery. Previous literature reported
individual differences in visual-object (visualizing pictorial appearances in terms of
shape, color, and texture) and visual-spatial (visualizing spatial relations and
transformations) imagery (Kozhevnikov
& Blazhenkova, 2013; Kozhevnikov, Kosslyn, & Shephard, 2005). In particular, Kozhevnikov et al. distinguished between two types of
individuals: object and spatial visualizers. Object visualizers tend to experience vivid and
colorful mental images and to excel in object visualization tasks (e.g., recognizing
degraded objects), whereas spatial visualizers tend to use imagery for representing spatial
relations and transformations and to excel in tasks that require spatial visualization
(e.g., mental rotation). This distinction in individual differences in imagery was based on
evidence of the dissociation between ventral “visual-object” and dorsal “visual-spatial”
pathways in the brain, underpinning processing of different aspects of visual information
(Farah, Hammond, Levine, &
Calvanio, 1988; Kosslyn &
Koenig, 1992; Mazard,
Tzourio-Mazoyer, Crivello, Mazoyer, & Mellet, 2004; Ungerleider & Haxby, 1994). The relationship
between BE and individual differences in object versus spatial visual imagery was examined
by Munger and Multhaup (2016).
Using the Object-Spatial Imagery Questionnaire (OSIQ), assessing object and spatial imagery
(Blajenkova, Kozhevnikov, & Motes,
2006), Munger and Multhaup found a significant positive correlation between
spatial imagery and BE, but not between object imagery and BE. The
authors concluded that BE might be related to a superior visual-spatial rather than
visual-object imagery ability. In addition, there is evidence that individual differences in
spatial but not object imagery are related to spatial dispersion of eye movements during the
recall of scenes (Johansson, Holsanova,
& Holmqvist, 2011). Furthermore, Mullally et al. (2012) asked participants to
visualize in their imagination the scenes extending beyond the current view and to rate the
vividness of their subjective imagery. Researchers found that patients with selective
bilateral hippocampal lesions had significant impairments in the ability to visually imagine
spatially coherent scenes (e.g., spatial relationships and locations). Markedly, the same
patients demonstrated attenuated BE (thus, better memory) than control participants. These
findings indicate that the subjective vividness of scene imagery may be associated with the
increased BE. It is important to note that Mullally et al. (2012) measured vividness associated
with imagination of spatially coherent scenes. As reported in Blazhenkova (2016), vividness that refers to imagery
of spatial properties (locations, spatial structure, and relationships) versus pictorial
object properties (color, texture, and shape) constitute separate, spatial and object,
vividness dimensions. Thus, the results of Mullally et al. (2012) may be interpreted as a
finding of the relationship between spatial imagery and BE for scenes.At the same time, the existing evidence suggests the possibility of positive association
between object imagery and BE. While some studies demonstrated that vivid
and pictorial object imagery may facilitate memory (Marks, 1973; Vannucci, Pelagatti, Chiorri, & Mazzoni, 2016),
other literature showed that the elaboration of sensory details during the encoding leads to
the increase in source memory errors (Thomas, Bulevich, & Loftus, 2003). Research has suggested that vivid and rich
in sensory detail imagery experiences may lead to later confusion between real and imagined
experiences, causing false memory errors (Gonsalves & Paller, 2000; Gonsalves et al., 2004). Markham and Hynes (1993) found the relationship
between individual vividness of imagery and memory errors in a task that, during the
encoding, required to imagine half-shapes as complete symmetrical geometric forms. The
participants were divided in “high-imagery” and “low-imagery” groups, according to their
scores on Vividness of Visual Imagery Questionnaire (VVIQ; Marks, 1973). During the recall of shapes,
high-imagery participants made more reality monitoring errors, confusing half-shapes with
complete shapes, than low-imagery participants. Notably, in contrast to Mullally et al. (2012) assessing
spatial vividness, the VVIQ instrument (Marks, 1973) measures object vividness (Blajenkova et al., 2006; Blazhenkova, 2016). Therefore, the results of Markham
and Hynes may be interpreted as a finding of the relationship between object imagery
vividness and mental extrapolation errors for objects (symmetrical shapes), but not for
scenes. Besides, numerous reports provided evidence that higher VVIQ scores were associated
with higher reality monitoring errors and false memory (Dobson & Markham, 1993; Gonsalves et al., 2004; Hyman & Pentland, 1996; Mazzoni & Memon, 2003). Overall, these findings
suggest that vividness and strength of individual imagery may lead to increased BE errors.
However, it is not yet known how BE may be affected by the type of imagery (object vs.
spatial) and the content of the image (a single object on a blank background vs. scene). The
present work examined the relationship between BE and individual differences in the two
types of imagery: object versus spatial. The present investigation implemented nonspatial
stimuli: faces on blank backgrounds. On the basis of the previous literature, it was
expected that larger BE in face processing would be associated with superior object, but not
spatial, visual imagery.In addition, because the present research used face stimuli conveying emotions, the
measures of emotional processing were also included along with the imagery
assessments. Previous studies indicated that, similar to vividness, emotional content might
induce the creation of false memories (Hyman & Pentland, 1996; Porter, Spencer, & Birt, 2003). However, the evidence regarding the
relationship between emotional processing and BE is quite controversial. Several studies
suggested that BE effect may depend on the emotional content of scenes. For example, Ménétrier et al. (2013) demonstrated
that positively valenced stimuli (i.e., actors showing happiness and pleasure through facial
and postural expressions) led to BE effect, whereas negatively valenced stimuli (i.e.,
expressing anger and irritation) did not produce directional memory distortion. Safer et al. (1998) showed that
negative emotional content of an image may lead to “tunnel memory effect” or boundary
restriction (BR), opposite to BE effect. However, other research found no difference in
magnitude of BE for pictures with emotionally neutral and emotionally charged content (Candel, Merckelbach, Houben, & Vandyck,
2004; Candel, Merckelbach,
& Zandbergen, 2003). Mathews and Mackintosh (2004) found that scene extrapolation interacts with
individual differences in emotionality: BE for very negative scenes was reduced in
high-trait-anxious individuals. Blazhenkova and Kozhevnikov (2010) indicated positive association between
individual differences in emotion and object, but not spatial, imagery. Therefore, in the
present study, it was expected that both, individual differences in emotional ability and
object imagery, similar to each other, would be positively related to BE in face
processing.To summarize, there is an increasing body of research on BE in scene perception and memory.
However, there are gaps in the knowledge regarding BE for nonscene representations, and in
particular, face images. The present work examined BE errors in face processing using
cropped face images on black backgrounds. Study 1 explored BE in different facial parts.
Using short-term and long-term memory conditions, it examined whether the location of a crop
(forehead vs. chin) affects BE. Study 2 examined the relationships between the measures of
BE in face images and different assessments of individual differences in object/spatial
imagery and emotion.
Study 1
The goal of Study 1 was to examine BE using face stimuli, cropped either in a forehead or
chin. The task was administered in two memory conditions to explore the robustness and the
possible differences in BE effect. On the basis of previous research that showed BE for both
long-term and short-term retention (Intraub & Dickinson, 2008; Safer et al., 1998), it was expected to find this
effect in both memory conditions. The early onset of BE indicates that this phenomenon
occurs at the border between perceiving and remembering (Intraub & Dickinson, 2008; Roediger, 1996). The comparison between the
long-term and short-term memory conditions could elucidate the role of perception and memory
in BE. Furthermore, it was hypothesized that participants would make BE errors for both crop
locations. However, based on the attentional asymmetry in face processing, forehead
extension was expected to be different from chin extension. In addition, this study explored
the effect of the expressed emotion on scene extrapolation. Eye-tracking data were collected
alongside with the task performance data.
Method
Participants
Eighty-three
participants were Sabancı University students (19–26 years old,
M age = 22). Participants were run either in the short-term memory
(22 males, 19 females) or in the long-term (19 males, 23 females) condition. They were
reimbursed with course credits for their participation. The research was approved by the
Sabanci University Research Ethics Council. All participants provided written informed
consent.
Materials and procedure
All participants were tested individually. They completed either a short-term or a
long-term condition of the Faces Task. This task included Encoding
& Emotional Identification and Recognition parts. During the Encoding, participants
observed photographs of faces, cropped either in a forehead or in a chin area. The
stimuli were created based on Karolinska Directed Emotional Faces picture set (Lundqvist, Flykt, & Öhman,
1998). They comprised color pictures of female faces displaying six different
emotional states (anger, disgust, happiness, neutrality, sadness, and surprise)
photographed from the front. Each emotional state was presented in four different faces
(two cropped in the forehead and two in the chin area), thus making 24 pictures in
total. The same Encoding face stimuli were used in both conditions. Forehead- and
chin-cropped faces had fixed width, but varied in height. The height of full uncropped
original Karolinska Directed Emotional Faces images corresponded to the full height of
the screen. The cropped stimuli were shifted so that the bottom of forehead-cropped
faces touched the bottom of the screen, and the top of chin-cropped faces touched the
top of the screen, thus leaving some space for a mental continuation of a cropped part.
Each face appeared for 4 s, and it was followed by Emotional Identification.The instruction included two parts. The Encoding & Emotional Identification
instruction was as follows: You will see 24 pictures of people with different facial
expressions. Try to identify the emotions conveyed in each of these photos. After each
face presentation, you will be automatically taken to the page with a question about
this face emotion. During the Emotional Identification, participants were asked, “What
was the emotion expressed by this face?” They had to select among the seven answer
options: “fear,” “anger,” “disgust,” “happiness,” “neutral,” “sadness,” and “surprise.”
The answer time was not limited, and program proceeded to the next trial after the
response. The recognition of the emotional expressions was included in the Faces Task to
encourage participants focusing on the other face properties rather than local areas of
crops and to separate the presentation of the Encoding and Recognition face stimuli.The second part of the instruction was as follows: You will see the same picture again
among the other options. Try to recognize the picture that you saw before. Select this
picture by clicking your mouse on it. During the Recognition, participants had to
recognize the previously seen images and to make a forced-choice selection among the
four options. During the first trial of the Emotional Identification task, to ensure
that a participant understands the task, the experimenter noted and showed that, though,
all the choice images represent the same face, but the crop area is different. These
selection options included the same faces with two extended and two reduced crops,
representing different degrees of possible BE and BR errors. To increase the sensitivity
of the test, none of the answer options were the same as an image presented during the
Encoding (so there was no correct answer option). Some other BE studies also did not
include the selection option identical to the studied view (Mathews & Mackintosh, 2004; Quinn & Intraub, 2007). For
example, Mathews and Mackintosh
(2004) used a forced-choice recognition task that presented four alternatives
with different degrees of BE and restriction but did not present the originally
memorized scene. Remarkably, research on eyewitness memories demonstrated that
participants confidently and falsely identified someone in the lineup of potential
suspects as being the perpetrator, when the actual perpetrator was not in the lineup
(Wells et al., 1998).In the short-term memory condition, both parts of the instruction were presented prior
to start of the test. In the long-term condition, the second part of the instruction
came after each block of Encoding & Emotional Identification. In the short-term
memory condition, each Encoding image was immediately followed by the Emotional
Identification and then Recognition (Figure 1). The response time for Emotional Identification was not limited and
not recorded. Pilot participants testing showed that it took approximately 3 to 10 s.
Because the time interval between the stimuli Encoding and Recognition did not exceed
30 s, this condition was labeled as “short-term” memory (Baddeley & Hitch, 1974). In the long-term
memory condition, the encoding images were presented in four blocks, each containing six
trials. Each Encoding block was followed by a Recognition block. This condition required
long-term memory because the time interval between the Encoding and Recognition stimuli
exceeded 30 s (Atkinson &
Shiffrin, 1968; Baddeley
& Warrington, 1970).
Figure 1.
Schematic of trial and block sequences in the short-term and long-term conditions
of the Faces Task.
Schematic of trial and block sequences in the short-term and long-term conditions
of the Faces Task.The same Recognition alternatives were used in both conditions. However, in the
short-memory condition, half of the Recognition options increased in size from right to
left, reflecting the magnitude of BE/BR in the solution options (i.e., Large BR, Small
BR, Small BE, Large BE), and another half had the reversed order. In the long-memory
condition, all choice options increased in size from right to left. To have a visual
diversity in the selection options, the size of the face crops varied, and it was
balanced between the chin and forehead areas. The face width was constant between the
four alternative options. However, because faces were cropped differently in either
forehead or chin, their heights were different. The order of face images with varying
crop locations was intermixed. It was fixed for the Encoding and Recognition, which
allowed to keep the time between the corresponding encoding and Recognition stimuli for
the same face relatively constant. Behavioral responses (mouse clicks on the selected
pictures) were recorded.
The response time for Recognition was not limited. Eye-tracking data were
collected using Tobii TX300 Eye Tracker (data rate 120, framerate 5, fixation filter
I-VT) and Tobbi Studio software. Stimuli were presented full-screen on a 23” monitor at
a resolution of 1,920 × 1,080 pixels.
Results - Behavioral Data
Types of errors during the Recognition
For the Recognition data analysis, Regions of Interest (ROIs) were created around each
answer option (Figure 2).
Because none of the answer options represented a correct answer, the Recognition was
analyzed in terms of different types of errors (i.e., Large BE, Small BE, Small BR, and
Large BR), corresponding to the four answer options. Error Frequency
was the number of mouse clicks on different Recognition options. The two types of BE
answers and the two types of BR answers were obtained in the forced-choice task, leading
to two categories of responses: Extension and Restriction Errors. The mean frequencies
of responses falling in these two categories were further compared for faces cropped in
the forehead or chin areas.
Figure 2.
ROIs around different types of Recognition errors in the Faces Task.
ROIs = Regions of Interest.
ROIs around different types of Recognition errors in the Faces Task.ROIs = Regions of Interest.
Recognition analysis
A mixed 2 × 2 × 2 three-way (two within- and one between-subjects factors) analysis of
variance (ANOVA) was conducted to assess the impact of two repeated measures variables,
Error Type (BE, BR) and cropping Location (Forehead, Chin), on participants’ recognition
(Error Frequency), across the two memory Conditions (Short-Term, Long-Term).
The adjustment for multiple comparisons was done with Bonferroni correction. The
analysis revealed a considerable main effect for Error Type, F(1,
81) = 71.757, p < .001, partial η2 = .470, showing that
participants selected more extension (M = 7.617,
SE = 0.191) and less restriction (M = 4.383,
SE = 0.191) answer options. There was a significant but weak
interaction between Error Type and Location, F(1, 81) = 4.562,
p = .036, partial η2 = .053, suggesting that Error Type
effect differed depending on the cropping Location. While the extension effect was
present for both locations of crop, it was greater for forehead-cropped images
(MD = 3.922, SE = 0.464,
p < .001; M
= 7.961, SE = 0.232; M
= 4.039, SE = 0.232) than for chin-cropped
(MD = 2.546, SE = 0.532,
p < .001; M
= 7.273, SE = 0.266; M
= 4.727, SE = 0.266) images. There was no interaction between
Error Type and Condition, F(1, 81) = 1.380, p = .243,
partial η2 = .017, suggesting no difference in the Error Type effect between
the short-term and long-term conditions. There was a weak but significant Error
Type × Location × Condition interaction, F(1, 81) = 11.511,
p = .001, partial η2 = .124. A separate analysis for two
conditions revealed that forehead versus chin asymmetry of BE effect was present in
short-term condition, F(1, 40) = 15.426, p < .001,
partial η2 = .278, but not in the long-term condition, F(1,
41) = .783, p = .381, partial η2 = .019. Figure 3 presents mean Error
Frequencies of BE and BR responses for both conditions.
Figure 3.
Recognition data: Error Frequencies for different types of errors in the Faces
Task.
ROI = Region of Interest; BE = Boundary Extension; BR = Boundary Restriction.
Recognition data: Error Frequencies for different types of errors in the Faces
Task.ROI = Region of Interest; BE = Boundary Extension; BR = Boundary Restriction.
The effect of Emotion
The possible effect of Emotion on boundary extrapolation in cropped face images was
additionally explored. A mixed 6 × 2 × 2 three-way ANOVA was conducted to assess the
impact of Emotion (Anger, Disgust, Happiness, Neural, Sadness, Surprise) and Error Type
on Error Frequency, across the two memory Conditions.
The adjustment for multiple comparisons was done with Bonferroni correction. The
analysis revealed a main effect of Error Type, F(1, 81) = 41.819,
p < .001, partial η2 = .340, demonstrating BE effect.
There was no interaction between Error Type and Condition, F(1,
81) = 1.019, p = .316, partial η2 = .012. There was a weak
but significant interaction between Emotion and Error Type, F(5,
405) = 7.895, p < .001, partial η2 = .089, suggesting
that Error Type effect depended on the Emotion. This interaction effect was analyzed
using a simple main effects analysis. Pairwise comparisons of the Emotion × Error Type
interaction revealed that Mean Difference (BE-BR) was significant only for “Angry”
(MD = 1.015, SE = 0.220,
p < .001); “Happy” (MD = 1.424,
SE = 0.221, p < .001); “Sad”
(MD = .891, SE = 0.227,
p < .001); and “Surprise”
(MD = 1.034, SE = 0.212,
p < .001) emotions. Furthermore, to compare the size of BE effect
for different types of emotions, the additional repeated measures analysis with BE-BR as
a dependent measure and Emotion as a within-subjects variable. Pairwise comparisons
revealed that BE-BR difference for “Happy” emotion was significantly larger compared
with BE-BR for “Disgust” (MD = 1.036, SE = 0.298,
p = .012) and “Neutral” (MD = 1.494,
SE = 0.261, p < .001) emotions. BE-BR discrepancy
was significantly smaller for “Neutral” emotion compared with BE-BR for “Angry”
(MD = −1.084, SE = 0.262,
p = .001), “Sad” (MD = −.964,
SE = 0.241, p = .002), and “Surprise”
(MD = −1.108, SE = 0.273, p = .002)
emotions. Figure 3 shows mean
Error Frequencies of BE and BR answers for different types of emotions. In addition, for
a clearer demonstration of Error Type effect for each emotion, this figure displays mean
difference between BE and BR measures (BE-BR). There was no interaction between Emotion,
Error Type, and Condition, F(5, 405) = 1.381,
p = .231, partial η2 = .017.
Results - Eye-Tracking Data
Encoding analysis
To analyze the visual processing during the Encoding, five ROIs were created for each
cropped face image: “cropped” area, “border” area, “eyes” area, “mouth” area, and
“noncropped” face area. As illustrated on the Figure 4, the “cropped” ROI was defined as a
missing part of a face, whereas “noncropped” ROI was the area above the “eyes” in
chin-cropped images or the area below the “mouth” the forehead-cropped faces. The “eyes”
ROI was created around the eyes, between the top of the alar nasal sulcus and just above
the superciliary arch. The “mouth” ROI included the area below the “eyes” ROI and
mentolabial sulcus. The “border” ROI included the area between the “eyes” and “cropped”
ROIs in the forehead-cropped faces or area between the “mouth” and “cropped” ROIs in the
in chin-cropped images. Two different eye-tracking metrics, based on the ROIs defined in
Figure 4, were used in the
analysis. Visit Duration was the total time in seconds spent within a
particular ROI. Visit Count was the total number of visits to an ROI,
where visit is defined as a time interval between the first fixation inside the ROI and
the next fixation outside the ROI.
Figure 4.
ROIs around different parts of the face viewed during the Encoding of the Faces
Task.
ROIs = Regions of Interest.
ROIs around different parts of the face viewed during the Encoding of the Faces
Task.ROIs = Regions of Interest.A mixed 5 × 2 × 2 three-way (two within- and one between-subjects factors) ANOVA was
conducted to assess the impact of two repeated measures variables, Face Area (“cropped,”
“border,” “eyes,” “mouth,” “noncropped”) and cropping Location (Forehead, Chin), on
oculomotor variables (Visit Duration, Visit Count) across the two Conditions
(Short-Term, Long-Term). The adjustment for multiple comparisons was done with
Bonferroni correction. The data were analyzed separately for Visit Duration and Visit
Count. Eye-tracking heat maps representing relative visit durations during the Encoding
are presented in Figure 5.
Figure 5.
Encoding eye-tracking heatmaps representing relative Visit Duration in the
short-term and long-term conditions of the Faces Task.
Encoding eye-tracking heatmaps representing relative Visit Duration in the
short-term and long-term conditions of the Faces Task.The analysis of Visit Duration revealed a substantial main effect for
Face Area, F(4, 324) = 259.601, p < .001, partial
η2 = .762. Pairwise comparisons demonstrated that “eyes”
(M = 19.965, SE = 0.694) attracted significantly
more attention than any other face areas all p’s < .001). The
“mouth” was the second most salient area; mean Visit Duration was significantly greater
for the “mouth” ROIs (M = 12.407, SE = 0.582) than for
other ROIs (all p’s < .001), except for the “eyes” areas. Visit
Duration for “cropped” (M = 1.880, SE = 0.227) and for
“noncropped” (M = 2.736, SE = 0.226) areas were
significantly smaller than all other ROIs (all p’s < .001), but not
different from each other (p = .087). Mean Visit Duration for the
“border” areas (M = 4.065, SE = 0.318) was greater
than for the “noncropped” ROIs (p = .005). The effect of cropping
Location was nonsignificant, F(1, 81) = 1.705,
p = .195, partial η2 = .021; there were no difference in
Visit Duration between forehead- and chin-cropped images. The effect of Condition was
not significant, F(1, 81) = .094, p = .760, partial
η2 = .001. There was a significant interaction between Face Area and
cropping Location, F(4, 324) = 29.360, p < .001,
partial η2 = .266, a significant but weak interaction between Face Area and
Condition, F(4, 324) = 14.886, p < .001, partial
η2 = .155, as well as a significant but weak interaction between Location
and Condition, F(1, 81) = 18.751, p < .001, partial
η2 = .188. Mean Visit Duration for the “cropped,” “border,” and “uncropped”
areas was greater in the short-term condition than in the long-term condition, but it
was greater for the “eyes” area in the long-term condition than in the short-term
condition (all p’s < .001). Face Area × Location × Condition
interaction was also significant, F(4, 324) = 10.416,
p < .001, partial η2 = .114, indicating that the
discrepancy in Face Area effect between the forehead- and chin-cropped images was
different in short and long-term conditions. In both conditions, there were no
differences in mean Visit Duration between the forehead versus chin areas for the
“cropped” and the “mouth” ROIs. For the “border” and the “eyes” ROIs’, Visit Duration
was greater for forehead- than for chin-cropped faces in the short-term condition
(p’s < .001), but not in the long-term condition. Mean Visit
Duration for the “noncropped” ROIs’ was greater for the chin- than for the
forehead-cropped faces in the short-term condition (p < .001), but
not in the long-term condition. Figure
6 demonstrates the results of comparisons between mean Visit Duration for
different ROIs in the forehead- and the chin-cropped faces, separately for the two
conditions.
Figure 6.
Faces Task Encoding data: Visit Duration and Visit Count for the different
ROIs.
ROIs = Regions of Interest.
Faces Task Encoding data: Visit Duration and Visit Count for the different
ROIs.ROIs = Regions of Interest.Consistent with Visit Duration, the analysis of Visit Count revealed a
substantial main effect for Face Area, F(4, 324) = 746.494,
p < .001, partial η2 = .902. Pairwise comparisons
demonstrated that the “eyes” (M = 33.497, SE = 0.714)
attracted significantly more attention than any other face areas all
p’s < .001. The “Mouth” was the second most salient area; mean Visit
Count was significantly greater for the “mouth” ROIs (M = 27.132,
SE = 0.758) than for the other ROIs (all
p’s < .001), except for the “eyes” areas. Mean Visit Count for the
“cropped” areas (M = 3.481, SE = 0.325) was the
smallest compared with all other ROIs (all p’s < .001). There was no
difference between mean Visit Count for the “border” (M = 7.948,
SE = 0.461) and the “noncropped” (M = 6.614,
SE = 0.396) ROIs (p = .155). The effect of cropping
Location was nonsignificant, F(1, 81) = 3.838,
p = .054, partial η2 = .045. The effect of Condition was
significant but weak, F(1, 81) = 6.136, p = .015,
partial η2 = .070. Mean Visit Count was greater for the short-term
(M = 16.568, SE = 0.479) than for the long-term
(M = 14.900, SE = 0.473) condition. There was a
significant interaction between Face Area and Location, F(4,
324) = 53.325, p < .001, partial η2 = .397, as well as a
significant but weak interaction between Face Area and Condition, F(4,
324) = 14.874, p < .001, partial η2 = .155. Visit Count
for the “cropped,” “border,” and “uncropped” areas was greater in the short-term
condition than in the long-term condition (all p’s < .001), but it
was greater for the “mouth” area in the long-term condition than in the short-term
condition (p = .041). There was no significant interaction between
Location and Condition, F(1, 81) = .090, p = .765,
partial η2 = .001. Face Area × Location × Condition interaction was
significant but weak, F(4, 324) = 12.731, p < .001,
partial η2 = .136. In both conditions, there were no differences in mean
Visit Count between the forehead versus chin areas for the “cropped” and the “mouth”
ROIs. For the “border” and the “eyes” ROIs’, mean Visit Count was greater for the
forehead- than for the chin-cropped faces in the short-term condition
(p’s < .001), but not in the long-term condition. Visit Duration
for the “noncropped” ROIs’ was greater for the chin- than for the forehead-cropped faces
in both, the short-term (p < .001) and the long-term
(p = .002), conditions. Figure 6 demonstrates the results of comparisons
between mean Visit Count for different ROIs in forehead- and chin-cropped faces,
separately for the two conditions.Consistent with the behavioral data analysis, a mixed 2 × 2 × 2 ANOVA was conducted to
assess the impact Error Type and Cropping Location on participants’ oculomotor variables
across the two memory Conditions. The data were analyzed separately for Visit Duration
and Visit Count. The adjustment for multiple comparisons was done with Bonferroni
correction. The eye-tracking data were examined using the same ROIs as the behavioral
Recognition analysis (Figure 2).
The eye-tracking heatmap visualizations are presented in Figure 7.
Figure 7.
Recognition eye-tracking heatmaps representing relative Visit Duration and mouse
clicks for the short-term and long-term conditions of the Faces Task.
Recognition eye-tracking heatmaps representing relative Visit Duration and mouse
clicks for the short-term and long-term conditions of the Faces Task.The analysis of Visit Duration revealed a substantial main effect for
Error Type, F(1, 81) = 104.030, p < .001, partial
η2 = .562, demonstrating that the participants had higher mean Visit
Duration for extension (M = 34.252, SE = 1.352) than
for restriction (M = 25.161, SE = 1.026) answer
options. There was a substantial main effect for cropping Location,
F(1, 81) = 70.645, p < .001, partial
η2 = .466, showing that the participants had higher mean Visit Duration for
the forehead-cropped (M = 32.265, SE = 1.259) than for
the chin-cropped (M = 27.149, SE = 1.040) images.
There was a significant Error Type × Location interaction, F(1,
81) = 31.698, p < .001, partial η2 = .281. The difference
between the mean Visit Duration for BE and BR answer options for the forehead-cropped
images (MD = 13.639, SE = 1.209,
p < .001; M
= 39.085, SE = 1.654; M
= 25.446, SE = 1. 080) was greater than for the chin-cropped
images (MD = 4.543, SE = 1.197,
p < .001; M
= 29.420, SE = 1.219; M
= 24.877, SE = 1.181). The effect of Condition was not
significant, F(1, 81) = 2.791, p = .099, partial
η2 = .033. There was no Error Type × Condition interaction,
F(1, 81) = .146, p = .703, partial
η2 = .002. There was no Location × Condition interaction,
F(1, 81) = .677, p = .413, partial
η2 = .008. There was a significant Error Type × Location × Condition
interaction between, F(1, 81) = 26.185, p < .001,
partial η2 = .244, indicating that the Error Type × Location interaction
varied between the long-term and the short-term conditions. A separate analysis for the
two conditions revealed that forehead versus chin asymmetry of BE effect (Error
Type × Location interaction) was present in the short-term condition,
F(1, 40) = 37.554, p < .001, partial
η2 = .484, but not in the long-term condition,
F(1, 41) = .270, p = .606, partial
η2 = .007. Figure 8
presents recognition results for the different Error Types, Conditions and cropping
Locations, based on eye-tracking data.
Figure 8.
Recognition data: Visit Duration and Visit Count for the different types of
errors in the Faces Task.
ROI = Region of Interest; BE = Boundary Extension; BR = Boundary Restriction.
Recognition data: Visit Duration and Visit Count for the different types of
errors in the Faces Task.ROI = Region of Interest; BE = Boundary Extension; BR = Boundary Restriction.The analysis of Visit Count revealed a substantial main effect for
Error Type, F(1, 81) = 131.129, p < .001, partial
η2 = .618, showing that the participants had higher mean Visit Count for
the extension (M = 60.603, SE = 1.789) than for the
restriction (M = 46.165, SE = 1.501) answer options.
There was a substantial main effect for cropping Location, F(1,
81) = 85.516, p < .001, partial η2 = .514, showing that
the participants had higher mean Visit Count for the forehead-cropped
(M = 57.355, SE = 1.671) than for the chin-cropped
(M = 49.413, SE = 1.506) answer options. There was a
significant Error Type × Location interaction, F(1, 81) = 39.501,
p < .001, partial η2 = .328. The difference between the
mean Visit Count for BE and BR answer options for the forehead-cropped images
(MD
= 20.974, SE = 1.624, p < .001;
MBE = 67.842, SE = 2.144;
MBR = 46.868, SE = 1.518) was greater
than for the chin-cropped images (MD
= 7.903, SE = 1.645, p < .001;
MBE = 53.364, SE = 1.723;
MBR = 45.461, SE = 1.709). The effect of
Condition was not significant, F(1, 81) = .006,
p = .937, partial η2 = .000. There was a significant but
weak Error Type × Condition interaction, F(1, 81) = 8.249,
p = .005, partial η2 = .092. The Error Type effect was
less pronounced in the short-term (MD
= 10.817, SE = 1.794, p < .001;
M
= 58.671, SE = 2.588; M
= 47.854, SE = 2.135) than in the long-term condition
(MD
= 18.060, SE = 1.772, p < .001;
M
= 62.536, SE = 2.527; M
= 44.476, SE = 2.110). There was no interaction between Location
and Condition, F(1, 81) = 2.085, p = .153, partial
η2 = .025. There was a significant Error Type × Location × Condition
interaction, F(1, 81) = 38.500, p < .001, partial
η2 = .322. A separate analysis for the two conditions revealed that the
forehead versus chin asymmetry of BE effect (Error Type × Location interaction) was
present in the short-term condition, F(1, 40) = 54.347,
p < .001, partial η2 = .576, but not in the long-term
condition, F(1, 41) = .005, p = .941, partial
η2 = .000. Figure 8
presents the recognition results for the different Error Types, Conditions and cropping
Locations, based on the eye-tracking data.Consistent with the behavioral data analysis, a mixed 6 × 2 × 2 ANOVA was conducted to
assess the impact of Emotion and Error Type on the eye-tracking measures, across the two
memory Conditions. The data were analyzed separately for Visit Duration and Visit Count.
The adjustment for the multiple comparisons was done with Bonferroni correction.The analysis of Visit Duration revealed a substantial main effect of
Error Type, F(1, 81) = 98.561, p < .001, partial
η2 = .549, demonstrating that the participants fixated at the BE answer
options longer than at the BR answer options. It was a significant, though weak, effect
of Emotion, F(5, 405) = 10.861, p < .001, partial
η2 = .118. Pairwise comparisons demonstrated that the faces representing
“Angry” emotion evoked longer Visit Duration than all the other emotions (all
p’s < .001, but for the “Sad” emotion p = .054).
There was a significant but weak interaction between Emotion and Error Type,
F(5, 405) = 4.216, p < .001, partial
η2 = .049, suggesting that the Error Type effect differed depending on the
Emotion. Pairwise comparisons of the Emotion and Error Type interaction revealed that
Mean Difference (BE-BR) was significant for all emotions: “Angry”
(MD = 3.771, SE = 0.470,
p < .001); “Disgust” (MD = 1.612,
SE = 0.477, p = .001); “Happy”
(MD = 3.817, SE = 0.427,
p < .001); “Neutral” (MD = 3.524,
SE = 0.564, p < .001); “Sad”
(MD = 2.113, SE = 0.451,
p < .001); and “Surprise”
(MD = 2.869, SE = 0.622,
p < .001) emotions. Furthermore, to compare the size of BE effect
for different types of emotions, the additional repeated measures analysis with BE-BR as
a dependent measure and Emotion as a within-subjects variable. Pairwise comparisons
demonstrated that mean BE-BR for the “Angry” emotion was significantly greater compared
with the “Disgust” (MD = 2.147, SE = 0.626,
p = .014). BE-BR for the “Happy” emotion was also greater compared
with the “Disgust” (MD = 2.204, SE = 0.601,
p = .007) and “Sad” (MD = 1.700,
SE = 0.554, p = .044) emotions. No other significant
differences were revealed. The effect of Condition was not significant,
F(1, 81) = 2.428, p = .123, partial
η2 = .029. There was no interaction between Type and Condition,
F(1, 81) = .013, p = .909, partial
η2 < .001. There was a weak but significant interaction between Emotion
and Condition, F(5, 81) = 2.780, p = .017, partial
η2 = .033, suggesting that the Emotion effect differed depending on the
Condition. Figure 8 presents
mean Visit Duration of BE and BR answers for the different types of Emotions in both
Conditions. In addition, for a clearer demonstration of Error Type effect for each
emotion, this figure displays the difference between BE and BR measures (BE-BR). The
interaction between Emotion, Error Type, and Condition, F(5,
81) = 26.185, p < .001, partial η2 = .244 was not
significant.Similarly, the analysis of Visit Count revealed a substantial main
effect for Error Type, F(1, 81) = 121.330,
p < .001, partial η2 = .600, demonstrating that the
participants fixated at the BE answer options more often than at the BR answer options.
There was a significant effect for Emotion, F(5, 405) = 27.227,
p < .001, partial η2 = .252. Pairwise comparisons
demonstrated that the faces representing “Angry” emotion evoked longer Visit Count than
all the other emotions (all MD > 3.705,
SE < 0.689, p’s < .001). Visit Count for faces
with the “Surprise” emotion was lower than the “Neutral”
(MD = − 1.684, SE = 0.527,
p = .030) and “Sad” (MD = −2.412,
SE = 0.592, p = .002) emotions. There was a
significant but weak interaction between Emotion and Error Type,
F(5, 405) = 4.827, p < .001, partial
η2 = .056. The effect of Condition was not significant,
F(1, 81) = .058, p = .811, partial
η2 = .001. There was a weak but significant interaction between Type and
Condition, F(1, 81) = 6.102, p = .016, partial
η2 = .070. There was a weak but significant interaction between Emotion and
Condition, F(5, 405) = 74.313, p = .024, partial
η2 = .031. Pairwise comparisons of the Emotion and Error Type interaction
revealed that Mean Difference (BE-BR) was significant for all emotions: “Angry”
(MD = 6.274, SE = 0.732,
p < .001); “Disgust” (MD = 3.183,
SE = 0.709, p < .001); “Happy”
(MD = 6.028, SE = 0.560,
p < .001); “Neutral” (MD = 4.934,
SE = 0.751, p < .001); “Sad”
(MD = 3.472, SE = 0.582,
p < .001); and “Surprise”
(MD = 3.997, SE = 0.769,
p < .001) emotions. Furthermore, to compare the size of BE effect
for different types of emotions, the additional repeated measures analysis with BE-BR as
a dependent measure and Emotion as a within-subjects variable. Pairwise comparisons
demonstrated that BE-BR for the “Angry” emotion was significantly greater compared with
the “Disgust” (MD = 3.084, SE = 0.927,
p = .020) and the “Sad” (MD = 2.783,
SE = 0.880, p = .033) emotions. BE-BR for the
“Happy” emotion was also greater compared with the “Disgust”
(MD = 2.843, SE = 0.832, p = .015)
and the “Sad” (MD = 2.542, SE = 0.764,
p = .020) emotions. The interaction between Emotion, Error Type, and
Condition, F(5, 405) = .908, p = .476, partial
η2 = .011, was not significant.
Discussion
The analyses of behavioral (Error Frequency) and eye-tracking (Visit Duration and Visit
Counts) Recognition data yielded highly consistent results. The main effect of Error Type
was found: The BE answer options were selected significantly more frequently and attracted
more attention than the BR answer options. Thus, Study 1 demonstrated the evidence of BE
effect in face processing. The analysis revealed no interaction between Error Type and
Condition for Error Frequency and Visit Duration measures; however, there was a weak but
significant interaction for Visit Count measure. These results provide the robust evidence
of the BE effect for the different retention intervals, though, Visit Count results
indicate that this effect may be somewhat more conspicuous in the long-term
condition. Furthermore, the BE effect was more pronounced for the
forehead-cropped images. This demonstrates the evidence of the asymmetry of BE effect in
face images. However, the forehead-biased asymmetry of BE effect manifested only in the
short-term Condition. Markedly, during the Encoding, the forehead-biased asymmetry in Face
Area effect was also found in the short-term, but not in the long-term Condition.The additional analyses revealed the effect of Emotion on Visit Duration and Count: The
“angry” emotion evoked more attention than all the other emotions. Furthermore, the found
interaction between Emotion and Error type suggests that the strength of BE effect may
differ depending on the type of Emotion. The analysis of the eye-tracking measures
revealed that the BE effect was greater for the “Angry” and the “Happy” emotions and
smaller for the “Disgust” and the “Sad” emotions. Similarly, but not fully consistent, the
analysis of mouse clicks revealed that the BE effect was greater for the “Happy” emotion
but smaller for the “Disgust” and the “Neutral” emotions.
Study 2
After Study 1 has established BE in processing of cropped face images, Study 2 aimed to
examine the relationship between the strength of BE and individual differences in imagery
and emotion. Thus, the same participants as in the Study 1 were asked to perform additional
tests on individual differences in object and spatial imagery and emotional processing.Thirty-nine participants from the short-term memory condition and 38 participants from
the long-term memory condition completed the full set of assessments.The participants received four different questionnaires and tasks: OSIQ, Emotion
Vividness task, Geneva Emotion Recognition Test (GERT), and Range and Differentiation of
Emotional Experience Scale (RDEES). In addition, Emotion Recognition performance data
from Faces Task (Study 1) were analyzed.
OSIQ
This is a self-report measure assessing individual differences in visual-object and
visual-spatial imagery (Blajenkova
et al., 2006). The OSIQ consists of 15 statements assessing object mental
visualization and 15 statements assessing spatial mental visualization. Participants
had to rate these items on a 5-point scale from total agreement to total disagreement.
The scores for object and spatial imagery are calculated by averaging the
corresponding 15 ratings per subscale. The internal reliabilities (Cronbach’s alpha)
of the object and spatial imagery subscales are .83 and .79, respectively (Blajenkova et al., 2006).
Emotion Vividness task
In this task, participants had to imagine six basic emotions (anger, surprise,
happiness, disgust, sadness, fear; see Ekman, 1992) as well as neutral facial
expression inside the empty face outline, presented for 5 s on a computer screen.
After imagining each emotion, participants rated the vividness of their subjective
mental image using the 5-point scale adopted from the VVIQ (Marks, 1973): 5 perfectly clear and as vivid as
normal vision), 4 (clear and reasonably vivid), 3 (moderately clear and vivid), 2
(vague and dim), 1 (no image at all, you only “know” that you are thinking of the
object). This task was developed by the author of this study, and it is not a
validated task of imagined emotion vividness.
GERT
This test is assessing the ability to accurately recognize emotional states (Schlegel, Grandjean, & Scherer,
2014). Participants watched 83 short video clips with sound, in which 5 male
and 5 female actors expressed different emotions conveyed both by facial expressions
and voice (using pseudolinguistic sentences). Participants had to select the emotion
word (from 14 emotions), which best describes the emotion expresses in the video.
RDEES
This is self-report assessing individual differences in emotional complexity, which
is defined as having emotional experiences that are broad in range and
well-differentiated (Kang &
Shaver, 2004). Seven items tap Range (e.g., “I have experienced a wide range
of emotions throughout my life”), and seven Differentiation of Emotional Experiences
(e.g., “I am aware of the subtle differences in the feelings that I have”).
Participants rated each item on a 5-point scale (1 = does not describe me very well
and 5 = describes me very well). The scores for Range (RDEES-r) and Differentiation
(RDEES-d) subscales were computed by averaging the corresponding ratings. The internal
reliability (Cronbach’s alpha) of this questionnaire is .85 (.82 for the RDEES-r
subscale and .79 for RDEES-d subscale; Kang & Shaver, 2004).
Emotion Recognition
Based on the responses given in a Faces Task (Study 1), the accuracy of emotional
recognition was computed as a number of correctly identified emotions. Note, this is
not a validated task of emotional recognition, and also it was done simultaneously
with memorization. The reliability was quite low (Cronbach’s alpha = .210). Thus,
another variable, Emotion RecognitionΔ accuracy was
computed by excluding the most inconsistent 4 items with accuracy below 50%
(Cronbach’s alpha was still quite low = .259).
Results
The relationships between different memory errors, imagery, and emotional measures were
analyzed using Pearson’s Correlational Analysis. In the short-term memory
condition (Table
1), the analysis revealed no relationship between the frequency of BE/BR errors and
any of imagery measures. Emotion RecognitionΔ accuracy in the Faces Task tended
to be positively correlated with BE and negatively with BR errors
(p’s = .091). Furthermore, longer Visit Duration and Count (both for
BE/BR images) were positively associated with higher scores on GERT (all
p’s ≤ .015). Visit Count was negatively associated with Vividness of
Emotional Imagery (p = .054 for Extension and p = .015
for Restriction Visit Count).
Table 1.
Correlations Between all the Measures in the Short-Term Condition of the Faces
Task.
1
2
3
4
5
6
7
8
9
10
11
12
13
1. Extension errors
1.00
2. Restriction errors
−1.00*
1.00
3. Extension duration
.303†
−.303†
1.00
4. Restriction duration
−.050
.050
.874*
1.00
5. Extension count
.314†
−.314†
.873*
.793*
1.00
6. Restriction count
−.056
.056
.717*
.853*
.876*
1.00
7. OSIQ-object
−.011
.011
.050
.093
−.034
.015
1.00
8. OSIQ-spatial
.048
−.048
.108
.066
−.157
−.212
.028
1.00
9. Emotion vividness
.049
−.049
−.075
−.192
−.310†
−.386*
.232
.164
1.00
10. GERT
.073
−.073
.420*
.437*
.461*
.495*
.241
−.281†
−.077
1.00
11. RDEES-r
−.088
.088
.163
.232
.089
.153
.426*
−.077
.339*
.485*
1.00
12. RDEES-d
.124
−.124
.008
.035
.058
.081
.401*
−.214
.295†
.148
.492*
1.00
13. Emotion recognition
.141
−141
−.251
−.284†
−.150
−.193
−.071
−.161
.008
.016
−.029
.029
1.00
Emotion recognitionΔ
.275†
−.275†
−.016
−.066
.037
−.079
−.057
−.054
.049
.161
.086
.058
.870*
Note. OSIQ = Object-Spatial Imagery Questionnaire; GERT = Geneva
Emotion Recognition Test; RDEES = Range and Differentiation of Emotional
Experience Scale.
p < .05. †p < .10.
Correlations Between all the Measures in the Short-Term Condition of the Faces
Task.Note. OSIQ = Object-Spatial Imagery Questionnaire; GERT = Geneva
Emotion Recognition Test; RDEES = Range and Differentiation of Emotional
Experience Scale.p < .05. †p < .10.In the long-term memory condition (Table 2), the analysis revealed significant
positive relationship between the frequency of BE errors and object imagery
(p = .038). Inversely, the relationship between the frequency of BR
errors and object imagery was negative (p = .020). In addition, attention
to BR images was significantly and negatively associated with object imagery
(p = .039 for Visit Duration and p = .014 for Visit
Count). Furthermore, RDEES-d was positively associated with frequency of BE Errors
(p = .042), and negatively with frequency of BR Errors
(p = .036), as well as restriction Visit Count
(p = .025). The similar, but nonsignificant, trends were observed between
the BE Errors and Emotion Vividness as well as GERT. Emotion Recognition accuracy in the
Faces Task tended to be negatively associated with visual attention, both for BE/BR images
(p’s ≤ .033 for Emotion RecognitionΔ and Visit Count,
p = .032 for Visit Duration).
Table 2.
Correlations Between all the Measures in the Long-Term Condition of the Faces
Task.
1
2
3
4
5
6
7
8
9
10
11
12
13
1. Extension errors
1.00
2. Restriction errors
−1.000*
1.00
3. Extension duration
.341*
−.341*
1.00
4. Restriction duration
−.333*
.333*
.671*
1.00
5. Extension count
.316†
−.316†
.826*
.519*
1.00
6. Restriction count
−.398*
.398*
.512*
.868*
.580*
1.00
7. OSIQ-object
.372*
−.372*
−.128
−.334*
−.146
−.406*
1.00
8. OSIQ-spatial
−.032
.032
.174
−.019
.077
−.017
−.231
1.00
9. Emotion vividness
.267
−.267
−.033
−.140
−.093
−.143
.322*
.035
1.00
10. GERT
.186
−.186
.082
.003
.147
.115
.117
−.075
.253
1.00
11. RDEES-r
.228
−.228
−.054
−.119
−.124
−.235
.451*
−.068
.095
.036
1.00
12. RDEES-d
.335*
−.335*
−.115
−.254
−.223
−.425*
.562*
−.096
.135
−.056
.677*
1.00
13. Emotion recognition
.104
−.104
−.114
−.202
−.195
−.344*
.439*
.005
−.003
.083
.010
.145
1.00
Emotion recognitionΔ
.119
−.119
−.184
−.315†
−.323*
−.462*
.378*
−.059
−.052
.079
−.141
.142
.856*
p < .10. *p < .05.
Note. OSIQ = Object-Spatial Imagery Questionnaire; GERT = Geneva
Emotion Recognition Test; RDEES = Range and Differentiation of Emotional
Experience Scale.
Correlations Between all the Measures in the Long-Term Condition of the Faces
Task.p < .10. *p < .05.Note. OSIQ = Object-Spatial Imagery Questionnaire; GERT = Geneva
Emotion Recognition Test; RDEES = Range and Differentiation of Emotional
Experience Scale.In addition, the relationship between imagery and emotional measures was examined using
the combined data from the long-term and short-term conditions (N = 79).
Object imagery scale of the OSIQ was positively correlated with Emotion Vividness
(r = .289, p = .010), RDEES-r
(r = .439, p < .001), RDEES-d
(r = .508, p < .001), as well as with Emotion
Recognition accuracy in the Faces Task (r = .223,
p = .049)/Emotion RecognitionΔ (r = .193,
p = .091). GERT was positively correlated with RDEES-r
(r = .268, p = .018), but negatively and marginally
significantly with spatial imagery scale of the OSIQ (r = −.215,
p = .058). Emotion Vividness tended to correlate with RDEES-r
(r = .216, p = .058) and RDEES-d
(r = .220, p = .053), and the latter two were also
interrelated (r = .589, p < .001).Study 2 demonstrated positive relationship between BE measures and individual differences
in object but not spatial imagery. This indicates that individuals with higher object
imagery showed greater BE and lesser BR effect. Similarly, individuals who reported higher
emotional differentiation tended to have more pronounced BE errors. However, the relation
between BE and imagery or emotional measures was observed primarily in the long-term
condition. Consistent with the previous research (Blazhenkova & Kozhevnikov, 2010), object, but
not spatial imagery, tended to be positively associated with different emotional measures,
which tended to be interrelated. Overall, object imagery and emotional measures’
relationships with BE measures were similar.
General Discussion
The present work examined BE phenomenon in face processing. The results of the Study 1
demonstrated significantly more BE than restriction recognition errors, as revealed by both
performance and eye-tracking measures. Thus, consistent with previous literature (Hubbard et al., 2010; Intraub & Bodamer, 1993), the
present findings suggest that people are prompted to perceptual and memory errors, in which
they see or remember seeing parts of images that were not visible during the presentation
but were possible to be inferred from the visible information. The present study
demonstrated BE effect in both short-term and long-term memory
conditions. This is consistent with other research that found BE for a wide range
of retention intervals (Intraub &
Dickinson, 2008; Intraub &
Richardson, 1989; Safer
et al., 1998).The present results indicate that BE effect could be somewhat more pronounced in
the long-term condition. The found difference in BE effect between the two memory
conditions was rather small and appeared only in Visit Count analysis. From the multisource
model perspective (Intraub,
2012), this finding may suggest that the increased retention time could lead to a
greater difficulty in distinguishing between amodally generated and remembered visual
sensory information, thus resulting in a greater BE in the long-term condition. In addition,
the difference between the two memory conditions may be explained by the difference in the
awareness about the nature of the memory task. In the short-term condition, after the very
first trial, participants became aware of the subsequent memory test that required boundary
knowledge, whereas in the long-term condition, participants became aware about the memory
test only after the first block of trials. Moreover, due to a significant delay between the
Encoding and Recognition in the long-term condition, the awareness about memorization demand
during the Encoding could decrease. It is important to note that during the Encoding,
participants had to simultaneously identify the emotion and to memorize the face, which
could be considered as a dual task (Pashler, 1994). Emotional identification was a primary task, and Recognition came
afterward. Eye-tracking data indicated that the participants, first, focused on the
emotional identification. The majority of the first fixations were in the “eyes” area,
whereas other facial parts were examined later. Because the differences between the
Recognition answer options appeared solely in the place of a crop, the “border” area
provided important cues for face memorization, critical for the subsequent Recognition. As
it can be seen from the eye-tracking visualizations, with more trials, participants
increased their attention toward the “border” areas, most probably because they were
realizing the importance of this area for the Recognition performance. Indeed, several
participants spontaneously mentioned that they were focused on border areas in demand to the
expected subsequent Recognition task. Similarly, Gagnier et al. (2013) suggested that the cropped
region is a salient marker of boundary placement and found multiple fixations to this area.
In the long-term condition, the Recognition was delayed, so, possibly, the participants were
more focused on emotional identification and less focused on the memorization for further
Recognition task. Indeed, in the short-term condition, the attention to the “border” and its
neighboring “cropped” areas was greater than in the long-term condition, while the attention
to the “eyes” and “mouth” areas was greater in the long-term condition. These results are
similar to Gagnier et al. (2013)
finding that the participants who were told that the boundary memory was tested had higher
oculomotor activity near the boundaries of pictures than the participants who were not
informed about the nature of the test. Thus, consistent with previous research that showed
the reduction of BE error when participants were informed about the nature of the task
(Gagnier et al., 2013; Intraub & Bodamer, 1993), the
short-term memory condition in the present research yielded somewhat reduced BE error.The present research demonstrated BE for isolated face images on blank
backgrounds. This is unlike the majority of the studies, which found BE primarily
for images containing scenes with background texture but not in response to isolated objects
removed from the background context (Chadwick et al., 2013; Gottesman & Intraub, 2002, 2003; Intraub et al.,
1998; Mullally et al.,
2012). It should be noted that stimuli used in the present work contained cropped
faces, which are likely to elicit the sense of continuation beyond the picture boundaries.
Thus, the perception of close-up cropped faces yields BE, as the perception of scenes does.
Probably, cropped faces, similar to truncated scenes, elicit BE, which may be not true for
other types of nonface objects. Further research is needed to examine whether this effect
could be found for different nonface objects. The present findings advocate some common
mechanisms in face and scene perception. As suggested by previous literature, the perception
of faces involves configural processing, for example, a perception of the relations between
the separate features rather than their processing in piecemeal or analytical way (Tanaka & Farah, 1993; Tanaka & Sengco, 1997).
Configural processing of faces implies holistic processing (building a gestalt from separate
individual features), as well as relational arrangement of features (two eyes above the nose
and a mouth below) and distances between these features (Maurer, Grand, & Mondloch, 2002). Similarly,
scene perception involves configural and relational processing (Chun, 2003). Furthermore, Gallagher et al. (2005) found that the salience of
the scene background increases the magnitude of BE. Indeed, a large body of evidence
advocates that faces are extremely salient visual stimuli (Farah et al., 1998). Besides, it may be the case that
face images elicited BE in the current study because the stimuli represented close-up views.
As indicated by previous research (Intraub et al., 1992), the degree of BE grows as increasingly close-up views are
presented.Furthermore, the present findings provide a new evidence on asymmetry of
BE. Consistent with predictions, the results of the Study 1 demonstrated
significantly more pronounced BE errors for forehead- than for chin-cropped face area. The
Recognition analysis showed that, overall, the measures of BE in the forehead-cropped faces
exceeded the measures of BE in the chin-cropped faces. In support of the attentional
explanation of the asymmetry in BE, the present work found the asymmetry in attention in
processing different facial parts. The present findings are consistent with the literature
on face processing, showing that greater attention is paid to the upper part of the face,
and in particular, to eyes area (Janik
et al., 1978; McKelvie,
1976; Mertens et al.,
1993; Pellicano et al.,
2006; Walker-Smith et al.,
1977; Yarbus, 1967).
The eye-tracking analysis of the Encoding demonstrated that, in both memory conditions, for
both forehead- and chin-cropped faces, most attention was paid to the “eyes” and then
“mouth” regions. Possibly, the presence of eyes may cue the attention to the neighboring
“border” areas. The attentional cueing by the “eyes” may elicit BE in the upper part of the
face and suppress it on the uncued side (see also Intraub et al., 2006). The present findings that BE
was greater for the upper than the lower part the face may be explained by the increased
attention to the “eyes” area. Surprisingly, while the “eyes” was the most salient area in
both conditions, the asymmetry of BE effect was evident only in the short-term, but not in
the long-term condition. It is possible that with a more delayed recognition testing, the
memory representation may include less visual information and more information suggested by
the mental schema. The enhanced visual attention in the upper part of a face may trigger the
forehead-biased asymmetry of BE. However, visual detail may be not remembered after a
significant delay, thus the asymmetry of BE could be observed only in the short-term, but
not in the long-term condition. Also, the intriguing difference in asymmetry of BE between
the two memory conditions could be related to the difference in attentional distribution
during the Encoding. The analysis of oculomotor behavior during the Encoding demonstrated
that in the short-term condition, the “eyes” areas in the forehead-cropped faces attracted
more attention than in the chin-cropped faces. Even when excluding “eyes,” the upper
“border” areas attracted more attention than the lower “border” face areas. Besides,
“noncropped” upper face areas attracted more attention than “noncropped” lower face areas.
Markedly, this increased attention to the upper face parts in the forehead-cropped compared
with chin-cropped faces was found only in the short-term condition. This indicates that the
differences in processing forehead- and chin-cropped faces between the two memory conditions
already appeared during the Encoding. Overall, the earlier comparisons between the two
memory conditions suggest that the specific patterns of BE errors, that is, forehead-biased
asymmetry, revealed in memory (during the Recognition) were consistent with the specific
patterns of attention revealed during the perception (Encoding).Besides the attentional asymmetry in face processing, there are other possible explanations
why such an asymmetry in BE in face processing may exist. One possibility could be that in
our everyday life, we are more used to see forehead- than chin-cropped faces (e.g., people
wearing hats); due to this experience, we may be better in mental continuation of the upper
parts rather than lower parts of faces. This possibility could be explored by examining BE
in face images of human-like animals (e.g., monkeys), who do not typically cover their
heads. Furthermore, because previous research found that asymmetry in scene representation
may be caused by the anticipatory processing in a certain direction, and in particular,
anticipated motion (Courtney &
Hubbard, 2004), it is possible that “eyes” may convey some sort of motion
information, causing BE. There is evidence that gaze direction may indicate anticipated
action and intention (Klin, Jones,
Schultz, & Volkmar, 2003). Additional research is needed to explore the
possibility whether gaze direction (e.g., looking to the left or to the right side, looking
up or down, closed or opened eyes) may influence the asymmetry of BE. Alternatively, a
greater BE of the upper face part could be because the upper face area is greater in size
than the lower one. Experiments, using artificially modified faces with equalized forehead
and chin area, can address this possibility. Another possibility, which also can be tested
by using artificially modified face stimuli, is that the asymmetry of BE in faces may be
caused by the perceptual differences between the top of the head, which does not continue,
and the chin area, which extends to the neck and below body parts. Further research is
necessary to determine which explanation is best supported. Notably, because both conditions
used the same face stimuli, the found difference in the asymmetry of BE between the two
conditions indicates that this asymmetry could not fully be due to a bias caused by the
characteristics of the stimuli. Irrespective of why such an asymmetry may occur, the present
finding of asymmetry of BE in face images expands the existing evidence that the content of
an image may influence BE. Furthermore, it contributes to understanding the roles of
perception and memory in BE phenomenon.The results of Study 2 showed that individual differences in imagery relate to
BE. In particular, object, but not spatial, imagery was associated with BE. These
findings support our hypothesis that, depending on the content of the image (scene vs.
isolated object or face), BE may relate not only to the strength of spatial
imagery, as in Munger and Multhaup
(2016) study, which used scenes, but also to the strength of
object imagery, as in the present study, which used faces. Besides, the
present results demonstrated that object imagery was positively associated with vividness of
imagined emotional expressions, which requires visualizing pictorial perceptual details
inside the empty face contour.Emotion vividness, similar to object imagery, tended to be positively associated with
extension and negatively with restriction errors. The present results extend the research of
Mullally et al. (2012), who
found that individual differences in spatial scene imagery vividness was associated with BE
for scenes, whereas the present findings show the association between object imagery and BE
for faces. In addition to measures of individual differences in imagery, the current
research included measures of emotional processing. Consistent with expectations and
previous research (Blazhenkova &
Kozhevnikov, 2010), emotional measures were positively related with object but not
with spatial imagery. Similar to object imagery, the emotional measures tended to have
positive relationship with BE errors. Intriguingly, the relationships between the BE
measures and individual differences in imagery were evident mostly in the long-term
condition. Thus, the individuals with greater imagery abilities tended to have a greater BE.
Probably, because imagery involves a top-down mental constructive processing (Ganis et al., 2004; Mechelli et al., 2004), these
individuals were more likely to make commission memory errors. The relationship between BE
and imagery could manifest only in the long-term condition because a delayed recognition
testing task required less sensory information and more mental schema suggestions. Previous
explanations of BE mechanisms indicated the involvement of mental imagery or mental
visualization of nonvisible image parts, extending beyond the borders. Foley, Foley, Scheye, and Bonacci (2007) indicated
the resemblance of BE errors to object completion errors, resulting from mental filling-in.
Perceptual filling-in involves mental completion of visual image and the perception of
visual features that are physically present only the surroundings (Komatsu, 2006; Walls, 1954). It is phenomenologically experienced
as a perception of the completed textured and colorful objects (Kanizsa, 1979; Ramachandran & Gregory, 1991) and having a
coherent object representation (Shipley
& Kellman, 1992; Yantis,
1995). However, previous research that showed that BE for scenes occurred
regardless of whether a boundary cropped an object or not, ruled out the object completion
explanation of BE (Intraub et al.,
1992; Intraub & Bodamer,
1993; Intraub & Richardson,
1989). Besides, Intraub et al.
(1998) indicated that imagination of background outward of central objects
(visual-spatial information) resulted in BE, whereas imagination of the color within the
central object (visual-object information) did not. In contrast to scene images, possibly,
in case of face cropped images, the anticipatory representation of the cropped face part may
cause the expectation of a continuation of the face itself (a central cropped object),
rather than a scene background outward the face, thus it may involve mental completion
processes. Research showed that eye scanpaths during visual imagery may reenact those of
perception and reflect visual scene content (Brandt, & Stark, 1997; Laeng, &
Teodorescu, 2002; Mast, & Kosslyn, 2002). Thus, it is probable that the eye fixations in
the cropped areas indicate that participants mentally visualized the missing facial part. As
can be seen from the eye-tracking encoding heat maps (Study 1), in the cropped areas,
participants were mostly focused inside the anticipated missing face part rather in the
background of a scene outward the face. As suggested by Gottesman and Intraub (2003), image boundaries can be
considered a partial occluder of a larger scene that extends behind it. In case of amodal
completion, the information about the occluded parts of an object is not supported by
sensory modalities. Hubbard et al.
(2010) proposed that, although object completion is unlikely to be the cause of BE,
BE might induce amodal continuation/completion of a cropped object, as in case of a
partially occluded object (Hubbard
et al., 2010).Gottesman and Intraub (2002)
proposed that amodal continuation of surfaces plays fundamental role in scene perception,
especially for surfaces truncated by boundaries. They claimed that spatial extrapolation
does not require detailed and rich background but is just the suggestion of a continuity
beyond the image boundaries. However, it is possible that in case of close-up face images,
BE may involve some sort of modal completion rather than (or in addition to) amodal
continuation. Nonspatial object imagery, which refers to visualization of sensory surface
properties such as color and texture, may be more involved in extrapolation of a cropped
face image beyond the boundaries. Future efforts to establish the nature
of mental extrapolation in processing of cropped face images could provide insight about the
characteristics of imagery involved in BE. The current study brings a question whether the
scene construction schema activated by a limited view is purely spatial and how it depends
on the content of the scene. While the existing neuroimaging research indicates that BE is
supported by the scene-selective brain regions but not by those specific to object
recognition (Park et al., 2007),
the present results raise a question whether BE of cropped face images may be underpinned by
brain areas involved in visual-object processing. Further research is needed to examine the
cognitive mechanisms and neural underpinnings of BE during cropped face processing. The
current literature indicates the existence of both visual-spatial and visual-object aspects
in scene representations. A dual-path model of scene processing, based on the neurological
“what” versus “where” distinction (Ungerleider & Haxby, 1994), dissociate between processing object and spatial
information (Chun, 2003).
Similarly, Aguirre and D’Esposito
(1997) distinguished between processing of large-scale environmental scene
pictorial appearance (e.g., landmark knowledge) and spatial locations within a scene (e.g.,
survey knowledge), subserved by distinct ventral “what” and dorsal “where” neural pathway.
Although visual-spatial properties such as position were shown to be better encoded in scene
representations than visual-object surface properties such as color (Aginsky & Tarr, 2000), object visualization may
play a significant role in scene processing and BE, depending on scene content. The present
work adds a new dimension to the research on BE and encourages the future investigation on
individual differences in object versus spatial imagery and emotion in relation to BE
depending on stimuli type.Furthermore, the exploratory analyses of the effect of different emotions on BE revealed
mixed findings. Faces with “Angry” and “Happy” emotions tended to evoke a greater boundary
extrapolation. These results are inconsistent with previous research that showed no
difference in BE for images with different emotional content (Candel et al., 2003, 2004). They are only partially consistent with the
research that demonstrated BE only for scenes with positive but not negative emotional
content (Ménétrier et al., 2013;
Safer et al. 1998). Such a
discrepancy in findings could be due to the nature of the stimuli that may mobilize
different cognitive mechanisms. In particular, Candel et al. as well as Safer et al. used
neutral and aversive scenes that evoked highly unpleasant and arousing emotions, whereas the
present study used faces and the task that involved emotional recognition. Ménétrier et al.
used dynamic stimuli representing human bodies cropped in the midthigh or knee level,
whereas the present study used static pictures representing cropped faces. The present
findings suggest that both positive and negative emotions may induce BE. Notably, both Angry
and Happy emotions are categorized as “aroused,” while disgust and sad are “not aroused”
(Russel, 1980). Possibly, the level or arousal may affect the strength of BE. Contrary to
the suggestion that BE may be “immune to emotional content because it involves early
perceptual processes, rather than relatively late reconstructive memory processes” (Candel et al., 2003, pp. 694–695),
short-term and long-term memory conditions produced similar effects of emotion on BE.
However, present results should be treated with caution due to a small number of stimuli per
Emotion. Future studies are needed to examine the effect of different emotions on spatial
extrapolation in face stimuli.One of the limitations of the present study was using only female faces as stimuli.
Unlikely, but possibly, the found results may not be observed for stimuli using faces of
males. The investigation of this possibility requires an additional investigation. For the
present data, the comparison between males and females on BE and BR measures (mean Error
Frequencies, Visit Duration, and Count) did not reveal any significant sex differences.
Another limitation was that presentation of Recognition answer options was not exactly the
same for the short-term (half increased from right to left, half increased from left to
right) and the long-term condition (all increased from right to left). This discrepancy was
due to the fact that the long-term condition was developed first. After the consideration of
the long-term memory limitations, the research design was improved in the short-term
condition. Despite this difference, the results indicate the evidence of BE for both orders
of the Recognition choices. Furthermore, the present study did not use any deliberate
imagery instructions, and during the encoding, participants were prompted to focus on
recognition of facial expressions and face memorization. As suggested by previous research,
spontaneous imagery, in contrast to deliberately induced imagery, may encourage source
memory errors (Foley et al.,
2010). Possibly, such an instruction amplified the found BE effect. Further research
is needed to examine the effects of imagery instructions on BE, manipulating not only
deliberate versus spontaneous imagery but also object versus spatial imagery
instructions.In sum, the current work presented new evidence on BE in face processing. Furthermore, it
demonstrated that for short retention intervals, BE errors were more pronounced for
forehead, than for chin face areas, which provided a new evidence of asymmetry in BE.
Moreover, individual differences in emotional ability and object, but not spatial, imagery
were found to be positively associated with BE in face processing. This investigation raises
a number of research questions and highlights future research directions, which could
elucidate the nature of BE phenomenon.