Literature DB >> 35066907

Hierarchical Integration of Communicative and Spatial Perspective-Taking Demands in Sensorimotor Control of Referential Pointing.

Rui 睿 Liu 刘1, Sara Bögels1, Geoffrey Bird2,3, W Pieter Medendorp1, Ivan Toni1.   

Abstract

Recognized as a simple communicative behavior, referential pointing is cognitively complex because it invites a communicator to consider an addressee's knowledge. Although we know referential pointing is affected by addressees' physical location, it remains unclear whether and how communicators' inferences about addressees' mental representation of the interaction space influence sensorimotor control of referential pointing. The communicative perspective-taking task requires a communicator to point at one out of multiple referents either to instruct an addressee which one should be selected (communicative, COM) or to predict which one the addressee will select (non-communicative, NCOM), based on either which referents can be seen (Level-1 perspective-taking, PT1) or how the referents were perceived (Level-2 perspective-taking, PT2) by the addressee. Communicators took longer to initiate the movements in PT2 than PT1 trials, and they held their pointing fingers for longer at the referent in COM than NCOM trials. The novel findings of this study pertain to trajectory control of the pointing movements. Increasing both communicative and perspective-taking demands led to longer pointing trajectories, with an under-additive interaction between those two experimental factors. This finding suggests that participants generate communicative behaviors that are as informative as required rather than overly exaggerated displays, by integrating communicative and perspective-taking information hierarchically during sensorimotor control. This observation has consequences for models of human communication. It implies that the format of communicative and perspective-taking knowledge needs to be commensurate with the movement dynamics controlled by the sensorimotor system.
© 2021 The Authors. Cognitive Science published by Wiley Periodicals LLC on behalf of Cognitive Science Society (CSS).

Entities:  

Keywords:  Kinematics; Perspective-taking; Recipient design; Social interaction; Spatial representation

Mesh:

Year:  2022        PMID: 35066907      PMCID: PMC9287027          DOI: 10.1111/cogs.13084

Source DB:  PubMed          Journal:  Cogn Sci        ISSN: 0364-0213


Introduction

Pointing one's finger to a location in space in order to signal a referent for an addressee is one of the biomechanically simpler instances of human communicative behavior, yet referential pointing is cognitively complex: It invites communicators to consider the presumed knowledge of addressees, an instance of recipient design (Blokpoel et al., 2012; Clark & Murphy, 1982; Tomasello, Carpenter, & Liszkowski, 2007). It has been argued that referential pointing is a fundamental human communicative behavior, relevant to understand other forms of human communication, with or without accompanying speech (Cooperrider, 2016; de Ruiter, 2000; de Ruiter, Bangerter, & Dings, 2012; Enfield, Kita, & de Ruiter, 2007; Enfield & Sidnell, 2017; Liszkowski, Brown, Callaghan, Takada, & de Vos, 2012; McNeill, 1992; Tomasello, 2008; Tomasello et al., 2007). Referential pointing provides an ecologically valid testing ground for understanding the integration of sociocognitive information with sensorimotor processing (Creem‐Regehr, Gagnon, Geuss, & Stefanucci, 2013; Murillo Oosterwijk et al., 2017; Pouw, Trujillo, & Dixon, 2020; Winner et al., 2019). Previous works have shown how communicators organize their pointing movements incorporating the physical location of addressees (Özyürek, 2002; Cleret de Langavant et al., 2011; Pezzulo, Iodice, Ferraina, & Kessler, 2013; Murillo Oosterwijk et al., 2017; Winner et al., 2019). However, it remains unclear whether and how communicators’ inferences about addressees’ mental representation of the interaction space influence planning and control of referential pointing. Inferring others’ mental representation of space is conceptualized as spatial perspective‐taking, and it is differentiated in two levels according to the cognitive demands involved (Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Salatas & Flavell, 1976; Surtees, Apperly, & Samson, 2013). Level‐1 perspective‐taking (PT1) is required to infer whether part of a visual scene can be seen by another observer. It relies on calculating another observer's line‐of‐sight in an egocentric frame of reference. Level‐2 perspective‐taking (PT2) is required to infer how a visual scene is perceived by another observer. It relies on transforming frames of reference between self and others, adopting an allocentric frame of reference. Perspective‐taking is considered a prerequisite of intentional communication (Brown‐Schmidt & Hanna, 2011; Clements‐Stephens, Vasiljevic, Murray, & Shelton, 2013; Keysar, Barr, Balin, & Brauner, 2000). Previous studies have examined how interlocutors make use of perspective information in language comprehension and production, for example, by manipulating shared and privileged information (Director Task, Barr, 2008; Brown‐Schmidt & Hanna, 2011; Keysar et al., 2000). Those studies have gained insights into how our knowledge about the other's perspective influences the generation of linguistic cues and joint attention. However, it remains unclear whether and how perspective‐taking and communicative demands interface with sensorimotor processes during the production of communicative gestures. Here, we differentiate between three potential mechanisms coupling a communicator's sensorimotor system with the cognitive operations handling communicative and perspective‐taking demands. We developed a communicative perspective‐taking (CPT) task in which communicative and perspective‐taking demands were manipulated independently, in a full‐factorial design. In this task, a communicator and an addressee sit on the opposite side of a table. The communicator points to one among a set of digital stimuli presented at different locations on the table while facing the addressee. A visual barrier is interposed between the addressee and the digital stimuli, allowing the communicator a full view of the stimuli and the barrier, whereas the addressee can see only a subset of the stimuli through the barrier. The communicator points either to instruct the addressee which digital stimulus should be chosen, when the addressee does not know the target (communicative pointing, COM) or to predict which stimulus the addressee will choose, when the addressee knows the target (non‐communicative pointing, NCOM). To decide which stimulus to point to, on each trial, the communicator has to spontaneously infer either which target stimulus can be seen by the addressee through the barrier (PT1), or which of the stimuli seen by the addressee looks like the target when seen from the addressee's perspective (PT2). We recorded the kinematics of the communicator's right hand throughout the experiment and analyzed temporal and spatial features of the pointing movements. Following previous studies, we focused on kinematic features relevant to parameterizing the sensorimotor processes of movement planning and execution (Becchio, Sartori, & Castiello, 2010; Cleret de Langavant et al., 2011; Murillo Oosterwijk et al., 2017). Each pointing movement was segmented into four stages: planning, approach, holding, and returning, approximately corresponding to the “rest,” “stroke,” “hold,” and “retraction” stages used in the analysis of co‐speech gestures (Kita, Gijn, & Hulst, 1998; McNeill, 1992). Temporally, we focused on the duration of the planning stage (i.e., the period between presentation of the stimuli and movement onset) and of the holding stage (i.e., the period between arrival at the end‐position and onset of the return movement away from the end‐position). Spatially, we computed displacement and variability of the approach trajectories, parameters known to index movement magnitude and movement corrections, respectively (Bays & Wolpert, 2007; Gallivan, Chapman, Wolpert, & Flanagan, 2018; Körding & Wolpert, 2006). Building on existing work, we expect to validate the current experimental setup by reproducing the observations that PT2 requires longer planning time than PT1 (Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Surtees, Apperly, & Samson, 2016) and that communicative demands result in longer holding times and exaggerated pointing trajectories (Becchio et al., 2010; Cleret de Langavant et al., 2011; Krishnan‐Barman, Forbes, Hamilton, & de, 2017; Murillo Oosterwijk et al., 2017; Peeters, Chu, Holler, Hagoort, & Özyürek, 2015; Trujillo, Simanova, Özyürek, & Bekkering, 2020). The kinematic parameters considered in this study have been chosen to differentiate between three hypotheses on the mechanism coupling cognitive and sensorimotor systems during referential pointing (Table 1). Hypothesis 1 (“two steps model”) proposes that those systems are independent: A communicator first decides where to point, according to perspective‐taking demands (Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Surtees et al., 2016), and then plans and controls the movement according to communicative demands. This hypothesis predicts a null effect of perspective‐taking on the bulk of the pointing trajectory and limits perspective‐taking effects to planning time, while higher communicative demands would lead to more emphatic pointing trajectories as shown previously (Cleret de Langavant et al., 2011; Murillo Oosterwijk et al., 2017; Winner et al., 2019). Hypothesis 2 (“leaky two steps model”) proposes that perspective‐taking effects, occurring during the determination of the pointing location before movement onset, might leak into later stages of movement execution when the pointing finger approaches the end‐position (the “apex” in co‐speech gesture literature), as communicators might only consider the accuracy of the pointing end‐positions that directly determines the task performance. Under the framework of optimal feedback control (Schwartz, 2016; Scott, 2004, 2016; Todorov & Jordan, 2002, 2003), deviations away from the predetermined forward kinematics (i.e., the anticipatory internal model that specifies end‐effectors and joint kinematics) will be corrected only for the task‐relevant dimensions. Specifically, the principle of minimum intervention of sensorimotor control argues that only the end‐position of the movement, instead of the whole trajectory, is modified by the sensorimotor system, unless external perturbation while moving hampers task performance (Nashed, Crevecoeur, & Scott, 2012; Keyser, Medendorp, & Selen, 2017; Todorov & Jordan 2003). This sensorimotor coding mechanism would predict the leakage of perspective‐taking into late‐stage movement execution due to the control of end‐position; and it would occur in parallel to an early effect of communicative demand on the approach trajectory (Cleret de Langavant et al., 2011; Murillo Oosterwijk et al., 2017; Winner et al., 2019). Hypothesis 3 (“interactive model”) proposes that communicators might bear in mind the addressee's perspective as well as communicative demands while planning and controlling the pointing movements. This hypothesis predicts an under‐additive interaction effect of perspective‐taking and communicative demands in the trajectory. The reason is that the more control signals are present, the more noise emerges in the sensorimotor system (i.e., “signal‐dependent noise”; Faisal, Selen, & Wolpert, 2008; Harris & Wolpert, 1998; Wolpert & Ghahramani, 2000; Wolpert, Doya, & Kawato, 2003). Hypothesis 3 implies that the two types of cognitive demands might share a common coding in the sensorimotor interface, allowing those different cognitive demands to be integrated to generate a more efficient control policy (“control policy” refers to internal rules on how anticipated consequences of a movement and real‐time sensory information are used to generate motor commands; Schwartz, 2016; Scott, 2004, 2016; Todorov & Jordan, 2002, 2003).
Table 1

Hypotheses and predictions of perspective‐taking, communication, and their interaction effects on pointing trajectories

HypothesisDescriptionEffect on Trajectory
Two‐steps modelPerspective‐taking only determines referent selection and has no impact on movement trajectory. Communication solely affects movement trajectoryCommunication
Leaky two‐steps modelPerspective‐taking determines referent selection thus affects movement trajectory near the end‐position. Communication affects movement trajectory independent from perspective‐takingCommunication, perspective‐taking
Interactive modelPerspective‐taking and communicative demands jointly affect movement trajectory, indicating sensorimotor integration of the two demandsInteraction between Perspective‐taking and Communication
Hypotheses and predictions of perspective‐taking, communication, and their interaction effects on pointing trajectories

Methods

Participants

Thirty‐eight right‐handed healthy participants, naive to the goal of the study, acted as the communicator in the CPT task. All reported normal or corrected‐to‐normal vision gave informed consent in accordance with protocols approved by the local ethics committee (Committee on Research Involving Human Subjects, region Arnhem‐Nijmegen, The Netherlands. Study protocol registration number 3011157.01) and were offered financial compensation for their time. Seven were excluded due to task performance falling below a pre‐defined boundary (<70% trials where the communicator correctly identified the targets), leaving 31 participants (23 females, mean ± SD = 23.7 ± 3.5 years old) included in the final analysis. The excluded participants used inadequate strategies, such as simply flipping the stimuli upside‐down or left‐right instead of engaging in perspective‐taking. Although the rejected participants might have been able to use PT2 if we had explicitly instructed them to do so, they did not spontaneously use PT2 to solve the task. Given the absence of published effect size estimates for perspective‐taking and communicative demands, it was decided to recruit participants until moderate evidence was obtained for (BF = 3) or against (BF = 0.3) main and interaction effects of the experimental factors, with a maximum amount of 40 participants (Dienes, 2016; Lee & Wagenmakers, 2013; Rouder, 2014).

Experimental setup

The experimental setup is illustrated in Fig. 1a and b. A pair of co‐players sat at opposite sides of a round wooden table (radius: 120 cm). The participant played the role of the communicator; the confederate played the role of the addressee. A set of stimuli (four digital items next to each other, each item covering 5 × 5 cm2, 7.5‐cm apart, Fig. 1b) was presented along the center of the main screen (73 × 44 cm2; Fig. 1a‐i and b) via a projector (DELL M410HD, refresh rate: 60 Hz; Fig. 1a‐ii) under the table. A barrier (26 cm from the addressee's eyes, Fig. 1a‐iii) was vertically interposed between the main screen and the addressee. The barrier consisted of 70 Liquid Crystal Display (LCD) tiles (1 × 36 cm2 each) arranged along the whole extent of the glass panel. The transparency of each LCD tile could be independently controlled through a computer (transition time: 3.2 ms). By controlling transparency of each tile, we allowed or prevented the addressee from seeing some of the stimuli on the main screen (manipulation of PT1, Fig. 1b; also see Section 2.3 Experiment Task and Design). The upper part (77 × 14 cm2) of the barrier was covered to prevent eye contact between the co‐players. At either side of the base of the barrier, there were two additional small monitors (15.5 × 8.9 cm2), one facing the communicator (communicator screen; Fig. 1a‐iv) and the other the addressee (addressee screen, Fig. 1a‐v). These small monitors allowed us to manipulate communicative demand by showing or not showing the target on the addressee screen (see Section 2.3 Experiment Task and Design). The experimental setup also included a home key for the communicator (Fig. 1a), aligned with the midline of the communicator and positioned 38 cm from the center of the main screen; a chin rest for the addressee, aligned with the midline of the addressee and positioned 60 cm from the center of the main screen; and a computer mouse (Fig. 1a‐vi), held with the right hand of the addressee, which controlled a cursor displayed on the main screen after the communicator completed the pointing movement and returned to the home key.
Fig 1

Experimental settings and design. (a) Two co‐players sit at the opposite sides of a table. Four stimuli are presented on the main screen (i) through a projector (ii). The communicator (on the left side) needs to point at one of the stimuli, considering which target stimulus can be seen by the addressee through the barrier (iii), or which of the visible stimuli look like the target presented on the communicator screen (iv) when seen from the addressee's perspective. Subsequently, the addressee selects a stimulus from the main screen either based on the communicator's pointing or on the target presented on the addressee screen (v), by mouse click (vi). (vii) is the source of the magnetic field used by the motion tracking system. (b). The curve indicates a schematic profile of velocity of the pointing finger. Each trial is segmented into four movement stages according to predetermined criteria. The photographs illustrate the corresponding configuration of the experimental setup and the pointing response of the communicator. The last two rows show representative Level‐2 perspective‐taking (PT2) trials (in communicative pointing (COM) and non‐ communicative pointing (NCOM) conditions) illustrating the visual scene available to the communicator (upper row) and to the addressee (bottom row). The large black rectangle represents the main screen; the small black rectangle represents the communicator screen (upper row), and the addressee screen (bottom row). The array of shadowed (opaque) and white (transparent) stripes represents the barrier; the dashed lines between the barrier and the main screen (bottom row) illustrate the addressee's line‐of‐sight through the transparent parts of the barrier. For illustration purpose, we use the dashed circle to represents the correct referent to point to in the example, but the circle was not presented in the real experiment. (c) The upper panel shows the visual stimuli with symmetric (Level‐1 perspective‐taking (PT1)) or asymmetric (PT2) features (three other mirrored versions of the stimuli were also adopted in the experiment.). The bottom panel denotes the experimental conditions, arising from using symmetric or asymmetric stimuli (PT1, PT2) and from showing or not showing the target stimulus on the addressee screen (NCOM, COM). More precisely, the communicator screen presents the target in all trials. The addressee screen presents the target in the NCOM condition, and the text “Target not shown” in the COM condition.

Experimental settings and design. (a) Two co‐players sit at the opposite sides of a table. Four stimuli are presented on the main screen (i) through a projector (ii). The communicator (on the left side) needs to point at one of the stimuli, considering which target stimulus can be seen by the addressee through the barrier (iii), or which of the visible stimuli look like the target presented on the communicator screen (iv) when seen from the addressee's perspective. Subsequently, the addressee selects a stimulus from the main screen either based on the communicator's pointing or on the target presented on the addressee screen (v), by mouse click (vi). (vii) is the source of the magnetic field used by the motion tracking system. (b). The curve indicates a schematic profile of velocity of the pointing finger. Each trial is segmented into four movement stages according to predetermined criteria. The photographs illustrate the corresponding configuration of the experimental setup and the pointing response of the communicator. The last two rows show representative Level‐2 perspective‐taking (PT2) trials (in communicative pointing (COM) and non‐ communicative pointing (NCOM) conditions) illustrating the visual scene available to the communicator (upper row) and to the addressee (bottom row). The large black rectangle represents the main screen; the small black rectangle represents the communicator screen (upper row), and the addressee screen (bottom row). The array of shadowed (opaque) and white (transparent) stripes represents the barrier; the dashed lines between the barrier and the main screen (bottom row) illustrate the addressee's line‐of‐sight through the transparent parts of the barrier. For illustration purpose, we use the dashed circle to represents the correct referent to point to in the example, but the circle was not presented in the real experiment. (c) The upper panel shows the visual stimuli with symmetric (Level‐1 perspective‐taking (PT1)) or asymmetric (PT2) features (three other mirrored versions of the stimuli were also adopted in the experiment.). The bottom panel denotes the experimental conditions, arising from using symmetric or asymmetric stimuli (PT1, PT2) and from showing or not showing the target stimulus on the addressee screen (NCOM, COM). More precisely, the communicator screen presents the target in all trials. The addressee screen presents the target in the NCOM condition, and the text “Target not shown” in the COM condition.

Experimental task and design

The CPT task required a communicator to point at one of the four stimuli presented on the main screen. A trial started with the communicator pressing and holding on the home key with the right index finger, making the barrier opaque and a fixation cross to appear on the main screen. After 1000 ms, portions of the barrier turned transparent, and the set of stimuli was displayed on the main screen. Meanwhile, a target was presented on the communicator screen (Fig. 1b, planning). The communicator decided when and where to point, then released the home key and moved the right hand toward the relevant item on the main screen, before returning to the home key. Releasing the home key triggered the stimuli and target to disappear from the main screen and from the communicator screen and the whole barrier to become transparent (Fig. 1b, approach and holding). No constraints were imposed on the communicator's pointing movement, other than starting from and returning to the home key, and pointing without touching the main screen. Once the communicator returned to the home key, the stimuli reappeared on the main screen and the barrier became partially opaque as it was at the onset of the trial (Fig. 1b, returning). The addressee could then select an item from the stimuli array on the main screen with the mouse, triggering the appearance of a gray frame around the selected item, which was also visible to the communicator. After a jittered interval (range: 400 to 600 ms, uniform distribution), the next trial would start. We implemented a 2 × 2 within‐subject factorial design (Fig. 1c). First, the CPT task required the communicator to point for either a communicative (COM) or non‐communicative (NCOM) purpose. The communicator was informed that the addressee was supposed to select an item that matched the target shown on the addressee screen, which was identical to the one shown on the communicator screen. In COM trials, however, the target was shown on the communicator screen but not on the addressee screen. Therefore, the addressee could complete the trial only by observing the communicator's pointing movement; in other words, the communicator was required to instruct the addressee which item should be chosen by pointing. In NCOM trials, the target was shown on both the communicator screen and the addressee screen. In these trials, the communicator was asked to predict which item of the stimuli array would be selected by the addressee. COM and NCOM trials were grouped into four testing blocks of 48 trials each, counter‐balanced according to an “ABBA” scheme and the type of the first block was randomly assigned. At the onset of each block of trials, a written instruction was presented on the main screen (i.e., “Instruct” for COM blocks, “Predict” for NCOM blocks), coupled with a red or green frame at the edge of the main screen throughout the 48 trials of a block. The association of the red/green frame with COM/NCOM blocks was randomized across participants, but it was kept consistent within each participant. Second, the communicator was informed that the addressee's choice determined the success of each trial. Therefore, when deciding which item to point at, the communicator needed to adopt the addressee's perspective. Compliance with this crucial task requirement was ensured through these task features: The communicator needed to consider that the addressee could only select an item that was visible to her, and that matched the target shown on the communicator screen when seen from the addressee's view. Perspective‐taking demand was manipulated through the use of the barrier and of targets with symmetric and asymmetric features. In PT1 trials, the target was symmetric along the horizontal axis (e.g., “O” in Fig. 1c) so that the target appeared the same from the perspectives of both co‐players. However, due to the arrangement of the transparent and opaque elements of the barrier (Fig. 1b), some items of the stimuli array were visible to both co‐players, whereas other items (including at least one duplicate of the target item) were visible only to the communicator. This manipulation was introduced so that the communicator needed to consider which items were visible from the addressee's point of view since only one of the two displayed target items is the correct answer. This task feature ensured communicators engaged in PT1. In PT2 trials, the target was asymmetric (e.g., “Q” in Fig. 1c) so that it appeared different from the communicator's and addressee's perspectives (e.g., a “Q” from the communicator's point of view would correspond to a from the addressee's point of view). Note that the communicator could not be able to identify the correct item by simply mirroring or flipping the target (e.g., “Q” and are not reflected images), and such strategy would lower their task performance. Furthermore, the use of this strategy was discouraged by adding mirrored versions of the stimuli array on the main screen. Thus, in order to select the correct item, the communicator needed to adopt the addressee's perspective to determine how the target would look, that is, to engage in PT2. Each trial started with a new stimuli configuration such that the communicator needed to engage once more in perspective‐taking. Other task details minimized sensory differences between conditions and movement predictability. To match physical features across stimuli arrays, each trial contained both symmetric and asymmetric stimuli. Targets for the four experimental conditions were counterbalanced across the four stimuli locations on the main screen. To minimize stereotyped movements, the items were pseudo‐randomly arranged such that no more than three trials with targets at the same location on the main screen were presented in succession. Although the participants only played the role of communicator in the task, they also experienced and understood the role of the addressee during training blocks (see Section 2.4 Experimental Procedure).

Experimental procedure

The experiment started with training blocks in which participants familiarized themselves with the experimental setup and the roles of the CPT task (started as the role of communicator, then alternated between the roles). The training blocks consisted of at least three blocks of five COM trials and five NCOM trials. Additional blocks were performed until a block was performed with more than 90% accuracy when the participant played the role of communicator. In the training blocks, the players alternated in the roles of communicator and addressee, exchanging positions at the table and experiencing both points of view. During the training blocks, the addressee (played by the confederate) performed the NCOM trials by selecting stimuli on the main screen according to the target shown on the addressee screen (i.e., the players might select a different item based on their own judgement). Then, during the four testing blocks (separated by self‐paced breaks), regardless of the trial conditions, the addressee always selected stimuli indicated by the communicator rather than following the targets presented on the addressee screen. This procedure ensured that the communicator received exactly matched feedback from the addressee's behaviour across the COM and NCOM conditions. Trial‐by‐trial and overall accuracy were presented to the players on the main screen in training blocks, and only overall accuracy was provided at the very end of the testing blocks. Before and after the testing blocks, participants performed 20 pointing movements (baseline trials) to each of the four locations on the main screen where the stimuli array was displayed, without the presence of the addressee, and with the barrier fully opaque. These measurements provided a baseline reference of each participant's pointing movements in a solo setting without experimental manipulations.

Kinematic recording and analysis

We sampled the position and orientation of five sensors at 240 Hz, using an electromagnetic tracking system (LIBERTY, Polhemus). The sensors were attached to the communicator's right index finger (on the second distal phalanx, on the proximal phalanx, and the metacarpal joint), wrist (corpus radii), and to the right shoulder. Kinematic data were processed with Kinemagic Toolbox (Verhagen, Dijkerman, Medendorp, & Toni, 2012, see also https://github.com/Donders‐Institute/kinemagic) using MATLAB R2016b (MathWorks). The data were filtered using a third‐order bidirectional Butterworth low‐pass filter at a cut‐off frequency of 15 Hz. The velocity and acceleration of the index finger were calculated as the first and second derivatives of the position. Position was defined within a Euclidian reference system centered on the home key, with the x‐axis along the communicator's transverse plane, the y‐axis along the communicator's sagittal plane, and the z‐axis along the communicator's coronal plane (the coordinate system is illustrated in the bottom‐left panel of Fig. 1a). Movement onset‐and‐offset was defined using a criterion of index fingertip velocity in the sagittal plane (y‐z plane) over‐or‐under 5 cm/s for at least 200 ms. Hence, each pointing movement was segmented into four stages (Fig. 1b): Planning (or “rest” in some co‐speech gestures literature; Kita et al., 1998; McNeill, 1992)—the period between presentation of the stimuli and movement onset; Approach (“stroke”)—the period between movement onset and arrival near the location of the selected stimulus; Holding (“hold”)—the period between arrival near the selected stimulus and onset of the return movement away from the selected stimulus; and Returning (“retraction”)—from the onset of the return movement until the time the home key was pressed. Only the first three stages were analyzed since the returning movement does not concern referential pointing, and it was constrained by the additional requirement to press the home key. We examined temporal and spatial aspects of kinematic data. Temporally, we focused on the duration of two movement stages: planning and holding. The spatial features of the pointing movements were quantified as average displacement and variability of the trajectories in three‐dimensional space during the approach stage as indices of movement magnitude and movement correction, respectively (see Section 1 Introduction). To compare the temporal development of spatial features of the movement across participants, each approach trajectory was first normalized into 100 time points on a participant level with cubic‐spline interpolation. We then extracted time series of average displacement and variability of the trajectories relative to the corresponding values in the baseline movement (performed in the solo setting). Specifically, for each time point on the time‐normalized trajectories, we estimated an ellipsoid along the eigenbases of the covariance matrix of the trajectories. The center of the ellipsoid at each time point provided the average position of the pointing finger; the volume of the ellipsoid was computed by taking the 95% confidence interval of the Mahalanobis distance according to a Chi‐square distribution. Based on the ellipsoid at each time point, displacement was calculated as the Euclidian distance from the home key to the center of the ellipsoid; and trajectory variability was calculated as the volume of the ellipsoid. Trials with incorrect responses (i.e., pointing movement toward an incorrect item) were excluded from the kinematic analysis (mean: 6.45%; range: 0.52%–17%). One participant was excluded from the trajectory analysis because of invalid baseline measuring.

Statistical analysis

To quantify evidence for or against the null hypothesis, we used Bayes Factors (BFs). Stage durations were included into Bayesian repeated‐measure analysis of variance (anovas) in jasp 0.9.2 (JASP Team, 2019) with two factors (perspective‐taking and communicative demands) with two levels each. We used a multivariate Cauchy prior with a fixed effects scale factor of r = .50, a random effects scale factor of r = 1, and a scale covariate of r = .35 (Rouder, Morey, Speckman, & Province, 2012). Post hoc comparisons were based on the default t test with a Cauchy (0, r = .71) prior (Rouder, Speckman, Sun, Morey, & Iverson, 2009). Statistical inferences on the time courses of displacement and trajectory variability were made using non‐parametric cluster‐based permutation tests (Maris & Oostenveld, 2007), in order to control for multiple comparisons in the context of data with complex spatiotemporal dependencies. The cluster‐forming threshold was α = .05 (two‐tailed). P‐values corrected for multiple comparisons (p MC) were computed by ranking the test statistics among sample values from 1000 Monte Carlo runs. We used post hoc exploration of significant effects to evaluate specific main and interaction effects using a Bonferroni procedure (p corr) to correct for multiple comparisons in those pairwise contrasts. These tests were performed with matlab R2016b.

Results

Temporal effects of perspective‐taking and communicative demands on the pointing movements

We performed a Bayesian repeated‐measure anovas on the duration of three‐movement stages (planning, approach, and holding). Consistent with our predictions on planning and holding times (see Section 1 Introduction), the BFs suggested extreme evidence for a main effect of perspective‐taking on planning time (BF = 1.264e+21, Fig. 2a) and a main effect of communication on holding time (BF = 1.960e+3, Fig. 2c). More precisely, participants took longer to initiate a movement during PT2 trials than during PT1 trials (BF = 2.864e+15; PT2: 4.420 s; PT1: 2.459 s) and held their finger on the end‐position location for a longer time at the end of communicative than non‐communicative movements (BF = 151.400; COM: 0.249 s; NCOM: 0.209 s). There was no strong evidence to conclude that approach time differed across conditions (Table 2).
Fig 2

Temporal features of the pointing movements. Durations of (a) planning, (b) approach, and (c) holding stage. The curved areas illustrate kernel density estimates of the data distribution. The colored dots are jittered raw data and the black dots represent the group mean. The boxes indicate the interquartile range (IQR) and the whiskers are maximum and minimum with 1.5 IQR. The task conditions are coded with the color of light blue, dark blue, light red, and dark red for PT1 + NCOM, PT1 + COM, PT2 + NCOM, and PT2 + COM, respectively.

Table 2

Bayes factors (BFs) of the effects of perspective‐taking and communicative demands and their interaction on the duration of each movement stage (against the null hypothesis)

Effect (Duration)PlanningApproachHolding
Communication effect0.2852.0361.960e+3
Perspective‐taking effect1.264e+214.4680.272
Interaction effect0.2350.2870.297

Note. BFs > 3 or < 0.3 provide moderate evidence and > 10 or < 0.1 strong evidence in favor or against the presence of an experimental effect (Lee & Wagenmakers, 2013).

Temporal features of the pointing movements. Durations of (a) planning, (b) approach, and (c) holding stage. The curved areas illustrate kernel density estimates of the data distribution. The colored dots are jittered raw data and the black dots represent the group mean. The boxes indicate the interquartile range (IQR) and the whiskers are maximum and minimum with 1.5 IQR. The task conditions are coded with the color of light blue, dark blue, light red, and dark red for PT1 + NCOM, PT1 + COM, PT2 + NCOM, and PT2 + COM, respectively. Bayes factors (BFs) of the effects of perspective‐taking and communicative demands and their interaction on the duration of each movement stage (against the null hypothesis) Note. BFs > 3 or < 0.3 provide moderate evidence and > 10 or < 0.1 strong evidence in favor or against the presence of an experimental effect (Lee & Wagenmakers, 2013).

Spatial effects of perspective‐taking and communicative demands on the pointing movements

To compare the spatial dynamics of the pointing movements, we normalized the raw trajectories (Fig. 3a) with respect to equal time intervals per participant. There are three main findings. First, in line with previous findings, displacement of pointing trajectories showed an early effect of communicative demand (from 9% until 100% of the approach time, p MC < 0.001, Fig. 3b), with longer trajectories in COM than in NCOM trials (t 29 = 7.531, p corr < 0.001, Cohen's d = 1.375. A post hoc analysis revealed that the COM trajectories ended further away from the starting position, compared to the NCOM trajectories (BF = 1.460e+7; COM: 38.14 cm; NCOM: 38.04 cm). Second, there was also a main effect of perspective‐taking (from 23% until 100% of the approach time, p MC < .001, Fig. 3b), with longer trajectories in PT2 than in PT1 trials (t 29 = 8.394, p corr < .001, Cohen's d = 1.533). A post hoc analysis revealed that the PT2 trajectories ended further away from the starting position, compared to the PT1 trajectories (BF = 1.019e+13; PT2: 38.18 cm; PT1: 38.00 cm). Third, in favor of Hypothesis 3 (“Interactive model”, see Section 1 Introduction), there was an under‐additive interaction between perspective‐taking and communicative demands (from 40% until 85% of the approach time, p MC < 0.001, t 29 = −5.404, p corr < .001; Cohen's d = .987; Fig. 3b), with a smaller effect of perspective‐taking demand (longer trajectories in PT2 than PT1) during COM trials (t 29 = 4.670, p corr = 0.005, Cohen's d = 0.853) than during NCOM trials (t 29 = 17.344, p corr < 0.001, Cohen's d = 3.167). A Bayesian one‐sample t‐test suggests that this interaction effect occurred after the initial acceleration phase of the pointing movement (BF = 6521), that is, after the pointing finger reached peak velocity at approximately 35% of the approach time. Moreover, these trajectory displacement effects could not be attributed to changes in trajectory variability: The trajectories remained comparable in that respect across conditions throughout the approach time (p MC > 0.05, Fig. 3c).
Fig 3

Spatial features of the pointing movements. (a) Illustration of raw pointing trajectories in each experimental condition, both in 3D view (Panel a1) and over the horizontal plane (Panel a2). Each curve represents a randomly chosen trial from a participant. (b) Time courses of the difference in trajectory displacement between task‐related and baseline movements, taken as reference (dashed line). Each curve represents group averages; shaded areas indicate ±1 standard error of the mean (SEM) over participants. The bars along the x‐axis indicate significant effects of communicative demand (black bar; between 9% and 100% of the approach time, p MC < .001); perspective‐taking (dark gray bar; between 23% and 100% of the approach time, p MC < .001); and interaction between perspective‐taking and communicative demand (light gray bar; between 40% and 85% of the approach time, p MC < .001). (c) Time courses of the difference in trajectory variability between task‐related and baseline movements.

Spatial features of the pointing movements. (a) Illustration of raw pointing trajectories in each experimental condition, both in 3D view (Panel a1) and over the horizontal plane (Panel a2). Each curve represents a randomly chosen trial from a participant. (b) Time courses of the difference in trajectory displacement between task‐related and baseline movements, taken as reference (dashed line). Each curve represents group averages; shaded areas indicate ±1 standard error of the mean (SEM) over participants. The bars along the x‐axis indicate significant effects of communicative demand (black bar; between 9% and 100% of the approach time, p MC < .001); perspective‐taking (dark gray bar; between 23% and 100% of the approach time, p MC < .001); and interaction between perspective‐taking and communicative demand (light gray bar; between 40% and 85% of the approach time, p MC < .001). (c) Time courses of the difference in trajectory variability between task‐related and baseline movements. We also implemented a post hoc analysis to quantify differences between solo pointing and CPT pointing. We contrasted the trajectories of the four task conditions to those of baseline trials (solo pointing without an addressee). Task‐related movements had larger displacements starting around 30% of the approaching time and afterward (39%, 32%, 32%, and 26% for NCOM+PT1, NCOM+PT2, COM+PT1, and COM+PT2, respectively). The variability of task‐related trajectories was above baseline from 38%, 40%, 29%, and 25% of the approaching time and afterward in NCOM+PT1, NCOM+PT2, COM+PT1, and COM+PT2 condition, respectively; Monte Carlo p‐values were all smaller than .001. These results revealed that the mere presence of an addressee had a substantial influence on the organization of the pointing movement, encouraging communicators to move further away from the starting position. We present the group‐averaged differential length and differential variability (relative to the baseline per condition) of movement trajectories in Fig. 3b and c. Fig. 3a provides an example of raw single‐trial trajectories to the four stimuli locations in each of the four conditions and at baseline.

Discussion

This study tests whether and how communicative and perspective‐taking demands are integrated at the sensorimotor level during referential pointing. Communicators changed temporal and spatial features of their pointing movements according to the communicative demand of the task and the spatial perspective of the addressee. The CPT task successfully reproduced well‐documented effects in the existing literature on perspective‐taking and communicative pointing. Communicators took longer to initiate their movement when inferring how the addressee perceived the referents (PT2) than when discriminating what could be seen by the addressee (PT1; Galati & Avraamides, 2013; Galati, Dale, & Duran, 2019; Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Samson et al., 2010; Surtees et al., 2016). Furthermore, communicators held their finger near the selected stimulus for longer when their pointing movement was required to communicate the target identity to the addressee (Becchio et al., 2010; Cleret de Langavant et al., 2011; Krishnan‐Barman et al., 2017; Murillo Oosterwijk et al., 2017; Peeters et al., 2015). The novel finding of this study is that the sensorimotor system controls the trajectory of referential pointing by integrating the communicative demands of the task and the spatial perspective of the addressee (“interactive model”). Both communicative demand and PT2 processes influenced motor responses in a similar manner, that is, lengthened trajectories from an early stage of the pointing movement. Crucially, the effect of PT2 on pointing trajectories decreased when communicators conveyed instructive information to addressees (communicative pointing) as compared to when communicators predicted addressees’ response (non‐communicative pointing). This evidence for the “interactive model” argues against the alternative possibility of a functional independence between (i) identification of a relevant stimulus according to perspective‐taking demands, and (ii) movement specification according to communicative demand. This possibility would have resulted in an effect of perspective‐taking limited to the planning stage (“two‐steps model”) or independent effects of perspective‐taking and communicative demands during movement execution (“leaky two‐steps model”).

Effects of communicative demand

The CPT task effectively modulated the communicative demands experienced by the communicators; they held their pointing finger near the end‐position for longer when they thought the addressee had to identify the correct referent by means of their pointing movements. This observation is consistent with findings from other referential pointing studies (Cleret de Langavant et al., 2011; Murillo Oosterwijk et al., 2017; Peeters et al., 2015; Winner et al., 2019), as well as naturalistic communicative interactions (Clark & Murphy, 1982; Hilbrink, Gattis, & Levinson, 2015; Levinson, 2016; Richardson, Dale, & Kirkham, 2007). A longer holding time allows the addressee to accumulate more sensory evidence about the end location of the pointing finger (Bangerter, 2004; Sacheli, Tidoni, Pavone, Aglioti, & Candidi, 2013; Sartori, Becchio, Bara, & Castiello, 2009; Vesper, Schmitz, & Knoblich, 2017). Furthermore, this study also confirms that increasing communicative demand led participants to generate longer movement trajectories, starting shortly after movement onset (Murillo Oosterwijk et al., 2017; Winner et al., 2019). A longer holding time and exaggerated trajectory provide a cue that, by virtue of it being longer than necessary and thus instrumentally dysfunctional, ostensively marks the action as communicative (Cleret de Langavant et al., 2011; Cooney, Brady, & McKinney, 2018; Gredebäck & Melinder, 2010; Murillo Oosterwijk et al., 2017; Peeters et al., 2015; Winner et al., 2019). Details of the experimental design and empirical observations exclude a number of alternative interpretations. For instance, the effects of communicative demand were not a by‐product of increased trial‐by‐trial variability in the movement trajectories or systematic variations in the end location of the pointing finger. Neither could the effects be attributed to feedback from the addressee, as the addressee's responses were matched across conditions, that is, the confederate chose the item indicated by the communicators regardless of the task condition (see Section 2.4. Experimental Procedure). Moreover, the addressee's face was occluded by the barrier so that the communicators were unable to make use of social cues, like gaze or facial expressions (see Section 2.2. Experimental Setup and Fig. 1b). It is important to emphasize that the effect of communicative demand on movement trajectories occurred over and above the systematically longer trajectories and larger variance observed in the joint than in the solo‐task settings, across experimental conditions. Taken together, these observations confirm that communicative demand leads to kinematic re‐organization of pointing movements (Chu & Hagoort, 2014; Kita & Özyürek, 2003), generating a spatially exaggerated movement (Becchio et al., 2010; Cleret de Langavant et al., 2011; Krishnan‐Barman et al., 2017; Murillo Oosterwijk et al., 2017; Trujillo et al., 2020). These findings raise the possibility that communicative demand leads individuals to change their motor control policy. More precisely, studies of spatially guided movements have repeatedly shown that participants tested in isolation (as in the solo condition of this study) control the end‐position of the pointing finger (Nashed et al., 2012; Keyser et al., 2017; Todorov et al., 2003) and/or minimize spatial variability of the finger position (Harris & Wolpert, 1998; Wolpert & Ghahramani, 2000; Wolpert, Ghahramani, & Jordan, 1994). In contrast, during referential pointing, communicators might focus on trajectory control, leading to a tight integration between communicative demands and sensorimotor trajectory control (Chu & Hagoort, 2014; Murillo Oosterwijk et al., 2017; Winner et al., 2019).

Effects of perspective‐taking

When solving the CPT task, participants took longer to start their movement during trials that required them to infer how the addressee perceived the referent. This observation is in line with the established notion that PT2 requires additional cognitive operations as compared to PT1, for example, inhibition of one's own perspective and transformation from an egocentric to allocentric frames of reference (Galati & Avraamides, 2013; Galati et al., 2019; Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Samson et al., 2010; Surtees et al., 2016). The current study goes further in providing novel evidence of how visuospatial perspective‐taking affects spatial features of movement performance. Trials requiring PT2 evoked longer trajectories than trials requiring PT1. The effect started during the initial acceleration phase of the movement, an indication that perspective‐taking influences the forward kinematics of a movement (Gallivan et al., 2018; Körding & Wolpert, 2004; Todorov et al., 2003). The effect of PT2 on pointing trajectory cannot be explained by trajectory adjustments driven by visual feedback of the selected stimulus and its relation to the point of view of the addressee. During the performance of the pointing movements, both target and stimuli disappeared from the communicator screen and the main screen; and the whole barrier turned transparent. The results also revealed that variability of the pointing trajectories, indexing movement correction, remained comparable across levels of visuospatial perspective‐taking. These observations suggest that the difference in trajectory displacement is due to perspective‐taking effects on the initial forward kinematics of the pointing movements, computed during the planning phase before movement onset. It could be argued the effect of PT2 on pointing trajectory has a communicative goal, for example, participants ostensively mark their pointing movements for the addressee. However, this possibility does not explain why perspective‐taking demand would lead participants to selectively deploy an ostensive marker on the movement trajectory, but then remove it from the holding time, where there are no differences between the two perspective‐taking conditions. This inconsistency in PT2 effects over the approach and holding phases of the movement does not match with the effects of communicative demands, where participants ostensively tag both phases of the movement. Studies on the impact of different frames of reference in coding spatial targets of a movement offer an alternative interpretation of the effect of PT2 on pointing trajectory. Specifically, it is possible that the communicators’ coding of the location of the selected stimulus was influenced by the use of different frames of reference in perspective‐taking conditions. More precisely, an allocentric frame of reference in PT2 trials. An allocentric frame is required for communicators estimating how the stimuli were perceived by the addressee. In contrast, communicators could estimate whether the objects could be seen by the addressee (PT1 trials) by using an egocentric frame of reference (Crawford, Henriques, & Medendorp, 2011; Flavell, Everett, Croft, & Flavell, 1981; Hart & Moore, 1973; Kessler & Rutherford, 2010; Michelon & Zacks, 2006; Salatas & Flavell, 1976; Surtees et al., 2013). The sensorimotor system uses those frames of reference when planning and executing a movement (Carrozzo, Stratta, McIntyre, & Lacquaniti, 2002; Chen, Byrne, & Crawford, 2011, 2014; Fiehler, Wolf, Klinghammer, & Blohm, 2014; Proulx, Todorov, Taylor Aiken, & Sousa, 2016). In this study, participants aimed their movements at systematically different end‐positions in PT1 and PT2, with comparable levels of movement correction (no difference in movement variability), an indication of different initial forward kinematics in the two perspective‐taking conditions (Gallivan et al., 2018; Körding & Wolpert, 2004; Todorov et al., 2003). Therefore, it could be suggested that the effect of PT2 on pointing trajectory might arise from differences in the estimation of the stimulus location due to the different frames of reference used to solve PT1 and PT2 problems.

Interaction between perspective‐taking and communicative demands

The main finding of this study is the under‐additive interaction between perspective‐taking and communicative demands, indicating that the sensorimotor system makes use of the two types of information in an integrated manner. Communicators lengthened the pointing trajectories after implementing PT2 relative to PT1 but proportionally less when they engaged in intentional communication, that is, when knowing that the addressee's response depended on their pointing movements. The interaction effect occurred after the initial acceleration phase of the movement, an indication that the interaction is likely to reflect control processes occurring during movement execution (Gallivan et al., 2018; Körding & Wolpert, 2004; Todorov et al., 2003). The under‐additive nature of the interaction suggests a hierarchical order in how perspective‐taking and communicative information are integrated during sensorimotor control. In communicative pointing, the motor controller could privilege the trajectory constraints determined by the distal goal of the movement, that is, to produce a communicative signal that needs to be correctly interpreted by an addressee. Over‐lengthened trajectories, determined by a linear summation of the factors shaping the main effects of perspective‐taking and communicative demands, reduce the legibility of the movement (Winner et al., 2019) and increase the noise level in the sensorimotor system (Faisal et al., 2008; Harris & Wolpert, 1998; Schwartz, 2016; Scott, 2004, 2016; Wolpert et al., 2003), thus increasing the ambiguity experienced by the addressee when interpreting the observed movement. A motor controller privileging the communicative demands of the movement would imply a reduction of the weight allocated to visuospatial parameters estimated during PT2, that is, over‐estimation of target distance observed in the main effect of perspective‐taking. This hypothetical change of control parameters (not only the end‐position but also the entire trajectory) would provide communicators with the means to finely adjust the pointing movement for the addressee and to follow the Gricean maxim of quantity (Grice, 1975), that is, to generate behaviors as informative as required rather than overly exaggerated displays. These findings indicate that the sensorimotor system can actively integrate different types of conceptual knowledge. In summary, the effect of perspective‐taking on trajectory length is present in non‐communicative situations, but less so in communicative ones, because then it is overruled by the communicative demands of the situation. Other empirical observations rule out the possibility that the interaction is a generic effect of task difficulty, either from perspective‐taking demand before target selection or from communicative demand at the end of pointing movements (holding stage); the interaction effect was limited to a spatial feature of the pointing movements (trajectory displacement), in the context of statistical evidence against similar interactions on temporal features of those movements, throughout planning, approach, and holding phases of the movements. It is also unlikely that the interaction is due to a ceiling effect caused by biomechanical constraints (e.g., the participants could not adequately reach the intended stimulus given their arm length), which should increase as the movement approaches the intended stimulus; in that case, we should have observed stronger interaction effects in the later section of the movement. However, the interaction effect occurred in the middle section of the movement.

Interpretational issues

The experimental design relies on a previously established operationalization of visuospatial perspective‐taking (Surtees, Butterfill, & Apperly, 2013, 2012, 2016) and did not include additional conditions to demonstrate that participants considered the distance between an observer and the visual stimulus (PT1) or the angular disparity between a participant and an observer's viewpoints (PT2). The findings of the perspective‐taking manipulation are consistent with previous studies that have differentiated PT2 from PT1 but could be tested further by manipulating mental rotation demands in perspective‐taking. However, we would like to emphasize that the experimental setup ensures communicators used PT1 and PT2 in the corresponding trials. Namely, due to the varying arrangement of transparent and opaque elements of the barrier, during PT1 trials, the communicator needed to consider which of multiple potential stimuli were visible from the addressee's point of view. Simply pointing to one of the stimuli visible to the communicator was not a viable option. Similarly, during PT2 trials, we used targets that appeared different from the communicator's and addressee's perspectives, and that were intermixed with left‐right‐mirrored versions. Accordingly, during PT2 trials, the communicator needed to adopt the addressee's perspective to determine how the target would look. Simply pointing to the mirror copy of the target was not a viable option (see 2.1. Participants and 2.3. Experimental Task and Design). It can be seen that participants produced systematically longer trajectories in the joint than in the solo‐task settings, across experimental conditions. This observation extends the notion that joint task settings enhance perspective‐taking beyond the selection of the correct referent (Freundlieb, Kovács, & Sebanz, 2016, 2017), showing that joint‐task settings change how participants spatially organize their movements (Ciardo, Lugli, Nicoletti, Rubichi, & Iani, 2016; Coello, Quesque, Gigliotti, Ott, & Bruyelle, 2018; Creem‐Regehr et al., 2013; Ondobaka, Kilner, & Friston, 2017; Proffitt & Linkenauger, 2013; Proulx et al., 2016). More generally, the current findings raise the possibility that movements with different distal goals use different control policies, providing a structural ground to descriptive differentiations between instrumental, coordinative, and communicative functions of body movements (Vesper & Sevdalis, 2020). It could be argued that because the task involved several repetitions of the same movements, the participants might have produced stereotypical responses that are qualitatively different from the occasional referential pointing movements occurring during daily communicative interactions. However, participants consistently produced movements that were sensitive to the communicative demands. Specifically, participants produced more ostensive movements during communicative than non‐communicative trials throughout the experiment. To understand the mechanistic implementation of human communication, it is fundamental to use experimental settings that allow pairs of individuals to coordinate on a trial‐by‐trial basis (Redcay & Schilbach, 2019; Wheatley, Boncz, Toni, & Stolk, 2019) while retaining the ability to quantify their behaviour under controlled and reproducible conditions. However, it remains to be seen whether the current findings, obtained in the context of highly controlled experimental conditions, can be generalized to settings involving the same fundamental phenomenon, but more communicative options, as during natural dialogue with co‐speech gestures. It could also be argued that the use of a confederate might introduce biases in the findings, similar to those reported during face‐to‐face dialogue (Kuhlen & Brennan, 2013). However, the current settings involve non‐verbal stereotyped responses with considerably fewer degrees of freedom than multimodal utterances in spoken dialogue. Furthermore, there was no demand on the confederates to deceive the participants, nor possibilities for the confederates to provide their response in a different manner. Accordingly, the minimal risks of using a confederate in the current circumstances should be weighed against the interpretational advantage of excluding differences in the behaviour of the addressee as a source of experimental effects.

Conclusion

This study provides evidence that the sensorimotor system makes use of communicative and perspective‐taking information in planning and controlling pointing trajectories, suggesting that during movement execution, communicative constraints are prioritized over the spatial biases evoked by perspective‐taking. The finding contributes to understanding the interface between sensorimotor and conceptual abilities in humans, showing that communicative and perspective‐taking knowledge penetrates the sensorimotor system. Put differently, sensorimotor processes are not a simple production mechanism, downstream to and encapsulated from cognitive decisions. This observation has consequences for models of human communication. It implies that the format of communicative and perspective‐taking knowledge needs to be commensurate with the movement dynamics controlled by the sensorimotor system. The current study highlights the value of bridging largely separated research traditions in sensorimotor control and human communicative behaviour, including co‐speech gestures and other referential actions, and it opens the way for studying neural mechanisms integrating other agents’ knowledge into communicators’ behaviour under controlled yet ecologically relevant experimental conditions.

Open Research Badges

This article has earned Open Data and Open Materials badges. Data are available at https://osf.io/h8yad and materials are available at https://osf.io/z2pq4.
  67 in total

1.  Bayesian decision theory in sensorimotor control.

Authors:  Konrad P Körding; Daniel M Wolpert
Journal:  Trends Cogn Sci       Date:  2006-06-27       Impact factor: 20.229

2.  Bayesian t tests for accepting and rejecting the null hypothesis.

Authors:  Jeffrey N Rouder; Paul L Speckman; Dongchu Sun; Richard D Morey; Geoffrey Iverson
Journal:  Psychon Bull Rev       Date:  2009-04

3.  Three-dimensional transformations for goal-directed action.

Authors:  J Douglas Crawford; Denise Y P Henriques; W Pieter Medendorp
Journal:  Annu Rev Neurosci       Date:  2011       Impact factor: 12.449

4.  Direct and indirect measures of Level-2 perspective-taking in children and adults.

Authors:  Andrew D R Surtees; Stephen A Butterfill; Ian A Apperly
Journal:  Br J Dev Psychol       Date:  2011-10-14

5.  Perceptual distortion contributes to the curvature of human reaching movements.

Authors:  D M Wolpert; Z Ghahramani; M I Jordan
Journal:  Exp Brain Res       Date:  1994       Impact factor: 1.972

Review 6.  Movement: How the Brain Communicates with the World.

Authors:  Andrew B Schwartz
Journal:  Cell       Date:  2016-03-10       Impact factor: 41.582

7.  Electrophysiological and Kinematic Correlates of Communicative Intent in the Planning and Production of Pointing Gestures and Speech.

Authors:  David Peeters; Mingyuan Chu; Judith Holler; Peter Hagoort; Aslı Özyürek
Journal:  J Cogn Neurosci       Date:  2015-08-18       Impact factor: 3.225

Review 8.  Decision-making in sensorimotor control.

Authors:  Jason P Gallivan; Craig S Chapman; Daniel M Wolpert; J Randall Flanagan
Journal:  Nat Rev Neurosci       Date:  2018-09       Impact factor: 34.870

9.  The Two Forms of Visuo-Spatial Perspective Taking are Differently Embodied and Subserve Different Spatial Prepositions.

Authors:  Klaus Kessler; Hannah Rutherford
Journal:  Front Psychol       Date:  2010-12-06

10.  Flexible spatial perspective-taking: conversational partners weigh multiple cues in collaborative tasks.

Authors:  Alexia Galati; Marios N Avraamides
Journal:  Front Hum Neurosci       Date:  2013-09-26       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.