Bastien Berret1,2,3, Adrien Conessa1,2, Nicolas Schweighofer4, Etienne Burdet4. 1. Université Paris-Saclay CIAMS, Orsay, France. 2. CIAMS, Université d'Orléans, Orléans, France. 3. Institut Universitaire de France, Paris, France. 4. University of Southern California, Los Angeles, California, United States of America.
Abstract
Human movements with or without vision exhibit timing (i.e. speed and duration) and variability characteristics which are not well captured by existing computational models. Here, we introduce a stochastic optimal feedforward-feedback control (SFFC) model that can predict the nominal timing and trial-by-trial variability of self-paced arm reaching movements carried out with or without online visual feedback of the hand. In SFFC, movement timing results from the minimization of the intrinsic factors of effort and variance due to constant and signal-dependent motor noise, and movement variability depends on the integration of visual feedback. Reaching arm movements data are used to examine the effect of online vision on movement timing and variability, and test the model. This modelling suggests that the central nervous system predicts the effects of sensorimotor noise to generate an optimal feedforward motor command, and triggers optimal feedback corrections to task-related errors based on the available limb state estimate.
Human movements with or without vision exhibit timing (i.e. speed and duration) and variability characteristics which are not well captured by existing computational models. Here, we introduce a stochastic optimal feedforward-feedback control (SFFC) model that can predict the nominal timing and trial-by-trial variability of self-paced arm reaching movements carried out with or without online visual feedback of the hand. In SFFC, movement timing results from the minimization of the intrinsic factors of effort and variance due to constant and signal-dependent motor noise, and movement variability depends on the integration of visual feedback. Reaching arm movements data are used to examine the effect of online vision on movement timing and variability, and test the model. This modelling suggests that the central nervous system predicts the effects of sensorimotor noise to generate an optimal feedforward motor command, and triggers optimal feedback corrections to task-related errors based on the available limb state estimate.
Spatial and temporal regularities in human motion suggest that the neural control of movement involves a planning stage [1, 2]. Evidence for motor planning has been provided in behavioural experiments [3, 4] and through the observation of neural processes prior to movement generation [5-7]. Among the planned aspects of movement, the timing (i.e. speed and duration) and trial-by-trial variability are important determinants of successful actions [8]. However, the principles according to which the central nervous system (CNS) may determine these critical features is not well explained by existing models.The currently dominating theory of motor control, stochastic optimal (feedback) control (SOC) [9-12], can explain the coordination of the degrees-of-freedom of the sensorimotor system, the structure of trial-by-trial variability or the reactive behavior to external perturbations [13-16]. However, SOC does not account well for the timing of self-paced reaching movements. As with deterministic optimal control (DOC) models (e.g. [17, 18]) the costs minimised in SOC models, effort and error, decrease monotonically with increasing movement duration, thereby predicting infinitely slow visually-guided movements. Fig 1A illustrates this issue for the SOC model of [19] where both motor and observation noise are considered. Interestingly, when SOC is used to model movements without vision by increasing observation noise to reflect a degraded hand state estimate, a finite optimal duration can be obtained because endpoint variance now increases with duration. Therefore, SOC with large enough observation noise may determine the timing and variability of movements without vision (Fig 1B), but the same principle cannot be used directly for visually-guided movements.
Fig 1
Expected costs and endpoint variance for the SOC model of [19] for simulated movements with (A) and without (B) vision.
A 10-cm long reaching movement of a point mass model of an arm is simulated. These simulations rely on the extended linear-quadratic-Gaussian framework considering multiplicative (signal-dependent) and additive (constant) motor noise as well as additive observation noise. A. Simulations with a standard observation noise corresponding to a visually-guided movement, as proposed in the original model of [19]. The expected costs for different movement durations were estimated using the Monte Carlo method (100,000 samples). This model fails to predict a finite movement duration because the optimal expected effort and total cost (sum of effort and terminal error costs) monotonically decrease with duration and plateau. The positional endpoint variance (gray trace) can also be seen to decrease and plateau to a value which mainly corresponds to that of visually-guided movements. B. Simulations with a large observation noise (noise in A multiplied by 10), corresponding to movements without vision. In this case, an optimal duration can be determined as the minimum of the total cost (indicated by a black vertical dotted line).
Expected costs and endpoint variance for the SOC model of [19] for simulated movements with (A) and without (B) vision.
A 10-cm long reaching movement of a point mass model of an arm is simulated. These simulations rely on the extended linear-quadratic-Gaussian framework considering multiplicative (signal-dependent) and additive (constant) motor noise as well as additive observation noise. A. Simulations with a standard observation noise corresponding to a visually-guided movement, as proposed in the original model of [19]. The expected costs for different movement durations were estimated using the Monte Carlo method (100,000 samples). This model fails to predict a finite movement duration because the optimal expected effort and total cost (sum of effort and terminal error costs) monotonically decrease with duration and plateau. The positional endpoint variance (gray trace) can also be seen to decrease and plateau to a value which mainly corresponds to that of visually-guided movements. B. Simulations with a large observation noise (noise in A multiplied by 10), corresponding to movements without vision. In this case, an optimal duration can be determined as the minimum of the total cost (indicated by a black vertical dotted line).Several ad hoc solutions have been proposed to circumvent this issue. In several models considering sensorimotor noise, duration was selected as the minimum time to match a desired endpoint variance related to target’s width, based on the speed-accuracy trade-off underlying visually-guided movements [8, 20–22]. Alternatively, a number of studies have assumed a “cost of time” (reflecting neuroeconomical processes related to decision-making and explicitly penalizing duration) to explain the preferred timing of movement [23-30].However, the preferred movement timing may be predicted without requiring an ad hoc solution. In particular, the results of [31] suggest that the motor noise, with its constant and signal dependent components, is a relevant factor to determine this characteristic of motion planning. Specifically, the preferred duration of movements performed without vision was found to be longer than the minimum variance duration, thereby suggesting that movement timing was determined from a neuromechanical principle based on a trade-off between effort and variance in the presence of signal-dependent and constant motor noise. Such optimality principle could explain the stereotyped durations and trajectories of saccades [32], but its relevance for arm reaching has not been tested in particular for visually-guided movements.Because SOC cannot be used for this purpose (see Fig 1A), here we develop a new computational model to predict the timing and variability of arm pointing movements carried out with complete or degraded sensory feedback (e.g. when vision of the hand is prevented) from neuromechanical factors only. This stochastic feedforward-feedback control (SFFC) model assumes that the motor command comprises a feedforward and a feedback components. The feedforward component is computed using the stochastic optimal open-loop control (SOOC) framework, which was initially developed to account for the planning of mechanical impedance via muscle co-contraction [33, 34]. This feedforward command yields an expectation about a timed trajectory. The feedback component is then computed using the linear SOC framework from a local approximation of the task dynamics, which allows triggering motor corrections in reaction to deviations from the goal, based on an estimation of the system’s state from the available sensory information and internal predictions. As a result, the proposed model merges the main precepts of influential models highlighting either the role of feedforward-only or feedback-only control [8–10, 13, 17–19].Predictions from the SFFC model are first tested by simulating arm reaching movements carried out without visual feedback and comparing the results with the available experimental results of [31, 35], and [36]. Second, an experiment was conducted to analyse the timing and variability of movements performed with and without online visual feedback of the hand. The SFFC predictions for movements in these two conditions are then compared to these new experimental data.
Results
Stochastic feedforward-feedback control model
In the proposed model, the actual motor command is made of a feedforward component (i.e. determined prior to movement execution) and a feedback component (i.e. determined throughout movement execution based on an estimation of the current state) to correct task-related errors as illustrated in Fig 2. This is a classical approach in optimal control theory (e.g. see [37, 38]). However, feedforward control is usually associated with deterministic systems and feedback control with stochastic systems. In the approach presented here, the feedforward command is optimized for the system’s stochasticity (i.e. presence of both signal-dependent and constant motor noise) as in [8] and [33].
Fig 2
Scheme of stochastic optimal feedforward-feedback control (SFFC).
A feedforward command u(t) is formed by the CNS based on prior knowledge about the task dynamics represented by f and G. A representation of the associated expected system’s trajectory is also established, which allows building a local approximation of the task dynamics and working in terms of state/control deviations (z and v respectively) during movement execution. This is done by setting the matrices and . An estimate of the current state deviation is computed from multisensory information y. This allows triggering a feedback command online to correct task-relevant errors caused by unexpected internal and external perturbations due e.g. to motor noise or external forces. In this scheme, the actual motor command u(t) + v is the sum of the feedforward and feedback commands. The matrices L(t) and K(t) denote the optimal filter and feedback gains respectively, g is the output function and in the local approximation. The random processes and are implemented here as Brownian. D(t) is an observation noise matrix, the magnitude of which can be increased to simulate the absence of vision. The vector h denotes the deviation from the sensory prediction. The terms Cff and Cfb refer to the cost functions that determine the optimal signature of the feedforward and feedback commands. The definition and meaning of all the variables are given in the Results section.
Scheme of stochastic optimal feedforward-feedback control (SFFC).
A feedforward command u(t) is formed by the CNS based on prior knowledge about the task dynamics represented by f and G. A representation of the associated expected system’s trajectory is also established, which allows building a local approximation of the task dynamics and working in terms of state/control deviations (z and v respectively) during movement execution. This is done by setting the matrices and . An estimate of the current state deviation is computed from multisensory information y. This allows triggering a feedback command online to correct task-relevant errors caused by unexpected internal and external perturbations due e.g. to motor noise or external forces. In this scheme, the actual motor command u(t) + v is the sum of the feedforward and feedback commands. The matrices L(t) and K(t) denote the optimal filter and feedback gains respectively, g is the output function and in the local approximation. The random processes and are implemented here as Brownian. D(t) is an observation noise matrix, the magnitude of which can be increased to simulate the absence of vision. The vector h denotes the deviation from the sensory prediction. The terms Cff and Cfb refer to the cost functions that determine the optimal signature of the feedforward and feedback commands. The definition and meaning of all the variables are given in the Results section.As we shall see, considering a feedforward command for a stochastic plant allows predicting an optimal movement duration because this command considers the effects of additive noise in the temporal evolution of the state covariance. The actual movement timing and variability may however be affected by sensory-based motor corrections issued online to handle unexpected perturbations and critical deviations from the task’s goal. This is the role of the high-level feedback command. In SFFC, the motor plan is thus primarily composed of a feedforward motor command (i.e. an optimal open-loop control) and an expectation about the upcoming state trajectory. It is complemented by a locally-optimal feedback gain that combines with a limb state estimate throughout movement execution to determine a task-relevant corrective motor command. This estimate is based on both internal predictions and relevant sensory information (e.g. proprioception or vision). We describe below how the feedforward and feedback components of the model are computed.
Determining the feedforward motor command via nonlinear stochastic optimal open-loop control
Here we consider a minimum effort-variance model of motor planning with additive and multiplicative motor noise to determine the feedforward motor command. The expectation and covariance of a nominal state trajectory can be obtained from this sub-problem.Let us consider a general rigid body dynamics with n degrees of freedom such as to model human arm movements:
where is the joint coordinates vector, the net joint torque vector produced by muscles, the inertia matrix, the Coriolis/centripetal, the viscosity, and the gravity terms. This dynamical system is nonlinear due to the mechanical coupling between the different body segments and gravity. Let us assume that the torque change is the control variable as in [18]:In a standard SOC model the motor command would be a stochastic variable u that depends on the random fluctuations arising from motor and measurement noise as well as from any environmental perturbations. As stressed before and in Fig 2, here we rather assume that motor planning primarily builds a feedforward motor command. This enables the feedforward component of the motor command to consider all the internal and external dynamic effects that can be learnt (including the consequences of noise, the sensory delays, the instability due to the interaction with the environment etc.) during the planning stage. To derive such a feedforward motor command, we restrict the control to be open-loop (denoted by u(t) to stress its deterministic nature) while retaining the stochastic aspect of the system’s dynamics.To this aim, let us assume that the arm movements are affected by multiplicative motor noise (i.e. with signal-dependent variance) and additive motor noise (i.e. with constant variance), modeled as a M-dimensional standard Brownian motion, . The corresponding stochastic dynamics of the arm can be described by
where x is the stochastic state vector and and are respectively the drift and diffusion terms (here N = 3n). The matrix G includes both the constant and the signal-dependent noise terms.For the reaching task under consideration, the control objective is to move the arm from an initial position x0 to a given target in time T with minimum effort and minimum variance, that is, by minimizing an expected cost of the form
where ϕ is a quadratic function penalizing the final state of the process (typically related to its covariance here), and l is a cost depending on and the open-loop control u(t). The cost l can be thought as a measure of effort but it can also include terms like trajectory smoothness. The parameter r is a weighting factor to trade-off the variance and effort/trajectory costs.This stochastic optimal open-loop control problem can be solved using approximate solutions based on stochastic linearization techniques (see [33]).Let us denote the covariance of the process x byIt can be shown (e.g. [39], Chap. 12) that propagation of the mean m(t) and covariance P(t) can be approximated using a second order Taylor’s expansion for f by the following ordinary differential equations:These ordinary differential equations are important to reformulate the initial problem as a deterministic optimal control problem involving only the mean and covariance of the original stochastic state process x.To do so, it must be noted that the expected cost function can also be rewritten in terms of the mean and covariance of x only as
where Φ is a function of the terminal mean and covariance of the random variable x. Note that the trajectory cost l can be taken outside of the expectation because the control and mean are deterministic variables (by hypothesis and definition, respectively).We thus obtain a deterministic optimal control problem (approximately equivalent to the stochastic problem defined by Eqs 3 and 4) to solve for the augmented state (m, P). This problem is summarized as:Interestingly, the efficient theoretical and numerical tools developed for DOC can be used to solve the above problem (e.g. [40]). Note that hard constraints for the final mean or covariance of the state can also be added in this formulation. We did so for the final mean state to ensure that the arm exactly reaches the desired target on average even though we could have modelled this constraint in the cost itself. The latter choice has the disadvantage of introducing additional tuning weights in the cost but could allow accounting for a terminal bias. Here we rather left the final covariance free because it was penalized in the cost function and was needed to determine an optimal movement duration without having to preset a desired amount of endpoint variance. The above optimal control problem can be run in free time, which means that the duration T can be found automatically from the necessary optimality conditions of Pontryagin’s maximum principle (instead of a laborious trial-and-error search process as it has been done in previous approaches such as [8] or [41]). The problem can also be run in fixed time, which means that the duration is preset by the researcher. We did so when investigating the evolution of optimal costs with respect to various movement durations and to adjust noise magnitudes for fast or slow movements performed without vision (see Materials and methods section).
Determining the feedback motor command via linear stochastic optimal feedback control
During their execution, movements can be modified with incoming sensory information. This information can be exploited to form an optimal estimate of the current limb state, which can be used online through a linear locally-optimal feedback control scheme. Here, by linearizing the dynamics around the nominal expected state/control trajectories coming from SOOC, we will use the standard linear-quadratic-Gaussian framework [19, 37]. We shall also consider that, besides motor noise, there is some observation noise, the magnitude of which will depend on the available sensory modalities (e.g. with or without vision).At this stage, we have access to a nominal open-loop control and an expected state trajectory, denoted by u(t) and m(t) respectively, for t ∈ [0, T]. We next extended the time horizon T′ > T in order to consider that executed movements have a longer duration than initially planned, assuming that the system is at rest for t ≥ T.To compute a locally-optimal feedback control, the dynamics is linearized around m(t) and u(t) using Taylor’s expansions to obtain a linear-quadratic-Gaussian approximation in terms of state/control deviations (e.g. [42]) as follows:
where
andWe further assume that the noisy sensory feedback y is obtained during motion execution from the following output equation:
where and is the output function. The matrix specifies how observation noise affects sensory feedback, where is a L-dimensional standard Brownian motion process.Using again a Taylor’s expansion and defining , the output equation can be approximated locally in terms of state deviations z as
where h = y − y(t) with y(t) = ∫g(m(t))dt.For this sub-problem, a quadratic cost function to ensure task achievement with minimal effort is defined as follows:The locally-optimal feedback control law can be written as where K(t) is the feedback gain matrix and is the optimal estimate of the state deviation z obtained from the Kalman filter equation:
where L(t) is the optimal filter gain.The problem defined by Eqs 9, 13 and 14 is a linear-quadratic-Gaussian problem, which can be solved using standard algorithms (e.g. [19]).An overview of the SFFC model is given in Fig 2.While the cost functions for the feedforward and feedback components both minimize error and effort terms (see Eqs (4) and (14)), they differ in several fundamental aspects. On the one hand, the feedforward cost function relies on deterministic variables that can be computed or estimated prior to the movement start. It aims at determining an optimal feedforward motor command, from which an expectation about the upcoming trajectory can be obtained. This cost minimizes effort (and possibly other terms such as smoothness) and endpoint variance, which in turn allows to specify the shape and characteristics of mean arm trajectories as well as the state covariance that would result from feedforward control (i.e. without online sensory feedback). Critically, this knowledge allows linearizing the arm’s dynamics in order to apply the linear SOC framework subsequently. On the other hand, the feedback cost function depends on the stochastic deviations from the above expected control/state trajectories which will arise during movement execution. It ensures that the task will be achieved with a minimal amount of motor correction, in accordance with the minimal intervention principle [13]. This is done here by minimizing errors at the end of the movement (i.e. for times longer than the planned movement duration). The R term can be adjusted depending on the task, which will in turn determine the magnitude of the planned feedback gain K(t). To compute the feedback component of the motor command, related to online corrections, the system requires sensory information to update the estimate of the limb state throughout movement execution. Hence, internal or external perturbations inducing a deviation from the expected trajectory will be corrected via a task-dependent feedback mechanism. While SFFC assumes that the brain has some knowledge of the upcoming reach trajectory and feedforward command to reformulate the task in terms of state/control deviations, it must be noted that the feedback cost does not assume that a reference trajectory is tracked. If the task does involve trajectory tracking this can be handled by minimizing errors throughout the whole movement in the feedback cost or modelled by considering muscle viscoelasticity and mechanical impedance as in [33].
Simulation results and comparison to experimental data
To test the SFFC model we consider a pointing task with a two-link arm moving in the horizontal plane from an initial posture to a target in Cartesian space (e.g. [31, 35, 36]). More details about the task, model of the arm and choice of parameters can be found in the Materials and methods section.
Comparison to previous data of reaching movements without vision
We first tested if SOOC can determine an optimal movement timing from the principle illustrated in Fig 1B. As SOOC does not consider the influence of online sensory feedback, its prediction would mainly correspond to the behavior of deafferented patients without vision of the moving hand (e.g. [43, 44]). It must be noted that SOOC at least requires an estimate of the initial arm’s state to build the optimal feedforward motor command, in agreement with [43] who showed that prior vision of the arm improved movement precision in these patients.Fig 3A illustrates that the evolution of the optimal expected cost with respect to movement duration. This U-shape cost function yields an optimal duration, which can be thought as the limit case of Fig 1B when observation noise is infinite. Remarkably, the resulting optimal duration is longer than the duration of minimum variance, which is in agreement with the observation of [31]. Fig 3B and 3C show the corresponding mean hand trajectory, the path of which is approximately straight with bell-shaped velocity profile. This agrees with typical findings in healthy subjects and also with the overall strategy of deafferented patients without online vision [44]. Interestingly, SOOC also provides information about the final variability of the pointing under pure feedforward control (i.e. without feedback component; represented as a confidence ellipse in Fig 3B). The large final variability (compared to target size) is compatible with the relatively large endpoint variability exhibited by deafferented patients [44].
Fig 3
Expected costs and trajectory predicted by SOOC for a horizontal point-to-point arm movement without feedback.
A. Evolution of the optimal costs with respect to the movement duration. The total cost function exhibits a U-shape, i.e. additive motor noise yields a minimal movement duration. The minimal duration (minimum of the black solid trace) is larger than the duration of minimum variance (the gray curve). B. Mean hand path and predicted endpoint variability (here depicted as a 90% confidence ellipse). These data can be computed from m(t) and P(t) respectively (converted from joint space to Cartesian space). The black circle depicts the target. C. Corresponding mean velocity profile. The dashed horizontal line indicates the threshold at which velocity profiles are cut in the experimental data. This figure is generated using simulation data of free-time optimal control computed with SOOC.
Expected costs and trajectory predicted by SOOC for a horizontal point-to-point arm movement without feedback.
A. Evolution of the optimal costs with respect to the movement duration. The total cost function exhibits a U-shape, i.e. additive motor noise yields a minimal movement duration. The minimal duration (minimum of the black solid trace) is larger than the duration of minimum variance (the gray curve). B. Mean hand path and predicted endpoint variability (here depicted as a 90% confidence ellipse). These data can be computed from m(t) and P(t) respectively (converted from joint space to Cartesian space). The black circle depicts the target. C. Corresponding mean velocity profile. The dashed horizontal line indicates the threshold at which velocity profiles are cut in the experimental data. This figure is generated using simulation data of free-time optimal control computed with SOOC.Next, we focus on previous experimental observations in healthy subjects performing movements without online visual feedback of the hand [31, 35, 36]. Healthy subjects typically have a smaller endpoint variability than deafferented patients without vision. In healthy subjects, proprioceptive feedback is indeed available and this sensory information can be used by the brain to build an estimate of the limb state. However, the absence of vision (and thus of multisensory integration) may degrade the hand state estimate [45], which can be accounted for in our model by assuming a relatively large observation noise in this case.Fig 4A and 4B show the path of trials in the N-W and N-E directions when simulating the experiment of [31], where parameters were selected to reproduce the data in the N-W direction. The trajectories predicted by the model are similar to the trajectories experimentally measured in this task, with relatively straight hand path and bell-shaped velocity profile. The duration determined by the SFFC model was also in good agreement with the data. Interestingly, as in the experimental data of [31], the predicted duration was slightly longer for movements in the N-W than in the N-E direction, and the variance was also slightly larger in the N-W direction. It can also be noted that the endpoint variability is smaller with SFFC than with SOOC, thereby illustrating the improvement with proprioceptive feedback.
Fig 4
Simulations of horizontal arm pointing movements without online vision of the hand.
A-B. Simulation of the data of [31]. Hand paths of 20 trials are shown in panels A and B for the N-W and N-E directions, respectively. 90% confidence ellipses of the end points are depicted in blue, which were computed from 1,000 samples. Dotted ellipses, corresponding to the endpoint variance of SOOC solutions, are depicted for comparison. The corresponding mean velocity profiles are also depicted as insets. The target is depicted as a black circle. Movement duration and endpoint variance are reported for each direction. C-F. Simulation of the data of [35, 36]. Hand paths of 20 trials are shown in panels C and E for the different directions and different distances for the N-E direction. 90% confidence ellipses of the end point are depicted (estimated from 1,000 samples). The corresponding durations are reported in panels D and F. In panel D, the acceleration predicted from the hand mobility matrix is depicted (and calculated as in [36]). The direction-dependent and distance-dependent modulation of duration can be noticed.
Simulations of horizontal arm pointing movements without online vision of the hand.
A-B. Simulation of the data of [31]. Hand paths of 20 trials are shown in panels A and B for the N-W and N-E directions, respectively. 90% confidence ellipses of the end points are depicted in blue, which were computed from 1,000 samples. Dotted ellipses, corresponding to the endpoint variance of SOOC solutions, are depicted for comparison. The corresponding mean velocity profiles are also depicted as insets. The target is depicted as a black circle. Movement duration and endpoint variance are reported for each direction. C-F. Simulation of the data of [35, 36]. Hand paths of 20 trials are shown in panels C and E for the different directions and different distances for the N-E direction. 90% confidence ellipses of the end point are depicted (estimated from 1,000 samples). The corresponding durations are reported in panels D and F. In panel D, the acceleration predicted from the hand mobility matrix is depicted (and calculated as in [36]). The direction-dependent and distance-dependent modulation of duration can be noticed.We then analyzed how the optimal movement duration depends on its direction and amplitude by using the same model parameters and still focusing on movements carried out without online visual feedback of the hand. Fig 4C–4F show how movements vary with the direction and with the distance. As in the experimental results of [35] and [36], movements in directions requiring more effort are slower. Fig 4F further shows that the predicted movement duration increases monotonically with the target distance as in experimental data [36].These simulations suggest that the model can reproduce the basic characteristics of planar arm reaching movements without visual feedback, showing typical dependencies on distance and direction. Next, we compare movements carried out with and without vision. In particular, movements with vision are known to exhibit a smaller endpoint variability than analog movements without vision [45, 46].
Comparison to data of reaching movements with and without vision
We asked healthy participants to perform arm pointing movements to test the impact of online visual feedback of the hand on the preferred timing and variance of goal-directed movements. We wanted to estimate the extent to which movements with and without visual feedback differed in terms of timing and trial-by-trial variability, and whether these data could be replicated by the proposed SFFC model. Horizontal arm pointing movements of various directions {E, N-E, N-W, W} and amplitudes {0.06, 0.12, 0.18, 0.24} m both with and without vision of the moving hand (represented as a cursor on the screen, the actual arm being hidden) were recorded. Details about the task can be found in the Materials and methods section.Fig 5A illustrates the experimental hand paths in all the directions and amplitudes in one participant. As expected, movements without vision were clearly less precise than movements with vision. This was confirmed by the group analysis reported in Fig 6A where the endpoint variance was computed for each distance and direction.
Fig 5
Hand trajectories and velocities for an exemplar subject and for simulations.
Movements with vision and without vision are depicted in red and blue respectively. Panels A and B show experimental trajectories in the N-W and N-E directions respectively. The four distances {0.6, 0.12, 0.18, 0.24} m are represented by shifting the starting point for visibility. The real starting point was the same as described in Fig 8B. Targets are represented as 90% confidence ellipses for the endpoints, which were estimated from the five experiment’s trials. Corresponding mean speed profiles over the five trials are computed after cutting the start and end using a 1.0 cm/s threshold and a time normalization. The same information is shown in panels C and D for 1,000 simulated movements with SFFC, where only 5 trajectories are depicted for clarity. The blue traces correspond to large observation noise and red traces to normal observation noise when vision is available. The simulation parameters were chosen to reproduce the average behavior of the participants, and not the plotted data of a specific participant.
Fig 6
Comparison of experimental and simulated data.
A. Mean experimental endpoint variance (across participants) for each direction and distance. Error bars indicate standard deviations across the 16 conditions (distance-direction pairs). Movement with and without vision are reported in red and blue, respectively. Circles, diamonds, squares and triangles represent the E, N-E, N-W and W directions, respectively. B. Same information for simulated movements. C. Root mean squared deviation (RMSD) between the real and simulated endpoint variance, (in log(mm2)). D-I. Same by reporting duration (in s) and peak velocity (in cm/s) instead of endpoint variance.
Hand trajectories and velocities for an exemplar subject and for simulations.
Movements with vision and without vision are depicted in red and blue respectively. Panels A and B show experimental trajectories in the N-W and N-E directions respectively. The four distances {0.6, 0.12, 0.18, 0.24} m are represented by shifting the starting point for visibility. The real starting point was the same as described in Fig 8B. Targets are represented as 90% confidence ellipses for the endpoints, which were estimated from the five experiment’s trials. Corresponding mean speed profiles over the five trials are computed after cutting the start and end using a 1.0 cm/s threshold and a time normalization. The same information is shown in panels C and D for 1,000 simulated movements with SFFC, where only 5 trajectories are depicted for clarity. The blue traces correspond to large observation noise and red traces to normal observation noise when vision is available. The simulation parameters were chosen to reproduce the average behavior of the participants, and not the plotted data of a specific participant.
Fig 8
Arm reaching task in our experiments and simulations.
A: Two-link model of the arm and planar movement used to set the model parameters. B: Horizontal arm movements carried out with and without online vision of the hand (4 directions, W, N-E, N-W and E, and 4 distances, 6, 12, 18 and 24 cm) in our experiment. C: Arm parameters used in the simulations.
Comparison of experimental and simulated data.
A. Mean experimental endpoint variance (across participants) for each direction and distance. Error bars indicate standard deviations across the 16 conditions (distance-direction pairs). Movement with and without vision are reported in red and blue, respectively. Circles, diamonds, squares and triangles represent the E, N-E, N-W and W directions, respectively. B. Same information for simulated movements. C. Root mean squared deviation (RMSD) between the real and simulated endpoint variance, (in log(mm2)). D-I. Same by reporting duration (in s) and peak velocity (in cm/s) instead of endpoint variance.Two-way repeated measures ANOVAs confirmed a main effect of the visual condition (F1,20 = 426.83, p < 0.001) with movements without vision exhibiting much more endpoint variance. A main effect of distance was also found (F3,60 = 12.55, p < 0.001) and there was a significant interaction (F3,60 = 46.59, p < 0.001) revealing that, without vision, participants were less and less precise as movement amplitude increased. A main effect of the direction was also detected on the variance (F3,60 = 3.73, p < 0.05), and there was no interaction effect between direction and condition (p = 0.056).These empirical observations were well replicated by our model (Figs 5B and 6B). In particular, the increase of endpoint variance with distance is well predicted by the model with large observation noise and the gain of precision is also clear when vision is present and observation noise is thus reduced. To quantify the errors between the model predictions and the empirical data, we computed root mean squared deviations (RMSD). Fig 6C reports RMSD values averaged across all directions and distances for endpoint variance. The average RMSD was 0.40 and 0.28 log(mm2) for the without and with vision conditions respectively, which corresponded to 6.9% and 11.6% of the respective experimental mean values.We next analyzed the timing of movements performed with and without vision (Fig 6D). A visual inspection reveals that the durations of movements with and without visual feedback of the hand exhibit similar trends, although movements with vision may tend to have slightly longer durations.A two-way repeated measures ANOVA revealed no main effect of the visual condition on movement duration (p = 0.055). We found a main effect of distance (F3,60 = 242.40, p < 0.001) on movement duration (i.e. duration clearly increases with distance). A significant interaction effect between the visual condition and the distance was found (F3,60 = 10.06, p < 0.001). Post-hoc analyses revealed that only the 24 cm distance had significantly longer duration with vision compared to without vision (p = 0.014). Regarding the effect of direction on duration, a significant interaction effect between the visual condition and direction was found (F3,60 = 5.66, p = 0.002). Post-hoc analyses mainly revealed that N-W movements with vision lasted longer than the other directions of movement (p < 0.001). The model replicated the increase of movement duration with distance relatively well, although some variations with respect to direction were less clear in these data (see Fig 6E). Quantitative comparisons are reported in Fig 6F and reveal that, on average across distance and direction conditions, RMSD for duration was 70 and 113 ms for the without and with vision conditions respectively, which corresponded to 7.1 and 11.8% of the respective experimental mean values.To analyze the differences in movement timing with a variable less sensitive to terminal adjustments, we repeated the above analyses using peak velocity instead of duration (Fig 6G). We found neither a main effect of the visual condition (p = 0.052) nor an interaction effect (p = 0.262) with distance. Although there was a trend to have slightly lower peak velocities with vision, no statistical difference was observed on peak velocity for movements with and without vision (even for the largest distance, 24 cm, in contrast to the results found for duration). A main effect of distance on peak velocity was found as expected since peak velocity clearly increases with movement distance (F3,60 = 253.72, p < 0.001). Regarding the effect of direction, we found a significant interaction (F3,60 = 2.83, p < 0.05). Post-hoc tests mainly indicated that N-W movements were slower than those in other directions with and without vision (p < 0.01). The model replicated well the increase of peak velocity with distance (Fig 6H) and the dependence of peak velocity on direction was again less clear in these data. RMSD for peak velocity was on average 3.8 and 2.7 cm/s for the without and with vision conditions respectively, which corresponded to 13.4 and 10.0% of the respective experimental values (Fig 6I).Finally, a correlation analysis was carried out to analyse the extent to which the timing properties of movements performed with and without vision were related (Fig 7A for durations and Fig 7B for peak velocities). We found strong correlations in experimental data (R2 > 0.96), thereby confirming the consistency of movement timing with and without visual feedback. Similarly strong correlations were found in simulated data based on the proposed model. The main reason is that, in the model, movements with or without vision are both based on the same feedforward motor command and just differ here in the magnitude of observation noise (which was assumed to be ×10 larger for movements without vision than for movements with vision).
Fig 7
Correlations of duration or peak velocity between movements with and without vision.
A. Correlations for durations. Each data point represents one condition of distance and direction (averaged across participants). Experimental and simulated data are plotted respectively in black and grey. B. Correlations for peak velocities. Regression lines are plotted for the illustration.
Correlations of duration or peak velocity between movements with and without vision.
A. Correlations for durations. Each data point represents one condition of distance and direction (averaged across participants). Experimental and simulated data are plotted respectively in black and grey. B. Correlations for peak velocities. Regression lines are plotted for the illustration.
Discussion
This paper introduced the stochastic optimal feedforward-feedback control model (SFFC) of learned goal-directed arm movements unifying previous optimal control models that focused either on deterministic or stochastic aspects of movement. The SFFC suggests how the nervous system may cope with noise and delays by combining feedforward and feedback motor command components. It can be used to predict the nominal timing and variability of reaching movements with degraded sensory feedback, as was illustrated on movements carried out without visual feedback. We discuss below the main aspects of this new model in perspective with experimental results and previous models from the literature.
How existing models predict movement timing and variability
The development of SFFC was prompted by the difficulty to predict movement timing independently of endpoint variability with existing optimal control models. Optimal control being a versatile framework to model human motor control [47, 48], several classes of models have been proposed with prediction of movement timing and variability summarized in Table 1. Seminal deterministic optimal control (DOC) models can predict the shape of average arm trajectories corresponding to a given movement duration [17, 18, 49, 50]. The movement duration can be determined in ad hoc ways such as by setting the task’s effort [51-53] but DOC does not account for the trial-by-trial variability of human movement. Assuming signal-dependent motor noise, SOOC models have been proposed to extend deterministic models and predict a movement duration corresponding to a fixed level of endpoint variance (e.g. the width of the target) [8, 20], but these models will follow Fitts’ law [54], which does not hold for self-paced arm movements [55]. Here we showed that SOOC can explain the timing and variability of self-paced movements carried out without sensory feedback by considering the effects of motor noise together with a minimum effort-variance cost. However, SOOC will not account for the drastic reduction of variability exhibited by movements executed with proprioceptive and/or visual feedback. Muscle co-contraction and mechanical impedance that could be modelled as in [34] may contribute to reduce this variability but not to the level of movements with online multi sensory feedback.
Table 1
Predictions of movement timing (duration or speed) and endpoint variance (variability across trials) with different types of optimal control models.
Some ad hoc fixes have been introduced in some models to predict timing and/or variability, unlike the SFFC model.
Type
Timing
Variability
Models
vision
no vision
vision
no vision
DOC
N
N
N
N
[17, 18, 49]
DOC with cost of time or other fixes
Y
Y
N
N
[24, 25, 51–53, 56]
SOOC
Y
Y
N
Y
[8, 20, 34]
SOC
N
Y
Y
Y
[14, 19]
SOC with cost of time or other fixes
Y
Y
Y
Y
[22, 41]
SFFC
Y
Y
Y
Y
introduced in this paper
Predictions of movement timing (duration or speed) and endpoint variance (variability across trials) with different types of optimal control models.
Some ad hoc fixes have been introduced in some models to predict timing and/or variability, unlike the SFFC model.SOC emphasized the role of high-level feedback to reliably execute a motor task despite relatively large variability in repeated movements. In SOC, the motor command is a function of a limb state estimate built from internal dynamic predictions and delayed sensory information. By considering sensorimotor noise, and minimizing error and effort [13, 14, 19], these models correct task-relevant errors according to the minimal intervention principle. However, as the expected cost typically plateaus for visually-guided movements of long duration (see Fig 1A), SOC cannot predict a finite movement duration without ad hoc criterion. For instance, [22] determined duration in an infinite-horizon SOC formulation by comparing the magnitude of endpoint variance to the target’s width, which allowed to predict the speed-accuracy trade-off. To model the variability of movements without vision, SOC models will normally assume a large observation noise. This will degrade limb state estimates and make the controller more dependent on internal predictions corresponding to a feedforward mechanism. The fact that movements carried out with and without vision had a highly correlated timing in our experimental data is supporting the hypothesis that these two types of movement have a common origin, which can be captured by a feedforward motor command. The same conclusion was drawn by [45] who found a reduction of feedback gains in a reaching task when visual feedback was removed. The authors suggested that a feedforward motor command is needed to explain that movements with lower feedback gains had well preserved kinematics and timing. Note however that visually-guided movements have a tendency to be slower likely due to the integration of visual corrections at the end of the movement, to adjust more accurately the final cursor’s location. This is consistent with the observation that in our experiment the peak velocities with and without vision were even more similar than movement durations. Overall, SFFC appears as the first model that can explain timing and variability of arm movement trajectories carried out with or without visual feedback from neuromechanic considerations only.
Is movement timing due to neuromechanic or neuroeconomic factors?
Previous arm movement models [24, 25, 56] used a cost of time to limit the movement duration. In these DOC models, the cost-of-time parameters to accurately reproduce the movement timing observed experimentally can be determined using inverse optimal control techniques [25, 27]. By extension, SOC with a cost of time has been used to model saccades [41] and, in principle, SOC with a cost of time may be able to reproduce the above experimental results. However, it is not straightforward to find an optimal duration in SOC from computing the cost for all possible durations, and to adapt such models to the nonlinear dynamics of the human arm. In contrast, it would be straightforward to include a cost of time in SFFC (in the term l(x,u)) and to use it to determine the optimal movement duration from necessary optimality conditions. However our objective here was to understand if neuromechanical factors could explain the timing and variability of self-paced arm movements, which led us to develop the SFFC model. While above simulations and experimental results showed that SFFC could explain the timing of simple pointing movements toward targets without ad hoc hypothesis, further investigations would be required to determine whether it could also account for individual differences [27, 57–59] and sensitivity to reward [60-62], e.g. by varying the factor r to trade-off variance and effort for instance. Experiments may also be developed to test the model’s prediction that movement timing should increase with larger constant motor noise and decrease with larger signal-dependent motor noise.
The role of motion planning and feedforward control
Overall, this study suggests the importance of motion planning in the generation of goal-directed arm movements. A large body of experimental evidence has shown the critical role of motion planning to select a suitable motor solution for carrying out a task (see [4] for a review). The picture suggested by previous studies and above modeling is that the CNS executes well-learned, unperturbed movements using an important feedforward component to the motor command, given the intrinsic noise, delays and task dynamics. The sensorimotor plans required for such control strategy may be learned by gradually minimizing reflexes and integrating voluntary (e.g. visual) corrections after movement [63-65]. This learning will minimize the reliance on high-level feedback corrections to achieve the task and thus gradually incorporate in the feedforward motor command any feature that can be identified over trials.The behaviour after learning could be captured by the SFFC model that integrates feedforward and feedback control. The simulation results illustrated how SFFC combines the advantages of SOOC [8, 33, 34] and SOC [10-12] to explain the timing and the variability of arm movements performed with or without visual feedback of the moving limb, by minimizing the consequences of signal-dependent and constant motor noise on endpoint variance as well as effort or kinematic costs such as smoothness. One important aspect of SFFC is that the feedforward motor command already considers uncertainty about the task dynamics (e.g. motor noise or unknown perturbations, like in [8]) and can incorporate this knowledge in the plan to adjust the mechanical impedance to the task’s uncertainty [34, 66, 67]. This feedforward motor command is complemented by a high-level feedback motor command that corrects task-relevant deviations resulting from perturbations not handled by the feedforward motor command, such as accumulation of positional errors due to constant noise [31, 68], visually elicited corrections [64, 69] or long-latency proprioceptive feedback responses to large mechanical perturbations [70, 71]. The state-feedback gain is also part of the motor plan, the magnitude of which can be adapted depending on the task (by tuning the weights in the feedback cost function in SFFC). This general planning scheme highlights how feedforward motor commands (which determine the nominal shape, timing and variability of unperturbed trajectories) and feedback motor commands (which handle the corrections of task-related errors using current limb state estimates) could yield a skillful motor control strategy.
Materials and methods
Ethics statement
The experimental protocol was approved by the Université Paris-Saclay local Ethics Committee (CER-Paris-Saclay-2019–031). Written informed consent was obtained from each participant prior to starting with the experiment.
Experimental task and procedures
Participants and experimental setup
21 young adults (24.5 ± 2.0 years old [mean±std], height 1.74 ± 0.09 m, with 11 females and 4 left-handed) participated in this study. All participants had normal or corrected to normal vision, and no known neurological impairment or mental health issue. Each participant was seated, and had to move a stylus on a Wacom tablet (Wacom Intuos 4 XL) laid on horizontal table. The location of the stylus on the tablet was displayed on a monitor placed in front of the participant (i.e. on a vertical screen).
Pointing task in two conditions: With and without online visual feedback
When a participant was ready, a 5 mm diameter disk appeared on the screen indicating the start position, on which they was instructed to move the cursor. Once the center of the cursor was within the start disk for 1 second, this was replaced by a 3 cm diameter target disk placed at 6, 12, 18 or 24 cm from the start position. Reaching movements were carried out in four directions as indicated in Fig 8A and 8B. If the start position was at the bottom of the screen (x-y coordinates with respect to the shoulder [-15, 30] cm), the target was placed in the N-E or N-W direction. If the start was on the left of the screen (coordinates [-29, 34.5] cm, it was in the E direction, and if it was on the right of the screen (coordinates [5, 34.5] cm) in the W direction. This resulted in 16 different possible movement types.
Arm reaching task in our experiments and simulations.
A: Two-link model of the arm and planar movement used to set the model parameters. B: Horizontal arm movements carried out with and without online vision of the hand (4 directions, W, N-E, N-W and E, and 4 distances, 6, 12, 18 and 24 cm) in our experiment. C: Arm parameters used in the simulations.The participants were instructed to move the cursor at comfortable pace in order to reach the target, without leaning the arm on the tablet. Note that their arm was hidden by a cardboard box so that they could not see it. They had to perform reaching movements either without or with the cursor displaying their hand position on the screen during the movement. In the non-visual condition, the cursor disappeared at the beginning of the movement and reappeared 1 s after the end of the movement, to indicate the pointing error and thus avoid the endpoint to gradually drift trial after trial.Each participant started with a familiarization phase of 32 pointing movements including 16 consecutive trials per condition (with and then without visual feedback). Then they had to perform 160 trials, with 80 per visual feedback modality. The different movement types and the visual modalities were presented in pseudo-random order. This resulted in 5 trials of each amplitude and direction for each starting position. Every 10 trials a break of approximately 1 minute was scheduled, during which they could place the forearm on the tablet or the desk.
Data acquisition and parameters of interest
The stylus position was recorded at 125 Hz with MATLAB (The MathWorks, Inc.), and the Psychtoolbox [76] was used to display the stimuli on the screen. The system was calibrated so that a movement of the stylus on the tablet corresponded to a movement of the same length of the cursor on screen. The raw data were smoothened for further analysis using a 5th-order Butterworth low-pass filter with 12.5 Hz cutoff frequency and without delay. Velocity was computed via numerical differentiation. Among the parameters of interest, we computed the movement duration using a velocity threshold of 1 cm/s, the peak velocity of the maximal value of velocity profiles (in cm/s), and the endpoint variance (in log(mm2)). In every trial, the movement’s endpoint was determined by the last recorded position at the end of the movement time. Endpoint variance was then estimated from the trace of the covariance matrix of final positions and the logarithm of this value was computed as in [31].
Statistical analysis
Two-way repeated measures ANOVAs with condition (with vision and without vision) and amplitude (from 6 to 24 cm) or direction (from E to W) as within-subjects factors were carried out to assess the variation of movement timing (i.e. duration and peak velocity) and variance across conditions. Moreover, a correlation analysis was performed to assess the relationships between the timing of movements carried out with and without vision.
Numerical simulations
Arm reaching movements were simulated using a 2-link arm model with joint configuration vector
where q1 is the shoulder and q2 the elbow angles. The skeletal dynamics of the arm was described by the rigid body model of Eq (1) with:For the planar movements considered in this paper the gravity term is set as zero. {I, L, L} and {M} are the moments of inertia, lengths of segments, lengths to the centre of mass and mass of the segments.Furthermore, we have
where the parameters {σ} are used to set the magnitude of additive noise and {d} the magnitude of multiplicative noise.Regarding the feedforward cost function, we define ϕ(m(T), x) to estimate the covariance of the final hand position. Denoting by J(q) the Jacobian matrix of the two-link arm, an approximation of this function can be computed by:
where m(T) is the mean final position of the random variable q (i.e. the 2-dimensional vector of final joint positions) and tr denotes the trace of the matrix. The expectation of ϕ(m(T), x) can then be rewritten as a function of the mean and covariance of the state process x:
where P is the 2×2 covariance matrix of joint positions.The infinitesimal cost l(m, u) is defined as follows:
where x and y denote the mean Cartesian positions of the hand (which can be computed from m(t) and the forward kinematic function). This cost implements a compromise between effort (here measured as squared torque change) and smoothness (here squared hand jerk) through the α parameter. Evidence for composite cost function mixing kinematic and dynamic or energetic criteria has been found in previous works [50, 72]. The jerk term is useful to correct for abnormal asymmetries in velocity profiles which may arise partly from the minimum torque change model (e.g. [73]), but this term does not affect our results otherwise.For the linear-quadratic-Gaussian sub-problem, we set C = I6 (identity matrix) meaning that we assume that both position, velocity and force could be estimated from multisensory information as in [19]. We verified that the same results and conclusions were obtained by limiting the observation matrix to the position and velocity components only. The observation noise matrix D was taken of the form D = β
I6 where β specifies the overall magnitude of observation noise. This parameter can be varied depending on whether vision of the cursor is available or not during the movement. Finally, for the feedback cost function, we set R = ρdiag(1, 1, 0, 0, 0, 0) such that only the deviations about the final arm posture defined by the target location were penalized during the post-movement interval.The SOOC solutions were obtained with the optimal control software that approximates the continuous-time optimal control problem as a sparse nonlinear programming problem [40]. To compute the SFFC solutions, we considered a discrete time approximation of the linear-quadratic-Gaussian sub-problem around the SOOC solution with a time step of dt = 0.005 s. Standard discrete-time algorithms for linear-quadratic-Gaussian control were then used to compute the gains [19]. In our simulations, we extended the time horizon by 1 s (T′ = T+1) to consider movements longer than the planned duration T. We tested different extended horizon between 0.5 s and 2 s and it did not change the results. It is worth noting that sensory feedback delays can be easily handled at this stage due to the discrete time approximation. All the simulations were performed with MATLAB (Mathworks, Natick, MA).
Selection of model parameters
The arm parameters used in the simulations (from [42], in SI units) are given in Fig 8C.The remaining parameters of the model are related to cost functions (α, r and ρ) and noise magnitudes ({σ}, {d} and β). Some of these parameters affect the design of the feedforward command (α, r, {σ}, {d}) and the others affect the design of the feedback command (ρ, β). First, we verified that the qualitative predictions and principles of the model were robust to parameters choices. Second, to have simulations that correspond quantitatively to experimental data, we adjusted the parameters using the procedure described hereafter. Note that we did not try to find the best-fitting parameters using an automated procedure but adjusted the parameters to yield timing and variance of the same order of magnitude as experimental data.We first fixed α = 0.02 in all simulations, to implement a compromise between torque change and hand jerk. Note that we also considered α = 0 and the results revealed that the smoothness term contributes to get slightly more linear hand paths with more bell-shaped velocity profiles, but this does not affect the main findings. Second, to reduce the number of parameters, we assumed that the magnitude of additive and multiplicative motor noise are the same in the two joints of the arm, i.e. σ1 = σ2 = σ [rad/s3/2] and d1 = d2 = d [rad/(Nm s1/2)]. The three remaining free parameters for SOOC ({σ, d, r}) were then adjusted by considering a movement of 7.4 cm in the N-W direction by using the existing data of [31] as a reference. The initial arm configuration was approximately q1(0) = 50° and q2(0) = 100° in this experiment. In the N-W movement, both joint angles change significantly, so that the effects of noise magnitude can be estimated in the two degrees of freedom using the three steps as follows:Since additive noise dominates at low speed, the magnitude of constant noise was adjusted on 1400 ms long movement in order to obtain an endpoint variance larger than what has been found in [31]. Indeed, these data were obtained for movements without vision in healthy subjects, where proprioceptive feedback was still available, and analog movements in deafferented patients would exhibit a larger endpoint variance [43, 44]. This resulted in σ = 0.005 [rad/s3/2] and in about 6 log(mm2) of endpoint variance (which is larger than the 4.2 log(mm2) measured in [31]).Since multiplicative noise dominates for fast speed movements, which are less affected by feedback, multiplicative noise was adjusted on 350 ms long movements based on the data of [31]. d = 0.01 yielded an endpoint variance about 4.1 log(mm2).The variance weight r was then adjusted to fit the preferred duration of movements observed in the N-W direction. We found that r = 2,000 yields a movement time of about 1080 ms, which is similar to the preferred duration in [31].Once the SOOC solution was obtained, we determined the remaining parameters of the SFFC model, which are related to the linear-quadratic-Gaussian sub-problem. We first set the observation noise β and feedback cost weight ρ by assuming that visually-guided movements are performed accurately at the preferred speed. This resulted in β = 0.003 and ρ = 1,000 for the data of [31]. Without vision, only proprioceptive feedback can be used and we assume that this leads to an increase of sensory variance. This increase was chosen to match the endpoint variance observed without vision in [31] (<4 log(mm2)), and this led to β = 0.03 (i.e. ×10 larger than the magnitude of observation noise with vision). Note that we did not change ρ in the present simulations, but we also considered that the product ρβ could remain constant (i.e. ρ = 100 if β = 0.03), which reduced the feedback gain without affecting much the simulations for the unperturbed reaching movements under consideration.When simulating movements without vision (or without feedback at all), a basic stopping mechanism was added by increasing joint friction 50 ms before the planned movement end to ensure that the terminal velocity always falls below the threshold (we added 3.5 kg m2/s to , i = 1, 2), which corresponds to a larger muscle viscosity at low speed [74]. Note that to compare simulated and experimental durations, we systematically applied a 1 cm/s threshold on hand velocity in agreement with experimental data processing (see above and [31]).This set of parameters was used to simulate movements of different durations and directions, and compare the predictions to existing data. We used Eq (9) and the above parameters to generate reaching movements of duration 300–1450 ms in the N-W and N-E directions, and then computed the different optimal costs. Next, we tested the model predictions by computing optimal movement durations for increasing distances ({7.5, 12.5, 17.5, 22.5, 27.5} cm in the N-E direction), and in eight directions as in classical experiments of arm reaching movements without vision [35, 36].
Simulation of new experiment with SFFC
To simulate movements with and without vision described in Fig 8A and 8B, two previous parameters had to be adjusted to account for the larger variability and the shorter durations observed in our data compared to the experiment of [31]. This adjustment was made to have a better quantitative fit of the experimental data but the qualitative results would be the same if keeping previous parameters unchanged. These changes may be due to differences in experimental protocols (target’s width, arm’s weight support, instructions etc.). Therefore, to reflect larger variance and shorter durations, we set σ = 0.025 and r = 6,000 and kept the other parameters invariant. Here, to also investigate the influence of sensory delays, we performed simulations by considering a 50-ms delay in sensory feedback loops. This was done in the discrete-time approximation of the linear-quadratic-Gaussian sub-problem by using the classical procedure consisting of augmenting the system’s state to include delayed instances of the state process (e.g. see [75] for details). Note that delays did not affect much the present simulations. This was verified by simulating SFFC with and without delays and very similar quantitative results were obtained for the tested movements. We report the simulations for the delayed case.
Experimental data supporting Fig 6.
(XLS)Click here for additional data file.24 Jan 2021Dear Dr. Berret,Thank you very much for submitting your manuscript "Stochastic optimal feedforward-feedback control determines timing and variability of arm movements with or without vision" for consideration at PLOS Computational Biology.As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.While we have invited a resubmission, as you will see below some of the reviewers were conflicted about how to treat this paper. While there was general praise for the methodological aspects of the modeling, reviewers 1 and 2 especially questioned the assumptions that went into the modeling, such as considering reaching without vision as open loop control.It is not clear whether this criticism can be addressed, but if you believe that is possible we would be happy to consider the paper again.----We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.When you are ready to resubmit, please upload the following:[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).Important additional instructions are given below your reviewer comments.Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.Sincerely,Ulrik R. BeierholmAssociate EditorPLOS Computational BiologySamuel GershmanDeputy EditorPLOS Computational Biology***********************Reviewer's Responses to QuestionsComments to the Authors:Reviewer #1: The authors present a model of human upper limb control composed of a feed-forward specification of a control sequence determined by the minimization of variance and effort, coupled with a local, linearized state-feedback controller which penalises deviation from the expected trajectory under feedforward control. It is argued that the model captures natural features of human movement control, including implicit selection of reach time, and changes in movement variance dependent on the feedback condition. Other models typically specify movement time either arbitrarily or based on factors that do not result from the derivation of the control law.Overall the paper is very well written, very clear and the modelling exercise appears to have been performed very cautiously. The paper complies with utmost scientific standards in terms of the accuracy of mathematical content, and clarity of the writing. However, I emphatically disagree with the relevance of the model to explain features of human sensorimotor control. The reason is that several modelling choices are simply wrong; the first being that reaching with or without vision is equated to open loop or closed loop control, which cannot be less compatible with the physiology. A second critical modelling choice is the use of a control model that consists in trajectory tracking, which is motivated for mathematical reasons to derive locally optimal feedback control laws around a nominal or expected trajectory, yet the possibility that the human brain uses trajectory tracking is debated, and evidence thereof is elusive if existing.The first point is the fact that control without vision is not feedforward control. There is extended evidence that online feedback control uses limb afferent feedback to correct for error when the hand is unseen (Goodale, Pelisson and Prablanc, 1986, Nature, and following literature). Evidence for sophisticated feedback control based on limb afferent feedback has been extensively discussed and reviewed for instance in Scott, 2016 (TINS, 39 (8), 512-526). It is clear that vision contributes to reducing the variability of movement but the larger amount of variability observed when vision is absent should not be equated to feedforward control in a model of human reaching. Recent evidence of subtle effects of vision during reaching has been documented here (ito and Gomi, eLife, 9:e52380), and it is clear that changing visual information is not equivalent to open loop versus closed loop.The second main contentious point is the hypothesis of trajectory tracking, which has been questioned in the work of Todorov and Jordan, 2002, in the task with different numbers of via-points. This led to partially artificial debate about whether there is or not a trajectory tracking mechanism in the brain. In fact, it appears that the brain may or may not use a trajectory for movement control, it is a feature of task related instructions (Cluff and Scott, JNS 35(36):12465–12476). These two points are major limitations: that the paper uses open loop trajectories to model movements without vision is in complete disagreement with current evidence that 1- such open loop trajectory may not exist in the brain, and 2-movements without vision are not feedforward.It does not take anything away from the quality of the modelling work but as a model of human sensorimotor control, the model assumptions are not valid. I find it difficult to make a clear recommendation, the study could be published as is if the authors acknowledge that it is a bio-inspired model of control but it is not a model of the human sensorimotor system. Should the authors wish to make it a candidate model in the field of neurophysiology, then they should profoundly revise the underlying assumptions and abandon the idea of feed forward control in the absence of vision.Major points:The motivation for the model is not clearly apparent, it is interesting that a model can make predictions about movement time, but it is also not a big problem to consider it is a free parameter, that is we can decide movement time (i.e. I want to move slower or faster). Is there a genuine need for a model in which movement time is not under volitional control? Is there an experimental manipulation other than the empirical observation that movement time vary, which would justify that this parameter is linked to biomechanics and noise factors?In general, the work by the groups of Shadmehr and Ahmed on the cost of time and temporal discounting of reward is acknowledged but under-represented. In fact, it is not clear that the explanation of movement time based on biomechanics only is required at all given that the temporal discounting of reward is a well-established phenomenon and that movement time is also at least partially under volitional control. Thus the ability to account for changes in movement time with biomechanical factors is likely valuable for specific conditions.Specific points:Line 70: Stochastic optimal control considers state-feedback adjustments not necessarily linked to corrections for changes relative to a planEquation 3: Why not using the formalism of stochastic differential equations? The proposed equation is slightly abusive as dw almost certainly does not exist (i.e. it has infinite instantaneous variation with probability 1).Equation 9: It would be useful to unpack where the linearization comes from and show the derivation from the original system.Reviewer #2: This study presents a new model for the planning and control of reaching movements. This model combines open-loop control with feedback control and is able to explain both the timing and the variability of reaching movements with and without sensory feedback. To explain the movement duration, a cost function is used that includes both effort and variance resulting from constant and signal-dependent noise in the motor command. The predictions are compared to previously published data and to the data of a new experiment.GeneralModelling the planning and control of reaching movements has a long history in the field of computational motor control. Over the years, many different ingredients have been proposed to be essential, such as open-loop control, feedback control, motor noise, and cost functions that include factors such as smoothness, effort and variance. This model can be considered as a unique combination of earlier proposed ingredients. The approach is therefore not original, but that’s not a problem. If this particular model has more predictive power than earlier models, this is scientific progress and an important result. To what extent that is the case for this study will be evaluated below.The methodology seems generally sound and the paper is generally clearly written, although the writing can be improved as outlined below.SpecificMajor points1. The model proposed here (SFFC) includes a feedforward and a feedback component. The feedforward-only version of the model is used to explain reaching without visual feedback, whereas the full version is used to explain reaching with visual feedback. It is however not correct to consider reaching without vision as reaching without any sensory feedback as there is always proprioceptive feedback. Reaching without any sensory feedback is only possible for deafferented patients. I therefore recommend to use data of such patients to test the predictions the feedforward-only version of the model, and to use the full model with increased sensory variance for reaching without visual feedback (but with proprioception).2. The relevance of a model is determined by the quality of its predictions. The better a model predicts actual behaviour, the more relevant the model. The authors demonstrate that their model can explain many characteristics of human reaching movements. However, Figure 6 shows that there is one important aspect of reaching movements that the model does not predict correctly: the velocity profile. Whereas actual velocity profiles are well-known to be bell shaped, with the peak velocity occurring slightly before half the movement duration (see also the authors’ own data in Fig 6A,B), Figure 6C,D shows that the model predicts atypical velocity profiles with an initially slowly increasing speed, a peak velocity that is reached long after half the movement duration, followed by a rapid decrease to zero. Since these predicted velocity profiles differ strongly from actual velocity profiles, I cannot consider this as a model that predicts all important aspects of reach trajectories correctly. I recommend the authors to try to fix the model so that it predicts more realistic velocity profiles.3. Equation 12 suggests that the model corrects for both task-relevant and task-irrelevant deviations from the nominal plan. Is that correct? This is at odds with stochastic optimal feedback control which corrects only for task-relevant deviations, a phenomenon that has strong experimental support. I recommend modifying the model so that it does not correct for task-irrelevant deviations.4. The proposed model has a number of free parameters. The values of these parameters were determined on the basis of data of Wang et al. (2016), and then later used to make predictions for the new experiment. In principle, that’s a clean approach. However, in L254ff it is mentioned that some of the parameters were changed to compare the model predictions to the data. Are these then still model predictions, or is this just a fit of the model? I found this part of the paper rather vague. There is nothing wrong with fitting your new model, but why then first pretend as if the model was first fit to an independent data set, so that proper predictions can be made?5. To simulate movement with visual feedback, it is assumed that the visual feedback provides information about the full state. However, the state includes joint angles, joint angles velocities and net joint torques (Eq 3). It’s debatable whether vision provides information about joint angles (I don’t look at my elbow during reaching), but it is certainly questionable that it provides information about joint angle velocities. Most problematic however, is the assumption that vision provides feedback about net joint torques - that’s certainly not realistic.Minor pointsThe authors used ad hoc, non-standard acronyms for the various modelling approaches, such as DOC, SOC and SOOC. The logic of these acronyms is not always clear to me. For instance, SOC is used for stochastic optimal feedback control. Why not SOFC, as feedback is in my opinion the most crucial element of this approach?L41-42: “SOC does not account for (…) the larger variance exhibited by movements without vision.” That’s not true, by increasing the level of variance in the sensory signals, this model predicts larger variance.L56: greater -> longerThe Introduction is very long (it ends only after 201 of the 475 lines of text). In my opinion, L28-76 are the genuine introduction, whereas L77-201 could be absorbed into the Methods section.Fig 2, legend: I guess the corrective motor command can also be triggered by the consequences of noise in u.L110-111: “the originality of our proposal is to assume that the primary goal of motor planning is to build a feedforward motor command.” That does not sound very original. All the models not including feedback have this same goal.Eq 3: what is omega_t?L147-149: “Note that hard constraints for the final mean or covariance of the state can also be added in this formulation. We did so for the final mean state to ensure that the arm exactly reaches the desired target on average…” I don’t see why the final mean state had to be set. It’s more natural to use a cost function such as mean squared error that allows for a biases that are accompanied by lower variance. For instance, the well-known undershoot of saccades has been explained this way (Chris Harris), whereas it allowed Liu & Todorov (2007) to explain incomplete corrections for target jumps.L175-176: Please back up the claim that delays in the sensory feedback are not critical.Methods: what time step was used in the model simulations, and do the model predictions depend on this time step?L264: than -> asL279ff: “whose position was displayed on a monitor placed in from of the participant”. Whose position was displayed? Placed in FRONT of the participant? How was the monitor oriented?L295: “and reappeared at the end”. When exactly?L300: So there were only 5 repetitions of each movement? That’s way too few to obtain reliable estimates of the variability. At least several tens of repetitions are needed for this.L307: what was the order of the filter?L329: larger -> longerL333: smaller -> shorterL340: Fig 5C -> Fig 5DL347-400: The results of a large number of statistical tests are reported here. It is however unclear to me why all these tests were conducted. A statistical test is conducted to test a hypothesis. However, none of the tests addresses a hypothesis raised in this study. Instead, the focus of the study is on the model and the extent to which it can explain observed reaching movements. The comparison between predictions and data therefore deserves much more attention than the statistical tests on the data, but very little effort is put in this comparison. In summary, I recommend to remove all the statistical tests on the data and replace these by a critical comparison between the data and model predictions.L409-410: “This learning will minimize feedback…” Unclear what is meant.L417-418: “One original aspect of SFFC is that the feedforward motor command considers uncertainty about the task dynamics…” This is not really original. Harris & Wolpert did this more than 20 years ago.L451: plan -> predictReviewer #3: In the current manuscript, the authors developed a new motor control model that explains the characteristics of timing (speed and duration) and variability of human reaching movements under the conditions of with and without visual feedback. The model which they call SFFC combines a feedforward control (SOOC in their model) and an online feedback control (SOC). The SOOC calculates an expected state trajectory and feedforward motor command in the planning phase and generates the movement in a feedforward manner. The critical point is that the motor command is optimized for the system stochasticity that includes not only signal-dependent noise, which is well known as a determinant of motor control, but also additive motor noise, which has not been given much importance in the field. As these two types of noise have opposing effects on the expected cost when the movement duration becomes longer, the optimal duration to minimize the cost is determined by the balance between the two. The motor planning by the SOOC can explain the timing characteristic that the movement duration increases as the target distance increases. Furthermore, it can account for an interesting observation reported by Wang et al (2016), which previous models fail to explain, that the duration of human preferred movements is greater than the minimum variance duration. On the other hand, the SOC corrects deviations online from the pre-planned state trajectory in a locally optimal manner. The SOC is assumed to be active only when visual feedback is available, which explains the variability characteristic that the variability is reduced in the presence of vision compared to when visual feedback is not available. The validity of their model was evaluated through comparison with human reaching data.I think this model is original in that it explains the timing and variability characteristics without any ad hoc constraints, unlike previous models, but only by considering neuromechanical factors (noise properties, etc.). Also, this study tells us the importance of the motor planning considering additive motor noise in addition to signal-dependent noise. Thus, this study will bring new insight on the computational mechanisms underlying human motor control. However, I have three main concerns that will need to be addressed for this study to make a clear and reliable contribution to the literature.1) First, the authors need more discussion on the currently dominating model, optimal feedback control model (OFC) (Todorov and Jordan, 2002). I agree that the OFC underestimates the importance of motor planning. However, the OFC is still a powerful model that can explain many features of human motor control. In order for the reader to better understand the difference in explanatory power between the OFC and the SFFC, the authors should more clearly mention (A) things the OFC can also explain with some additional assumptions, albeit ad hoc, and (B) things the OFC can explain but SFFC cannot if any.(A): I acknowledge that the authors discuss carefully that the OFC can explain the timing characteristics by adding some assumptions as in previous studies. However, I feel the discussion on the difference between the variability with and without vision is insufficient. I wonder if the OFC could also explain an increased variability in the absence of vision when the weight of the observation noise increases due to only proprioceptive feedback available. If so, the authors should mention it.(B): As the SOC corrects only task-relevant movements (minimal intervention principle), variability (task-irrelevant) in the middle of the movement is greater than at the endpoint, which is consistent with human experimental data (Todorov and Jordan, 2002). In contrast, the SFFC corrects any deviation from a planned trajectory at every moment. Thus, I wonder if the SFFC cound not explain the greater mid-movement variability. I would recommend that the authors add some discussion of this point.2) The second is regarding the interpretation of the experimental data. While the data for variability (Fig 7A,) clearly supports the validity of the model, the data for timing (Fig 7C, E) does not appear to be consistent with the model simulation. On line 456, the authors state that “the fact that movements with and without vision had similar timing in our data reaffirmed the importance of ….”. However, the ANOVA analysis in fact provides marginal significances with p=0.055 (duration) and p=0.052 (peak velocity) for the main effect of the visual condition although the model predicts no difference. Thus, I ask the author to reconsider the interpretation of the data and discuss what possible causes make the difference.3) The third is the contribution of proprioception in the absence of vision. The SFFC assumes that in the absence of vision only the SOOC is active. However, it seems more plausible to assume that the proprioceptive feedback (with a large observation noise) is available and the SOC still works. Although the authors state that the SOC with a large observation noise amounts to use SOOC (lines 454-455), I feel that is a bit extreme. If the SFFC assumes the SOC with a large observation noise is active in the absence of vision, can it explain the increased variability in the absence of vision?Minor points4) Figure1:The simulation result seems to change depending on the amount of additive noise and signal-dependent noise. When I run a simulation with Todorov’s model, I found the endpoint variance increases over the duration when the weight of additive noise is much greater than that of signal-dependent noise. The effect of increasing additive noise on endpoint variance is mentioned in Fig. 1B of Todorov (2005). Thus, the authors may modify the sentence in the caption, “The positional endpoint variance (gray trace) can also be seen to decrease and plateau to a value which mainly corresponds to that of visually-guided movements.”5) Line 46: Please remove “A”.6) Figure 4 AB: What do the black circles indicate? What are MD and VAR?7) Figure 7: The presentation of experimental data requires a measure of variability.**********Have all data underlying the figures and results presented in the manuscript been provided?Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology
data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.Reviewer #1: YesReviewer #2: No:Reviewer #3: Yes**********PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: NoReviewer #2: NoReviewer #3: NoFigure Files:While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at .Data Requirements:Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.Reproducibility:To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions, please see23 Feb 2021Submitted filename: response_letter.pdfClick here for additional data file.24 Mar 2021Dear Dr. Berret,Thank you very much for submitting your manuscript "Stochastic optimal feedforward-feedback control determines timing and variability of arm movements with or without vision" for consideration at PLOS Computational Biology.As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.---------------As you will see below review 1 (and indirectly the response to the previous version by reviewer 2) indicate that the equating of movement without vision and feedforward-only control is erroneous, and that the claim is still present in the manuscript.This needs to be clear in the paper that movement without vision can not be fully explained by feedforward mechanisms. This may require a minor or major rewrite, depending on whether you agree with this statement.Of course it might be possible instead to test experimentally (using deafferented patients, or with perturbation study) whether the feedforward aspect of the model alone can explain explain movement without vision, but that may be going further than your aim with this paper.----------------We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.When you are ready to resubmit, please upload the following:[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).Important additional instructions are given below your reviewer comments.Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.Sincerely,Ulrik R. BeierholmAssociate EditorPLOS Computational BiologySamuel GershmanDeputy EditorPLOS Computational Biology***********************Reviewer's Responses to QuestionsComments to the Authors:Reviewer #1: The authors have extensively revised the manuscript but I admit I feel that the idea of feedforward control corresponding to movements performed without vision is still there. Movements performed without vision are now associated with larger uncertainty but this modification falls a bit short of the more profound change that I believe was necessary.To take a few examples, the abstract states the following: “[feedback] to correct task-related errors which are mostly noticeable when vision is available”. This is wrong and misleading, there is extended evidence for non-visually mediated corrections for task-related errors. In the authors’ summary, it is written “feedforward motor command before the movement starts, which can be observed for example in arm movements without vision […]”. I do not feel it is useful to reiterate that movements without vision are not feedforward. This idea is pushed forward again later in the intro where it is written that: “SOC is designed to account for visually guided movements…” is also wrong. There is a double fallacy in this statement, first it ignores that feedback in SOC models needs not be visual, second it implies that when there is no vision SOC is probably not useful. This pushes the idea that movements without vision are feedforward, which I indicated in my first review does not stand one second in front of evidence that movement without vision are controlled in closed loop. The overall experimental work compares feedforward with movements performed without vision. The statement in the discussion that movement are largely based on feedforward control is again not supported.The main message put forward is that changes in movement time can be captured by considering open loop minimization in the brain. This is interesting, but that does not imply that a feedforward controller is derived, much less executed. The experimental validation is not adequate and therefore the model is not supported. It is also conceivable that movement times are selected based on prior experience (including optimisation criteria, this is a contribution of the authors’), and that the implicit selection of movement time is used in feedback controllers with or without vision. To argue that movement performed without vision are similar to feedforward and establish that the selection of movement time could be due to the optimization of a feedforward controller requires at least a task that involves perturbations. In support to their model the authors must demonstrate in a perturbation task that there is less or no feedback correction when there is no vision.Much of the authors’ reasoning to establish a similarity between no-vision and feedforward control is based on movement variability, but this reasoning is inconsistent with the work of Ito and Gomi (2020, eLife). The authors argumentation imply that corrections are reduced when vision is withheld, but there is another phenomenon, that is the gain of long-latency responses in this case in fact increased. Long-latency responses are not impedance control, and responses in the same time window increase without vision. This produced an increase in variance, but not due to less feedback correction, instead it was due to more vigorous corrections. This is a sufficient piece of evidence to reject the authors’ model. That the model produces movement times that are consistent with human behaviour is interesting but the analogy stops there.Other major points:1. The limitation of SOC relative to changes in movement time is a straw-man, it simply follows from finite horizon formulation, so this model is not supposed to produce changes in movement time. The argumentation must be reworked throughout.2. Aspects of model presentation must be clarified: the procedure considering “fixed or free movement time” (just prior to Eqn. 4) was unclear and should be sketched out; pls sketch the argument why P(t) appears in the deterministic part of the SDE; it is unclear what manipulation was performed with the hard constraints that make the terminal penalty cost redundant or useless, what was the manipulation exactly?3. Page 7 middle: pls clarify, it is suggested that deafferented patients have both similar and higher level of variability, this section was unclear. The argument that variability in deafferented patients is comparable in some conditions to control groups is not telling that they use the same control strategy, again a perturbation paradigm would clearly highlight differences between the two.4. The correlation between timings of movements performed with or without vision is taken as evidence that they can share the same feedforward control command, an equally valid conclusion is that the consistent timing originated from the same state-feedback controller. This does not undermine the authors’ efforts to show that movement times may be selected optimally, although there is no conclusive evidene that this selection is based on a feedforward controller.5. While the model parameters have likely been selected carefully it is necessary to perform a sensitivity analysis to assess the robustness of the model simulations against variations in parameter settings.Reviewer #3: The new version submitted by the authors contains significant updates in their model as well as additional explanation/discussion on the model simulation and experimental data to answer to the points we raised in our first review. I appreciate the additional effort and think that the authors improved the quality of the paper which I consider now suitable for publication.**********Have all data underlying the figures and results presented in the manuscript been provided?Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology
data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.Reviewer #1: YesReviewer #3: Yes**********PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: NoReviewer #3: NoFigure Files:While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at .Data Requirements:Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.Reproducibility:To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols13 Apr 2021Submitted filename: response_letter_rev2.pdfClick here for additional data file.5 May 2021Dear Dr. Berret,We are pleased to inform you that your manuscript 'Stochastic optimal feedforward-feedback control determines timing and variability of arm movements with or without vision' has been provisionally accepted for publication in PLOS Computational Biology.Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology.Best regards,Ulrik R. BeierholmAssociate EditorPLOS Computational BiologySamuel GershmanDeputy EditorPLOS Computational Biology***********************************************************Reviewer's Responses to QuestionsComments to the Authors:Reviewer #1: The authors made commendable efforts to take all previous concerns into account and I recommend that the manuscript be accepted for publication,**********Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.Reviewer #1: Yes**********PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: No8 Jun 2021PCOMPBIOL-D-20-02130R2Stochastic optimal feedforward-feedback control determines timing and variability of arm movements with or without visionDear Dr Berret,I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!With kind regards,Katalin SzaboPLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol