Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT

User menu

Search

  • Advanced search
eNeuro

eNeuro

Advanced Search

 

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT
PreviousNext
Research ArticleNew Research, Sensory and Motor Systems

Individual Differences in Motor Noise and Adaptation Rate Are Optimally Related

Rick van der Vliet, Maarten A. Frens, Linda de Vreede, Zeb D. Jonker, Gerard M. Ribbers, Ruud W. Selles, Jos N. van der Geest and Opher Donchin
eNeuro 16 July 2018, 5 (4) ENEURO.0170-18.2018; DOI: https://doi.org/10.1523/ENEURO.0170-18.2018
Rick van der Vliet
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
2Department of Rehabilitation Medicine, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Rick van der Vliet
Maarten A. Frens
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
3Erasmus University College, 3011 HP, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Maarten A. Frens
Linda de Vreede
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Zeb D. Jonker
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
2Department of Rehabilitation Medicine, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Zeb D. Jonker
Gerard M. Ribbers
2Department of Rehabilitation Medicine, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
4Rijndam Rehabilitation Centre, 3015 LJ, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Gerard M. Ribbers
Ruud W. Selles
2Department of Rehabilitation Medicine, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
5Department of Plastic and Reconstructive Surgery, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Ruud W. Selles
Jos N. van der Geest
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Jos N. van der Geest
Opher Donchin
1Department of Neuroscience, Erasmus MC, 3015 CN, Rotterdam, The Netherlands
6Department of Biomedical Engineering and Zlotowski Center for Neuroscience, Ben Gurion University of the Negev, Be’er Sheva, 8499000 Israel
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Opher Donchin
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Individual variations in motor adaptation rate were recently shown to correlate with movement variability or “motor noise” in a forcefield adaptation task. However, this finding could not be replicated in a meta-analysis of adaptation experiments. Possibly, this inconsistency stems from noise being composed of distinct components that relate to adaptation rate in different ways. Indeed, previous modeling and electrophysiological studies have suggested that motor noise can be factored into planning noise, originating from the brain, and execution noise, stemming from the periphery. Were the motor system optimally tuned to these noise sources, planning noise would correlate positively with adaptation rate, and execution noise would correlate negatively with adaptation rate, a phenomenon familiar in Kalman filters. To test this prediction, we performed a visuomotor adaptation experiment in 69 subjects. Using a novel Bayesian fitting procedure, we succeeded in applying the well-established state-space model of adaptation to individual data. We found that adaptation rate correlates positively with planning noise (β = 0.44; 95% HDI = [0.27 0.59]) and negatively with execution noise (β = –0.39; 95% HDI = [–0.50 –0.30]). In addition, the steady-state Kalman gain calculated from planning and execution noise correlated positively with adaptation rate (r = 0.54; 95% HDI = [0.38 0.66]). These results suggest that motor adaptation is tuned to approximate optimal learning, consistent with the “optimal control” framework that has been used to explain motor control. Since motor adaptation is thought to be a largely cerebellar process, the results further suggest the sensitivity of the cerebellum to both planning noise and execution noise.

  • Bayesian statistics
  • cerebellum
  • motor learning
  • noise
  • optimal control
  • visuomotor adaptation

Significance Statement

Our study shows that the adaptation rate is optimally tuned to planning noise and execution noise across individuals. This suggests that motor adaptation is tuned to approximate optimal learning, consistent with “optimal control” approaches to understanding the motor system. In addition, our results imply sensitivity of the cerebellum to both planning noise and execution noise, an idea not previously considered. Finally, our Bayesian statistical approach represents a powerful, novel method for fitting the well-established state-space models that could have an influence on the methodology of the field.

Introduction

As children we all learned: some of us move with effortless grace and others are frankly clumsy. Underlying these differences are natural variations in acquiring, calibrating, and executing motor skill, which have been related to genetic (Frank et al., 2009; Fritsch et al., 2010; McHughen et al., 2010) and structural (Tomassini et al., 2011) factors. Recently, it has been suggested that differences between individuals in the rate of motor adaptation (i.e. the component of motor learning responsible for calibrating acquired motor skills to changes in the body or environment; Shadmehr et al., 2010), correlate with movement variability, or motor noise (Wu et al., 2014). However, this finding was not supported by a recent meta-analysis of adaptation experiments (He et al., 2016). This inconsistency may arise because motor noise has multiple components with differing relations to adaptation rate. Our study characterizes the relationship between adaptation rate and motor noise and suggests that adaptation rate varies optimally between individuals in the face of multiple sources of motor variability.

Motor noise has many physiologic sources, such as motor preparation noise in (pre)motor networks, motor execution noise, and afferent sensory noise (Faisal et al., 2008). Modeling (Cheng and Sabes, 2006, 2007; van Beers, 2009) and physiologic (Churchland et al., 2006; Chaisanguanthum et al., 2014) studies have divided the multiple sources of motor noise into planning noise and execution noise (see Fig. 1A). Planning noise is believed to arise from variability in the neuronal processing of sensory information, as well as computations underlying adaptation and maintenance of the states in time (Cheng and Sabes, 2006, 2007). Indeed, electrophysiological studies in macaques show that activity in (pre)motor areas of the brain is correlated with behavioral movement variability (Churchland et al., 2006; Chaisanguanthum et al., 2014). Similar results have also been seen in humans using fMRI (Haar et al., 2017). In contrast, execution noise apparently originates in the sensorimotor pathway. In the motor pathway, noise stems from the recruitment of motor units (Harris and Wolpert, 1998; Jones et al., 2002; van Beers et al., 2004). Motor noise is believed to dominate complex reaching movements with reliable visual information (van Beers et al., 2004). In addition, sensory noise stems from the physical limits of the sensory organs and has been proposed to dictate comparably simpler smooth pursuit eye movements (Bialek, 1987; Osborne et al., 2005). Planning and execution noise might affect motor adaptation rate in different ways.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Planning and execution noise have opposing effects on visuomotor adaptation. A, State-space model of visuomotor adaptation. The aiming angle on trial 2 Embedded Image is a linear combination of the aiming angle on the previous trial Embedded Image multiplied by a retentive factor A minus the error Embedded Image on the previous trial multiplied with adaptation rate B . In addition, the aiming angle is distorted by the random process η (planning noise). The actual movement angle Embedded Image is the aiming angle Embedded Image distorted by the random process ϵ (execution noise). The error Embedded Image is the sum of the movement direction Embedded Image and the external perturbation Embedded Image . B, Planning noise and optimal adaptation rate Embedded Image (defined as the Kalman gain). The optimal adaptation rate increases with planning noise Embedded Image . In this figure, Embedded Image was kept constant at Embedded Image . C, Execution noise and optimal adaptation rate Embedded Image (defined as the Kalman gain). The optimal adaptation rate decreases with execution noise Embedded Image . In this figure, Embedded Image was kept constant at Embedded Image . D, Simulated optimal learners. At trial 110, a perturbation (black line) is introduced that requires the optimal learners to adapt their movement. The gray learner has low planning noise Embedded Image and execution noise Embedded Image . The red learner has a higher planning noise Embedded Image than the gray learner Embedded Image . This causes the red learner to adapt faster. The green learner has a higher execution noise than the gray learner Embedded Image . This causes the green learner to adapt more slowly. For all learners, the thick line shows the average, and the thin line, a single noisy realization.

Motor adaptation has long been suspected to be sensitive to planning noise and execution noise. Models of visuomotor adaptation incorporating both planning and execution noise have been shown to provide a better account of learning than single noise models (Cheng and Sabes, 2006, 2007; van Beers, 2009). In addition, manipulating the sensory reliability by blurring the error feedback, effectively increasing the execution noise, can lower the adaptation rate (Baddeley et al., 2003; Burge et al., 2008; Wei and Körding, 2010; van Beers, 2012), whereas manipulating state estimation uncertainty by temporarily withholding error feedback, effectively increasing the planning noise, can elevate the adaptation rate (Wei and Körding, 2010). These studies not only suggest that adaptation rate is tuned to multiple sources of noise, but also indicate that this tuning process is optimal and can therefore be likened to a Kalman filter (Kalman, 1960). Possibly, differences in adaptation rate between individuals correlate with planning noise and execution noise according to the same principle, predicting faster adaptation for people with more planning noise and slower adaptation for people with more execution noise (He et al., 2016; Fig. 1B–D).

To test the relation between adaptation rate and planning noise and execution noise across individuals, we performed a visuomotor adaptation experiment in 69 healthy subjects. We fitted a state-space model of trial-to-trial behavior (Cheng and Sabes, 2006, 2007) using Bayesian statistics to extract planning noise, execution noise, and adaptation rate for each subject. We show that the adaptation rate is sensitive to both types of noise and that this sensitivity matches predictions based on Kalman filter theory.

Materials and methods

Subjects

We included 69 right-handed subjects between October 2016 and December 2016, without any medical conditions that might interfere with motor performance (14 men and 55 women; mean age = 21 years, range 18–35 years; mean handedness score = 79; range 45–100). Subjects were recruited from the Erasmus MC University Medical Center and received a small financial compensation. The study was performed in accordance with the Declaration of Helsinki and approved by the medical ethics committee of the Erasmus MC University Medical Center.

Experimental procedure

Subjects were seated in front of a horizontal projection screen while holding a robotic handle in their dominant right hand (Donchin et al., 2012). The projection screen displayed the location of the robotic handle (“the cursor”; yellow circle 5-mm radius), start location of the movement (“the origin”, white circle 5-mm radius), and target location of the movement (“the target”, white circle 5-mm radius) on a black background (see Fig. 2A). The position of the origin on the screen was fixed throughout the experiment, ∼40 cm in front of the subject at elbow height, while the target was placed 10 cm from the origin at an angle of –45°, 0°, or 45°. To remove direct visual feedback of hand position, subjects wore an apron that was attached to the projection screen around their necks.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Measurements of planning and execution noise and adaptation rate in a visuomotor adaptation experiment. A, Setup. The projection screen displayed the location of the robotic handle (“the cursor”), start location of the movement (“the origin”), and target of the movement (“the target”) on a black background. The position of the origin on the screen was fixed throughout the experiment, while the target was placed 10 cm from the origin at an angle of –45°, 0°, or 45°. B, Trial types. The experiment included vision unperturbed and perturbed trials and no-vision trials. In vision unperturbed trials, the cursor was shown at the position of the handle during the movement. The cursor was also visible in vision perturbed trials, but at a predefined angle from the vector connecting the origin and the handle. In no-vision trials, the cursor was turned off when movement onset was detected and therefore only visible at the start of movement to help subjects keep the cursor at the origin. C, Experimental design. The baseline block consisted of 225 vision unperturbed trials and 225 no-vision trials (indicated by vertical red lines). The perturbation block had 50 no-vision trials and 400 vision trials, with every block of nine trials containing 1 no-vision trial. Most vision trials were perturbed vision trials whose perturbation magnitudes formed a staircase running from –9° to 9°. D, Simulation of planning noise Embedded Image and standard deviation Embedded Image of the movement angle. Embedded Image increases with Embedded Image . Calculated for Embedded Image and Embedded Image with Embedded Image for the solid line and Embedded Image for the dashed line. E, Simulation of planning noise Embedded Image and lag-1 autocorrelation Embedded Image of the movement angle. Embedded Image increases with Embedded Image . Calculated for Embedded Image and Embedded Image with Embedded Image for the solid line and Embedded Image for the dashed line. F, Simulation of execution noise Embedded Image and standard deviation Embedded Image of the movement angle. Embedded Image increases with Embedded Image . Calculated for Embedded Image and Embedded Image with Embedded Image for the solid line and Embedded Image for the dashed line. G, Simulation of execution noise Embedded Image and lag-1 autocorrelation Embedded Image of the movement angle. Embedded Image decreases with Embedded Image . Calculated for Embedded Image and Embedded Image with Embedded Image for the solid line and Embedded Image for the dashed line. H. Simulated learners without vision. The green and red traces show a single realization of two learners with either high planning noise (red learner Embedded Image and Embedded Image ) or high execution noise (green learner Embedded Image and Embedded Image ). Both sources increase the movement noise, but planning noise leads to correlated noise, whereas execution noise leads to uncorrelated noise. This property can be seen from the relation between sequential trials. For the red learner, sequential trials are often in the same (positive or negative) direction. For the green learner, sequential trials are in random directions. This is captured by the lag-1 autocorrelation. I, Simulation of Embedded Image between the perturbation p and movement angle y , and adaptation rate B . Embedded Image gets more negative for increasing B (simulated with Embedded Image ). J, Simulated learners with perturbation. The gray and blue lines show a simulated slow (Embedded Image , Embedded Image ) and fast (Embedded Image , Embedded Image ) learner. The fast learner tracks the perturbation signal more closely than the slow learner. This property is captured by the covariance between the perturbation and the movement angle.

Subjects were instructed to make straight shooting movements from the origin toward the target and to decelerate only when they passed the target. A trial started with the presentation of the target and ended when the distance between the origin and cursor was at least 10 cm or when trial duration exceeded 2 s. At this point, movements were damped with a force cushion (damper constant 3.6 Ns/m, ramped up over 7.5 ms) and the cursor was displayed at its last position until the start of the next trial to provide position error feedback. Furthermore, timing feedback was given to keep trial duration (see definition below) in a tight range. The target dot turned blue if trial duration on a particular trial was too long (>600 ms) and red if trial duration was too short (<400 ms) and remained white if trial duration was in the correct time range (400–600 ms). During presentation of position and velocity feedback, the robot pushed the handle back to the starting position. Forces were turned off when the handle was within 0.5 cm from the origin. Concurrently, the cursor was projected at the position of the handle again and subjects had to keep the cursor within 0.5 cm from the origin for 1 s to start the next trial.

The experiment included vision unperturbed, vision perturbed, and no-vision trials (see Fig. 2B). In vision unperturbed trials, the cursor was shown at the position of the handle during the movement. The cursor was also visible in vision perturbed trials but at a predefined angle from the vector connecting the origin and the handle. In no-vision trials, the cursor was turned off when movement onset was detected (see below) and was visible only at the start of the trial to help subjects keep the cursor at the origin.

The entire experiment lasted 900 trials with all three target directions (angle of –45°, 0°, or 45°) occurring 300 times in random order. The three different trial types were used to build a baseline and a perturbation block (see Fig. 2C). We designed the baseline block to obtain (1) reliable estimates of the noise parameters and (2) variance statistics (standard deviation and lag-1 autocorrelation of the movement angle) related to the noise parameters. Therefore, we included a large number of no-vision trials (225 no-vision trials) as well as vision unperturbed trials (225 vision unperturbed trials). The order of the vision unperturbed trials and no-vision trials was randomized except for trials 181–210 (no-vision trials) and trials 241–270 (vision unperturbed trials). We designed the perturbation block to obtain (1) reliable estimates of the adaptation parameters and (2) variance statistics related to trial-to-trial adaptation (covariance between perturbation and movement angle). The perturbation block consisted of a large number of vision trials (400 vision trials) and a small number of no-vision trials (50 no-vision trials), with every block of nine trials containing one no-vision trial. Every 8 to 12 trials, the perturbation angle changed with an incremental 1.5° step. These steps started in the positive direction until reaching 9° and then switched sign to continue in the opposite direction until reaching –9°. This way, a perturbation signal was constructed with three “staircases” lasting 150 trials each (see Fig. 2C). Design of the gradual perturbation was optimized to provide a “rich” input for system identification, without sacrificing the consistency of the signal too much, as this has been shown to negatively affect the adaptation rate (Gonzalez Castro et al., 2014; Herzfeld et al., 2014), and is similar to the perturbation used by Cheng and Sabes (2007). The experiment was briefly paused every 150 trials.

Data collection

The experiment was controlled by a C++ program developed in-house. Position and velocity of the robot handle were recorded continuously at a rate of 500 Hz. Velocity data were smoothed with an exponential moving average filter (smoothing factor = 0.18 s). Trials were analyzed from movement start (defined as the time point when movement velocity exceeds 0.03 m/s) to movement end (defined as the time point when the distance from the origin is equal to or larger than 9.5 cm). Reaction time was defined as the time from trial start until movement start, movement duration as the time from movement start until trial end and trial duration as the time from trial start until trial end. Movement angle was calculated as the signed (+ or –) angle in degrees between the vector connecting origin and target and the vector connecting robot handle position at movement start and movement end. The clockwise direction was defined as positive. Peak velocity was found by taking the maximum velocity in the trial interval. Trials with (1) a maximal displacement below 9.5 cm, (2) an absolute movement direction larger than 30°, or (3) a duration longer than 1 s were removed from further analysis (2% of data).

Visuomotor adaptation model

Movement angle was modeled with the following state-space equation (see Fig. 1A; Cheng and Sabes, 2006, 2007):Embedded Image (1) Embedded Image (2) Embedded Image (3) Embedded Image (4)

In this model, Embedded Image is the aiming angle (the movement plan), and Embedded Image is the movement angle (the actually executed movement). Error e Embedded Image on a particular trial is the sum of Embedded Image and the perturbation Embedded Image . The learning terms are A , which represents retention of the aiming angle over trials, and adaptation rate B , the fractional change from error Embedded Image . The movement angle is affected by planning noise process η , modeled as a zero-mean Gaussian with standard deviation Embedded Image , and execution noise process ϵ , modeled as a zero-mean Gaussian with standard deviation Embedded Image .

Statistics

Our statistical approach is a Bayesian approach (an excellent introduction to Bayesian statistics for a nontechnical audience can be found in Kruschke (2010)). We used this approach to fit the state-space model described in Eqs. (1)–(4) because it offers a number of advantages over the expectation-maximization algorithm used in previous studies (Cheng and Sabes, 2006, 2007). Perhaps the most important advantage of the Bayesian approach is that it naturally allows hierarchical modeling that shares data across subjects, allowing greater regularization of the parameter fits for each subject, as well as simultaneous estimates of the population distribution of the parameters (Browne and Draper, 2006; Gelman, 2006). In a classic approach, each subject’s parameters are generally estimated independently, and the uncertainty in those estimates is often not propagated forward when calculating population estimates. Indeed, the output of a Bayesian approach is not the best possible estimate of the parameter or even a maximum-likelihood estimate with a confidence interval, but rather a sampling from the parameter’s probability distribution given the data (Kruschke and Liddell, 2018a). This allows the analysis to naturally refocus on parameter uncertainty rather than focusing on point estimates (Kruschke, 2013; Wagenmakers et al., 2014; Kruschke and Liddell, 2018b). The difficulty with point estimates has been a focus of much debate in the current discussion of the reproducibility crisis in science (Ioannidis, 2005; Cumming, 2014). The Bayesian approach also estimates the hidden (state) variables simultaneously with the parameters, rather than creating a somewhat arbitrary distinction between imputation and estimation (Carlin et al., 1992; Carter and Kohn, 1994). This allows analysis of how the state variable estimates change with the parameter estimates, an analysis that is tricky to do with an expectation-maximization approach. Finally, the Bayesian approach allows great flexibility in specifying the form of the model (Kruschke and Liddell, 2018a). This can be useful in defining constraints on the model parameters or transforming variables to lie in more relevant parameter spaces, as defined below.

Modern Bayesian approaches rely on a family of algorithms called the Markov chain Monte Carlo (MCMC) algorithms (Andrieu et al., 2003). These algorithms require definitions of the likelihood function (how the data would be generated if we knew the parameters), the prior probability for the parameters (generally chosen to be broad and uninformative, but see below), and return samples from the posterior joint-probability function of the parameters. Thus, once the model and priors are specified, the output of the MCMC algorithm is a large matrix where each row is a sample and each column is one of the parameters in the model. These samples can be, then, summarized in different ways to generate parameter estimates (usually the mean of the samples but often the mode) and regions of uncertainty (very often a 95% region called the high-density interval (HDI) which contains 95% of the posterior samples but also obeys the criterion that every sample in the HDI is more probable than every sample outside of it). They can also be used to assess asymmetry in the parameter distributions and covariance in the parameter estimates.

As outlined above, the Bayesian approach to state-space modeling we have taken requires us to define priors on the model parameters. We will justify our choices in the following section. The adaptation parameters Embedded Image and retention parameters Embedded Image were sampled in the logistic space instead of the regular 0-1 space:Embedded Image (5)

The logistic space spreads the range from 0-1 all the way from Embedded Image to Embedded Image . This means that the distance between 0.1 and 0.01 and 0.001 are all similar in the logistic space, as are the distances between 0.9, 0.99 and 0.999. This space, thus, reflects much more accurately the real effects of changes in the parameter than if we sampled in the untransformed space. This leads to much better sampling behavior and, thus, greater accuracy and less bias in the results. The priors for Embedded Image and Embedded Image were not actually specified in the description of the model. Only their shape was determined (normal in the logistic space). The actual prior was chosen by sampling hyperparameters for these normal distributions. For the hyperparameters, we did need to choose a specific prior, and here we choose highly uninformative priors to allow the posterior distribution to be influenced primarily by the data:Embedded Image (6) Embedded Image (7)

The sensitivity analysis (described below) showed that the choice to sample Embedded Image and Embedded Image from a normal distribution in the logistic space had no strong effect on the results. Following the standard Bayesian approach (Kruschke, 2010), we sampled the precision (inverse of the variance) and used a very broad gamma distribution as a prior for the precision:Embedded Image (8)

One reason the gamma distribution is a popular prior for the precision is that it is a conjugate prior which makes the algorithm more efficient. In any case, other choices of prior did not change our results in a meaningful way (see sensitivity analysis below).

MCMC sampling for the Bayesian state-space model was implemented in OpenBUGS (v. 3.2.3, OpenBUGS Foundation, available from: http://www.openbugs.net/w/Downloads) with three 50,000 samples chains and 20,000 burn-in samples. A single estimate per subject s was made for Embedded Image and Embedded Image , Embedded Image and Embedded Image . We used all 150,000 MCMC samples that represent the posterior distribution of the model parameters Embedded Image , Embedded Image , and Embedded Image given the data to calculate linear regressions and correlations between the model parameters across subjects. Results were presented as the mode of the effect size (either the correlation coefficient r or regression coefficient β) with 95% HDIs. Parameter estimates are plotted as the mode with 68% HDIs, similar to the standard deviation interval.

To demonstrate the test-retest properties of the Bayesian state-space model, we simulated two datasets with 50 learners on the visuomotor adaptation task outlined above. The first (optimal) dataset was simulated by drawing model parameters from the following distributions: Embedded Image , Embedded Image , and Embedded Image , and calculating Embedded Image as the Kalman gain. The goal of this analysis was to determine the test-retest correlations of the model parameters Embedded Image , Embedded Image , and Embedded Image and the ability to correctly estimate the relations between Embedded Image and the noise parameters. For the second (permuted) dataset, A [s], Embedded Image , and Embedded Image were kept constant but Embedded Image was permuted between learners. The motivation for this analysis was to show that our Bayesian state-space model does not introduce false relations between B and the noise parameters.

To evaluate the sensitivity of the main results to alternate prior distributions for the Bayesian state-space model, we repeated the entire analysis with (alternative priors 1) t-distributions with the hyperparameter for the degrees of freedom sampled from an exponential distribution (in line with recommendations from Kruschke (2013)) as priors for Embedded Image and Embedded Image ; (alternative priors 2) t-distributions as priors for Embedded Image and Embedded Image , and uniform distributions in the range Embedded Image as priors for Embedded Image and Embedded Image (in line with recommendations from Gelman (2006)); and (alternative priors 3) beta distributions with hyperparameters sampled from gamma distributions as priors for Embedded Image and Embedded Image and uniform distributions as priors for Embedded Image and Embedded Image . Finally, we addressed the concern that the between-subjects correlations of the model parameters might arise from within-subject correlations of the model parameters by permuting the MCMC samples differently for each parameter and recalculating the correlation and regression coefficients. The permuted distribution of the model parameters has the property that all correlations between the parameters within subjects are zero.

Code accessibility

BUGS/JAGS code for the Bayesian state-space model can be accessed without restrictions at: https://github.com/rickvandervliet/Bayesian-state-space.

Results

Simulations

We designed a visuomotor adaptation task (Tseng et al., 2007) to (1) fit the state-space model of adaptation and (2) investigate the validity of the parameter estimates Embedded Image , Embedded Image , and Embedded Image by correlating the estimates with the variance statistics of the data (see Fig. 2A–C).

The baseline block was designed to extract the standard deviation and the lag-1 autocorrelation of the movement direction and relate these measures to the parameter estimates of Embedded Image and Embedded Image . The standard deviation and lag-1 autocorrelation in our baseline block are well approximated by the following expressions:Embedded Image (9) Embedded Image (10)

In addition, we included a control segment of 30 trials without vision ( B = 0), to calculate estimates of the standard deviation and lag-1 autocorrelation which are independent of the adaptation rate B :Embedded Image (11) Embedded Image (12)

For both the expressions with vision (9)–(10) (solid lines) and without vision (11)–(12) (dashed lines), standard deviation Embedded Image increases with planning noise Embedded Image (see simulations in Fig. 2D) and execution noise Embedded Image (see simulations in Fig. 2F) whereas lag-1 autocorrelation Embedded Image increases with planning noise Embedded Image (see simulations in Fig. 2E) but decreases with execution noise Embedded Image (see simulations in Fig. 2G), with the strongest correlations between Embedded Image and Embedded Image , and Embedded Image and Embedded Image . We therefore expected similar relations between the noise parameters Embedded Image and Embedded Image , and the standard deviation Embedded Image and lag-1 autocorrelation Embedded Image of the baseline block (see simulations of planning and execution noise in the baseline block in Fig. 2H).

The perturbation block was designed to extract the covariance Embedded Image between the perturbation and the movement angle from the data and relate this parameter to the adaptation rate B . The covariance Embedded Image depends solely on the learning parameters A and B and becomes increasingly negative for higher adaptation rates because learning is compensatory (see simulations in Fig. 2I). Therefore, we expected a similar relation between the covariance Embedded Image and adaptation rate Embedded Image in the perturbation block of our experiment (see simulations of two learners with a low or high adaptation rate in Fig. 2J).

Next, we designed a Bayesian state-space model to estimate the model parameters. To demonstrate the test-retest properties of this approach, we simulated one dataset with optimal learners and one dataset wherein the adaptation rate of the optimal dataset was permuted across learners. Excellent test-retest correlations were found in both the optimal dataset (Embedded Image r = 1.00; 95% HDI = [1.00 1.00], Embedded Image r = 0.89; 95% HDI = [0.85 0.93], and Embedded Image r = 0.99; 95% HDI = [0.98 0.99]) and the permuted dataset (Embedded Image r = 1.00; 95% HDI = [1.00 1.00], Embedded Image r = 0.90; 95% HDI = [0.86 0.93], and Embedded Image r = 0.99; 95% HDI = [0.98 0.99]). In the optimal dataset, the Bayesian state-space model was able to uncover the relations between Embedded Image and the noise parameters Embedded Image β = 0.73; 95% HDI = [0.68 0.77] (see Fig. 3A) and Embedded Image β = –0.44; 95% HDI = [–0.51 –0.38]), which were 0.81 and –0.53 in the simulated data (see Fig. 3B). In the permuted dataset, the Bayesian state-space model did not falsely introduce relations between Embedded Image and the noise parameters Embedded Image β = 0; 95% HDI = [–0.09 0.08] (see Fig. 3C) and Embedded Image β = –0.01; 95% HDI = [–0.04 0.02]), as they were –0.01 and –0.04 in the original dataset (see Fig. 3D). Therefore, the Bayesian state-space model can reliably estimate the model parameters and the regression coefficients between the noise terms and the adaptation rate.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Test-retest properties of the Bayesian state-space model. A, B, Regression of Embedded Image onto Embedded Image (A) and Embedded Image (B) for the simulated optimal dataset. C, D, Regression of Embedded Image onto Embedded Image (C) and Embedded Image (D) for the simulated permuted dataset. Parameter estimates with 68% HDIs are shown for every simulated learner as a dot with error bars. The black solid line shows the regression on the model parameters estimated with the Bayesian state-space model, the green dashed line the regression on the original model parameters.

Experimental results

Sixty-nine subjects performed the visuomotor adaptation task outlined above. Overall, participants started moving 230 ms, IQR = [211 254] ms, after target presentation and completed the movement in 290 ms, IQR = [251 320] ms, resulting in a trial duration of 520 ms, IQR = [500 534] ms with 87% of trials IQR = [84 95]% in the correct time window between 400 and 600 ms. Standard deviation of movement angle calculated across the 69 subjects illustrates the differences in movement behavior between people (Fig. 4A). The group average aiming angle Embedded Image , calculated from 1,000 samples of the posterior distribution using the model (green dotted line), shows good agreement with the group average movement angle calculated directly from the data (brown solid line).

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

State-space model of visuomotor adaptation. A, Visuomotor adaptation. Average movement angle of the 69 subjects with standard deviations are shown in brown tone colors. The black line indicates the average perturbation signal, and the green line, the average posterior estimate of the aiming angle. B, Planning noise examples. The gray line shows a subject with low planning noise (Embedded Image ), and the red line, a subject with high planning noise (Embedded Image ). C, Execution noise examples. The gray line shows a subject with low execution noise (Embedded Image ), and the green line, a subject with high execution noise (Embedded Image ). D, Relation between the parameter estimate Embedded Image and baseline measure Embedded Image . The black line is a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image . E, Relation between the parameter estimate Embedded Image and baseline measure Embedded Image . The black line is a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image . F, Relation between the parameter estimate Embedded Image and baseline measure Embedded Image . The black line is a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image . G, Relation between the parameter estimate Embedded Image and baseline measure Embedded Image . The black line is a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image . H, Adaptation rate examples. The thick lines show a slow (gray, Embedded Image ) and fast (blue, Embedded Image ) subject smoothed with a 6th-order Butterworth filter. The black shows the perturbation signal for the fast subject. I, Relation between the parameter estimate Embedded Image and perturbation block estimate Embedded Image . Parameter estimates and 68% HDIs are shown for every subject as a dot with error bars.

Fig. 4B, C show example subjects with low or high planning noise Embedded Image (see Fig. 4B) and low or high execution noise Embedded Image (see Fig. 4C). We calculated the standard deviation and lag-1 autocorrelation using all trials in the baseline block and regressed these estimates onto Embedded Image and Embedded Image . Agreeing with our group-level predictions (see Fig. 2D–G), we found a positive relation between planning noise Embedded Image and standard deviation Embedded Image (β = 0.18; 95% HDI = [0.11 0.24]; see Fig. 4D), between planning noise Embedded Image and lag-1 autocorrelation Embedded Image (β = 0.42; 95% HDI = [0.29 0.55]; see Fig. 4E) and between execution noise Embedded Image [s] and standard deviation Embedded Image (β = 0.91; 95% HDI = [0.87 0.94]; see Fig. 4F) and a negative relation between execution noise Embedded Image and lag-1 autocorrelation Embedded Image (β = –0.14; 95% HDI = [–0.24 –0.07]; see Fig. 4G). Next, we calculated the standard deviation and lag-1 autocorrelation of trials 181–210 only, which are no-vision trials where adaptation rate B = 0. Here, we found similar correlations between (1) planning noise Embedded Image and standard deviationEmbedded Image (β = 0.12; 95% HDI = [–0.04 0.27]; (2) planning noise Embedded Image and lag-1 autocorrelation Embedded Image (β = 0.22; 95% HDI = [0.07 0.35]; (3) execution noise Embedded Image [s] and standard deviation Embedded Image (β = 0.44; 95% HDI = [0.39 0.49]), and (4) execution noise Embedded Image and lag-1 autocorrelation Embedded Image (β = –0.04; 95% HDI = [–0.10 –0.01]). Example subjects with a low and high adaptation rate are shown in Fig. 4H. Again, according to the model prediction (see Fig. 2I), we found a negative relation between adaptation rate Embedded Image and covariance Embedded Image on a group level (r = –0.69; 95% HDI = [–0.78 –0.60]; see Fig. 4I).

Next, we investigated the relation between adaptation rate and the noise terms. The results are illustrated with scatterplots of the parameter estimates for individual subjects (Fig. 5, left column), heatmaps of the parameter estimate distributions for the entire population (Fig. 5, middle column), and line plots of the regression and correlation coefficient densities (Fig. 5, right column). We regressed Embedded Image onto Embedded Image and Embedded Image and found a positive relation between Embedded Image and Embedded Image (β = 0.44; 95% HDI = [0.27 0.59]; see Fig. 5A–C) and a negative relation between Embedded Image and Embedded Image (β = –0.39; 95% HDI = [–0.50 –0.30]; see Fig. 5D–F) with a variance explained of 0.32; 95% HDI = [0.19 0.45]. This finding indicates that a significant proportion of the difference in adaptation rate between individuals can be explained from differences in their planning and execution noise with the direction of the correlations in agreement with Kalman filter theory (see Fig. 1B, C). In addition, we determined the steady-state Kalman gain for every subject from Embedded Image , Embedded Image , and Embedded Image and correlated the steady-state Kalman gain with Embedded Image . Steady-state Kalman gain was calculated by solving the Riccati equation for the steady-state covariance Embedded Image :Embedded Image (13) Embedded Image (14)

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Relation between noise and adaptation rate. A, D, G, Scatter plots of individual parameter estimates. Parameter estimates and 68% HDIs are shown for every subject as a dot with error bars. The black line is a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image , a linear regression of Embedded Image onto Embedded Image and Embedded Image for average Embedded Image (D) and the correlation between Embedded Image and Embedded Image (G). B, E, H, Heatmaps of the parameter estimate distributions. The heatmaps illustrate the distribution of the parameter estimates for the entire population of 69 subjects. The intensity represents the percentage of samples in a specific range for (B) Embedded Image and Embedded Image (B), Embedded Image and Embedded Image (E), and Embedded Image and Embedded Image (H). C, F, I, Effect size densities. The black line represents the probability density of the regression coefficient for Embedded Image and Embedded Image (C), the regression coefficient for Embedded Image and Embedded Image (F), and the correlation coefficient for Embedded Image and Embedded Image (I). The green lines indicate the 95% HDIs. The red line shows the mode.

On a group level, the Kalman gain was a good approximation for the adaptation rate as the difference between the mean Embedded Image , and the mean Embedded Image normalized with respect to the mean Embedded Image was 10%; 95% HDI = [6.6 14]%. On an individual level, we found a positive correlation between steady-state Kalman gain Embedded Image and Embedded Image (r = 0.54; 95% HDI = [0.38 0.66]; see Fig. 5G–I), adding support to the claim that individual differences in adaptation rate can be explained from differences in noise according to an optimal learning rule. To assess the robustness of our findings, we performed a sensitivity analysis for the model priors (see Table 1: alternative priors 1–3) and a control analysis for within-subject correlations (see Table 1: permuted samples) and found consistent results.

View this table:
  • View inline
  • View popup
Table 1.

Sensitivity and control analyses.

Finally, we investigated how planning and execution noise correlated with movement peak velocity. Execution noise originates from muscle activity and should increase with vigorous contraction when larger motor units are recruited which fire at a lower frequency and produce more unfused twitches (Harris and Wolpert, 1998; Jones et al., 2002). Indeed, by regressing peak velocity onto the noise terms, we found a negligible correlation between peak velocity and planning noise β = –0.12; 95% HDI = [–0.27 0.02] and a small positive correlation between peak velocity and execution noise β = 0.22; 95% HDI = [0.18 0.28].

Discussion

We investigated the relation between components of motor noise and visuomotor adaptation rate across individuals. If adaptation approximates optimal learning from movement error, it can be predicted from Kalman filter theory that planning noise correlates positively and execution noise negatively with adaptation rate (Kalman, 1960). To test this hypothesis, we performed a visuomotor adaptation experiment in 69 subjects and extracted planning noise, execution noise, and adaptation rate using a state-space model of trial-to-trial behavior. Indeed, we found that adaptation rate correlates positively with planning noise (β = 0.44; 95% HDI = [0.27 0.59]) and negatively with execution noise (β = –0.39; 95% HDI = [–0.50 –0.30]). In addition, the steady-state Kalman gain calculated from planning and execution noise correlated positively with adaptation rate (r = 0.54; 95% HDI = [0.38 0.66]). We discuss implications of our findings for the optimal control model of movement and cerebellar models of adaptation and identify future applications of Bayesian state-space model fitting.

Optimal control model of movement

The optimal control model of movement has been successful in providing a unified explanation of motor control and motor learning (Todorov and Jordan, 2002). In this framework, the motor system sets a motor goal (possibly in the prefrontal cortex) and judges its value based on expected costs and rewards in the basal ganglia (Shadmehr and Krakauer, 2008). Selected movements are executed in a feedback control loop involving the motor cortex and the muscles which runs on an estimate of the system’s states (Shadmehr and Krakauer, 2008). Both the feedback controller and the state estimator are optimal in a mathematical sense, the feedback controller because it calculates optimal feedback parameters for minimizing motor costs and maximizing performance, given prescribed weighting of these two criteria (Åström and Murray, 2008), and the state estimator because it optimally combines sensory predictions from a forward model (cerebellum) with sensory feedback from the periphery (parietal cortex), similar to a Kalman filter (Kalman, 1960; Wolpert et al., 1995). In the optimal control model of movement, motor adaptation is defined as calibrating the forward model, which is optimal in the same sense as the state estimator (Shadmehr et al., 2010).

Wu et al. (2014) is one of the first studies to suggest that there may be a positive relationship between motor noise and motor adaptation. They outlined two apparent challenges of their findings to the optimal control approach: first, they claimed that optimal motor control is inconsistent with a positive relation between motor noise and adaptation rate; second, they claimed that optimal motor control does not account for the possibility that the motor system shapes motor noise to optimize adaptation. We take a different view. Because we find that only the planning component correlates positively with adaptation rate, our results are predicted by Kalman filter theory (Kalman, 1960) and consistent with optimal control models of movement (Todorov and Jordan, 2002; Åström and Murray, 2008). However, we do agree that the mathematical structure used to express the optimal control approach does not provide a clear way to discuss shaping noise to optimize adaptation. While this may be a technical difficulty from the point of view of optimal feedback approaches, it is apparent that there is electrophysiological evidence that some animals do shape noise to optimize adaptation. This evidence can be found in monkeys (Mandelblat-Cerf et al., 2009). In addition, studies in Bengalese finches show that a basal ganglia-premotor loop learns a melody from reward (Charlesworth et al., 2012) by injecting noise (Kao et al., 2005) to promote exploration (Tumer and Brainard, 2007) during training (Stepanek and Doupe, 2010) and development (Olveczky et al., 2005). We suggest that a similar mechanism operates in humans during adaptation. This additional tuning mechanism could be an interesting topic of future studies into optimal control models of movement.

Cerebellar model of motor adaptation

Motor adaptation is the learning process which fine tunes the forward model and is believed to take place in the olivocerebellar system (De Zeeuw et al., 2011). How could this learning process be sensitive to planning noise and execution noise on a neuronal level?

Central to the forward model is the cerebellar Purkinje cell, which responds to selected sensory (Chabrol et al., 2015) and motor (Kelly and Strick, 2003) parallel fiber input with a firing pattern reflecting kinematic properties of upcoming movements (Pasalar et al., 2006; Herzfeld et al., 2015). When Purkinje cell predictions of the upcoming kinematic properties are inaccurate, activity of neurons in the cerebellar nuclei is proportional to the prediction error. This is apparently because inhibitory Purkinje cell input cannot cancel the excitatory input from mossy fibers and the inferior olive (Brooks et al., 2015). The sensory prediction error calculated by the cerebellar nuclei could be used to update either (1) motor commands in a feedback loop with (pre)motor areas (Kelly and Strick, 2003) or (2) state estimates of the limb in the parietal cortex (Grafton et al., 1999; Clower et al., 2001). During adaptation, parallel fibers to Purkinje cell synapses associated with predictive signals are strengthened and parallel fibers to Purkinje cell synapses associated with nonpredictive signals are silenced (Dean et al., 2010). These plasticity mechanisms are affected by climbing fibers originating from the inferior olive, which integrate input from the sensorimotor system and the cerebellar nuclei and act as a teaching signal in the olivocerebellar system (De Zeeuw et al., 1998; Ohmae and Medina, 2015).

No previous experimental or modeling work has considered how planning or execution noise might be conveyed to the cerebellum or how they might influence plasticity. We speculate that planning noise is reflected in synaptic variability of the parallel fiber to Purkinje cell synapse. Electrophysiological studies of CA1 hippocampal neurons have shown that synaptic noise can improve detection of weak signals through stochastic resonance (Stacey and Durand, 2000). Such a mechanism might help form appropriate connections at the parallel fiber to Purkinje cell synapse during adaptation. In addition, theoretical studies on deep learning networks have shown that gradient descent algorithms, which can be likened to error-based learning, benefit from adding noise to the gradient at every training step (Neelakantan et al., 2015). Furthermore, we speculate that execution noise affects adaptation through climbing fiber firing modulation. Execution noise will decrease reliability of sensory prediction errors because (1) the motor plan is not executed faithfully (motor noise; van Beers et al., 2004) and (2) the sensory feedback is inaccurate (sensory noise; Osborne et al., 2005). Therefore, when sensory information for a specific movement plan has been unreliable in the past, the olivocerebellar system might decrease its response to sensory prediction error, for example by decreasing climbing fiber firing in the inferior olive (De Zeeuw et al., 1998), which would lower the adaptation rate. The existence of such a mechanism has also been suggested by a recent behavioral study that showed a specific decline in adaptation rate for movement perturbations that had been inconsistent in the past (Herzfeld et al., 2014).

Two-rate models of adaptation

Our results are based on a one-rate learning model of adaptation (Cheng and Sabes, 2006, 2007; van Beers, 2009). However, recent studies have suggested that a two-rate model composed of a slow but retentive and a fast but forgetting learning system provides a better explanation for learning phenomena such as savings and anterograde interference (Smith et al., 2006). The fast learning system might represent an explicit process, which could be located in the cortex, and the slow learning system an implicit process, which could be located in subcortical areas such as the cerebellum (Mazzoni and Krakauer, 2006; Taylor et al., 2014; McDougle et al., 2015). How could we interpret our results in light of these two-rate models? In a two-rate state-space model, the two systems will add to produce the movement output (Smith et al., 2006). That is, the total adaptation rate is equal to the sum of the adaptation rates of the two systems, and the same goes for the planning noise. Of course, a two-rate model will still include only one term for execution noise. Therefore, a two-rate model can reproduce our results either if both systems are optimally tuned or if only one system is optimally tuned but is relatively dominant. With our current experimental design, we cannot differentiate between these two options. Future studies combining reporting-based approaches to discern the contributions of the implicit and explicit processes and the Bayesian statistical approach to state-space modeling presented in this paper could further unravel this question.

Extended data 1

Bugs/Jags code for running the Bayesian state-space model of visuomotor adaptation. Download Extended Data 1, TXT file.

Footnotes

  • The authors declare no competing financial interests.

  • This work was supported by ZonMw (project # 10-10400-98-008) and Stichting Coolsingel.

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.

References

  1. ↵
    Andrieu C, de Freitas N, Doucet A, Jordan MI (2003) An introduction to MCMC for machine learning. Mach Learn 50:5–43. doi:10.1023/A:1020281327116
    OpenUrlCrossRef
  2. ↵
    Åström KJ, Murray RM (2008) Feedback Systems: An Introduction for Scientists and Engineers. Princeton, NJ: Princeton University Press.
  3. ↵
    Baddeley RJ, Ingram HA, Miall RC (2003) System identification applied to a visuomotor task: near-optimal human performance in a noisy changing task. J Neurosci 23:3066–3075. doi:10.1523/JNEUROSCI.23-07-03066.2003
    OpenUrlAbstract/FREE Full Text
  4. ↵
    Bialek W (1987) Physical limits to sensation and perception. Annu Rev Biophys Biophys Chem 16:455–478. doi:10.1146/annurev.bb.16.060187.002323 pmid:3297091
    OpenUrlCrossRefPubMed
  5. ↵
    Brooks JX, Carriot J, Cullen KE (2015) Learning to expect the unexpected: rapid updating in primate cerebellum during voluntary self-motion. Nat Neurosci 18:1310–1317. doi:10.1038/nn.4077 pmid:26237366
    OpenUrlCrossRefPubMed
  6. ↵
    Browne WJ, Draper D (2006) A comparison of Bayesian and likelihood-based methods for fitting multilevel models. Bayesian Anal 1:473–514. doi:10.1214/06-BA117
    OpenUrlCrossRef
  7. ↵
    Burge J, Ernst MO, Banks MS (2008) The statistical determinants of adaptation rate in human reaching. J Vis 8:20. doi:10.1167/8.4.20
    OpenUrlAbstract/FREE Full Text
  8. ↵
    Carlin BP, Polson NG, Stoffer DS (1992) A Monte Carlo approach to nonnormal and nonlinear state-space modeling. J Am Stat Assoc 87:493. doi:10.1080/01621459.1992.10475231
    OpenUrlCrossRef
  9. ↵
    Carter CK, Kohn R (1994) On Gibbs sampling for state space models. Biometrika 81:541–553. doi:10.1093/biomet/81.3.541
    OpenUrlCrossRef
  10. ↵
    Chabrol FP, Arenz A, Wiechert MT, Margrie TW, DiGregorio DA (2015) Synaptic diversity enables temporal coding of coincident multisensory inputs in single neurons. Nat Neurosci 18:718–727. doi:10.1038/nn.3974 pmid:25821914
    OpenUrlCrossRefPubMed
  11. ↵
    Chaisanguanthum KS, Shen HH, Sabes PN (2014) Motor variability arises from a slow random walk in neural state. J Neurosci 34:12071–12080. doi:10.1523/JNEUROSCI.3001-13.2014 pmid:25186752
    OpenUrlAbstract/FREE Full Text
  12. ↵
    Charlesworth JD, Warren TL, Brainard MS (2012) Covert skill learning in a cortical-basal ganglia circuit. Nature 486:251–255. doi:10.1038/nature11078 pmid:22699618
    OpenUrlCrossRefPubMed
  13. ↵
    Cheng S, Sabes PN (2007) Calibration of visually guided reaching is driven by error-corrective learning and internal dynamics. J Neurophysiol 97:3057–3069. doi:10.1152/jn.00897.2006 pmid:17202230
    OpenUrlCrossRefPubMed
  14. ↵
    Cheng S, Sabes PN (2006) Modeling sensorimotor learning with linear dynamical systems. Neural Comput 18:760–793. doi:10.1162/089976606775774651 pmid:16494690
    OpenUrlCrossRefPubMed
  15. ↵
    Churchland MM, Afshar A, Shenoy KV (2006) A central source of movement variability. Neuron 52:1085–1096. doi:10.1016/j.neuron.2006.10.034 pmid:17178410
    OpenUrlCrossRefPubMed
  16. ↵
    Clower DM, West RA, Lynch JC, Strick PL (2001) The inferior parietal lobule is the target of output from the superior colliculus, hippocampus, and cerebellum. J Neurosci 21:6283–6291. pmid:11487651
    OpenUrlAbstract/FREE Full Text
  17. ↵
    Cumming G (2014) The new statistics. Psychol Sci 25:7–29. doi:10.1177/0956797613504966 pmid:24220629
    OpenUrlCrossRefPubMed
  18. ↵
    De Zeeuw CI, Hoebeek FE, Bosman LWJ, Schonewille M, Witter L, Koekkoek SK (2011) Spatiotemporal firing patterns in the cerebellum. Nat Rev Neurosci 12:327–344. doi:10.1038/nrn3011 pmid:21544091
    OpenUrlCrossRefPubMed
  19. ↵
    De Zeeuw CI, Simpson JI, Hoogenraad CC, Galjart N, Koekkoek SK, Ruigrok TJ (1998) Microcircuitry and function of the inferior olive. Trends Neurosci 21:391–400. pmid:9735947
    OpenUrlCrossRefPubMed
  20. ↵
    Dean P, Porrill J, Ekerot C-F, Jörntell H (2010) The cerebellar microcircuit as an adaptive filter: experimental and computational evidence. Nat Rev Neurosci 11:30–43. doi:10.1038/nrn2756 pmid:19997115
    OpenUrlCrossRefPubMed
  21. ↵
    Donchin O, Rabe K, Diedrichsen J, Lally N, Schoch B, Gizewski ER, Timmann D (2012) Cerebellar regions involved in adaptation to force field and visuomotor perturbation. J Neurophysiol 107:134–147. doi:10.1152/jn.00007.2011 pmid:21975446
    OpenUrlCrossRefPubMed
  22. ↵
    Faisal AA, Selen LPJ, Wolpert DM (2008) Noise in the nervous system. Nat Rev Neurosci 9:292–303. doi:10.1038/nrn2258 pmid:18319728
    OpenUrlCrossRefPubMed
  23. ↵
    Frank MJ, Doll BB, Oas-Terpstra J, Moreno F (2009) Prefrontal and striatal dopaminergic genes predict individual differences in exploration and exploitation. Nat Neurosci 12:1062–1068. doi:10.1038/nn.2342 pmid:19620978
    OpenUrlCrossRefPubMed
  24. ↵
    Fritsch B, Reis J, Martinowich K, Schambra HM, Ji Y, Cohen LG, Lu B (2010) Direct current stimulation promotes BDNF-dependent synaptic plasticity: potential implications for motor learning. Neuron 66:198–204. doi:10.1016/j.neuron.2010.03.035 pmid:20434997
    OpenUrlCrossRefPubMed
  25. ↵
    Gelman A (2006) Prior distributions for variance parameters in hierarchical models (comment on article by Browne and Draper). Bayesian Anal 1:515–534. doi:10.1214/06-BA117A
    OpenUrlCrossRef
  26. ↵
    Gonzalez Castro LN, Hadjiosif AM, Hemphill MA, Smith MA (2014) Environmental consistency determines the rate of motor adaptation. Curr Biol 24:1050–1061. doi:10.1016/j.cub.2014.03.049 pmid:24794296
    OpenUrlCrossRefPubMed
  27. ↵
    Grafton ST, Desmurget M, Epstein CM, Turner RS, Prablanc C, Alexander GE (1999) Role of the posterior parietal cortex in updating reaching movements to a visual target. Nat Neurosci 2:563–567. doi:10.1038/9219 pmid:10448222
    OpenUrlCrossRefPubMed
  28. ↵
    Haar S, Donchin O, Dinstein I (2017) Individual movement variability magnitudes are explained by cortical neural variability. J Neurosci 37:9076–9085. doi:10.1523/JNEUROSCI.1650-17.2017 pmid:28821678
    OpenUrlAbstract/FREE Full Text
  29. ↵
    Harris CM, Wolpert DM (1998) Signal-dependent noise determines motor planning. Nature 394:780–784. doi:10.1038/29528 pmid:9723616
    OpenUrlCrossRefPubMed
  30. ↵
    He K, et al. (2016) The statistical determinants of the speed of motor learning. PLOS Comput Biol 12:e1005023. doi:10.1371/journal.pcbi.1005023 pmid:27606808
    OpenUrlCrossRefPubMed
  31. ↵
    Herzfeld DJ, Kojima Y, Soetedjo R, Shadmehr R (2015) Encoding of action by the Purkinje cells of the cerebellum. Nature 526:439–442. doi:10.1038/nature15693 pmid:26469054
    OpenUrlCrossRefPubMed
  32. ↵
    Herzfeld DJ, Vaswani PA, Marko MK, Shadmehr R (2014) A memory of errors in sensorimotor learning. Science 345:1349–1353. doi:10.1126/science.1253138
    OpenUrlAbstract/FREE Full Text
  33. ↵
    Ioannidis JPA (2005) Why most published research findings are false. PLoS Med 2:e124. doi:10.1371/journal.pmed.0020124 pmid:16060722
    OpenUrlCrossRefPubMed
  34. ↵
    Jones KE, Hamilton AF, Wolpert DM (2002) Sources of signal-dependent noise during isometric force production. J Neurophysiol 88:1533–1544. doi:10.1152/jn.2002.88.3.1533 pmid:12205173
    OpenUrlCrossRefPubMed
  35. ↵
    Kalman RE (1960) A new approach to linear filtering and prediction problems. J Basic Eng 82:35. doi:10.1115/1.3662552
    OpenUrlCrossRef
  36. ↵
    Kao MH, Doupe AJ, Brainard MS (2005) Contributions of an avian basal ganglia-forebrain circuit to real-time modulation of song. Nature 433:638–43. doi:10.1038/nature03127 pmid:15703748
    OpenUrlCrossRefPubMed
  37. ↵
    Kelly RM, Strick PL (2003) Cerebellar loops with motor cortex and prefrontal cortex of a nonhuman primate. J Neurosci 23:8432–8444. doi:10.1523/JNEUROSCI.23-23-08432.2003
    OpenUrlAbstract/FREE Full Text
  38. ↵
    Kruschke JK (2013) Bayesian estimation supersedes the t test. J Exp Psychol Gen 142:573–603. doi:10.1037/a0029146
    OpenUrlCrossRefPubMed
  39. ↵
    Kruschke JK, Liddell TM (2018a) The Bayesian new statistics: hypothesis testing, estimation, meta-analysis, and power analysis from a Bayesian perspective. Psychon Bull Rev 25:178–206. doi:10.3758/s13423-016-1221-4
    OpenUrlCrossRef
  40. ↵
    Kruschke JK, Liddell TM (2018b) Bayesian data analysis for newcomers. Psychon Bull Rev 25:155–177. doi:10.3758/s13423-017-1272-1
    OpenUrlCrossRef
  41. ↵
    Kruschke JK (2010) Doing Bayesian data analysis: a tutorial with R, JAGS, and Stan. Cambridge, MA: Academic Press.
  42. ↵
    Mandelblat-Cerf Y, Paz R, Vaadia E (2009) Trial-to-trial variability of single cells in motor cortices is dynamically modified during visuomotor adaptation. J Neurosci 29:15053–15062. doi:10.1523/JNEUROSCI.3011-09.2009 pmid:19955356
    OpenUrlAbstract/FREE Full Text
  43. ↵
    Mazzoni P, Krakauer JW (2006) An implicit plan overrides an explicit strategy during visuomotor adaptation. J Neurosci 26:3642–3645. doi:10.1523/JNEUROSCI.5317-05.2006
    OpenUrlAbstract/FREE Full Text
  44. ↵
    McDougle SD, Bond KM, Taylor JA (2015) Explicit and implicit processes constitute the fast and slow processes of sensorimotor learning. J Neurosci 35:9568–9579. doi:10.1523/JNEUROSCI.5061-14.2015 pmid:26134640
    OpenUrlAbstract/FREE Full Text
  45. ↵
    McHughen SA, Rodriguez PF, Kleim JA, Kleim ED, Marchal Crespo L, Procaccio V, Cramer SC (2010) BDNF val66met polymorphism influences motor system function in the human brain. Cereb Cortex 20:1254–1262. doi:10.1093/cercor/bhp189 pmid:19745020
    OpenUrlCrossRefPubMed
  46. ↵
    Neelakantan A, Vilnis L, Le QV, Sutskever I, Kaiser L, Kurach K, Martens J (2015) Adding gradient noise improves learning for very deep networks. Mach Learning arXiv:1511.06807.
  47. ↵
    Ohmae S, Medina JF (2015) Climbing fibers encode a temporal-difference prediction error during cerebellar learning in mice. Nat Neurosci 18:1798–1803. doi:10.1038/nn.4167 pmid:26551541
    OpenUrlCrossRefPubMed
  48. ↵
    Olveczky BP, Andalman AS, Fee MS (2005) Vocal experimentation in the juvenile songbird requires a basal ganglia circuit. PLoS Biol 3:e153. doi:10.1371/journal.pbio.0030153 pmid:15826219
    OpenUrlCrossRefPubMed
  49. ↵
    Osborne LC, Lisberger SG, Bialek W (2005) A sensory source for motor variation. Nature 437:412–416. doi:10.1038/nature03961 pmid:16163357
    OpenUrlCrossRefPubMed
  50. ↵
    Pasalar S, Roitman AV, Durfee WK, Ebner TJ (2006) Force field effects on cerebellar Purkinje cell discharge with implications for internal models. Nat Neurosci 9:1404–1411. doi:10.1038/nn1783 pmid:17028585
    OpenUrlCrossRefPubMed
  51. ↵
    Shadmehr R, Krakauer JW (2008) A computational neuroanatomy for motor control. Exp Brain Res 185:359–381. doi:10.1007/s00221-008-1280-5 pmid:18251019
    OpenUrlCrossRefPubMed
  52. ↵
    Shadmehr R, Smith MA, Krakauer JW (2010) Error correction, sensory prediction, and adaptation in motor control. Annu Rev Neurosci 33:89–108. doi:10.1146/annurev-neuro-060909-153135 pmid:20367317
    OpenUrlCrossRefPubMed
  53. ↵
    Smith MA, Ghazizadeh A, Shadmehr R (2006) Interacting adaptive processes with different timescales underlie short-term motor learning. PLoS Biol 4:e179. doi:10.1371/journal.pbio.0040179 pmid:16700627
    OpenUrlCrossRefPubMed
  54. ↵
    Stacey WC, Durand DM (2000) Stochastic resonance improves signal detection in hippocampal CA1 neurons. J Neurophysiol 83:1394–1402. doi:10.1152/jn.2000.83.3.1394 pmid:10712466
    OpenUrlCrossRefPubMed
  55. ↵
    Stepanek L, Doupe AJ (2010) Activity in a cortical-basal ganglia circuit for song is required for social context-dependent vocal variability. J Neurophysiol 104:2474–2486. doi:10.1152/jn.00977.2009
    OpenUrlCrossRefPubMed
  56. ↵
    Taylor JA, Krakauer JW, Ivry RB (2014) Explicit and implicit contributions to learning in a sensorimotor adaptation task. J Neurosci 34:3023–3032. doi:10.1523/JNEUROSCI.3619-13.2014 pmid:24553942
    OpenUrlAbstract/FREE Full Text
  57. ↵
    Todorov E, Jordan MI (2002) Optimal feedback control as a theory of motor coordination. Nat Neurosci 5:1226–1235. doi:10.1038/nn963 pmid:12404008
    OpenUrlCrossRefPubMed
  58. ↵
    Tomassini V, Jbabdi S, Kincses ZT, Bosnell R, Douaud G, Pozzilli C, Matthews PM, Johansen-Berg H (2011) Structural and functional bases for individual differences in motor learning. Hum Brain Mapp 32:494–508. doi:10.1002/hbm.21037 pmid:20533562
    OpenUrlCrossRefPubMed
  59. ↵
    Tseng Y-WW, Diedrichsen J, Krakauer JW, Shadmehr R, Bastian AJ (2007) Sensory prediction errors drive cerebellum-dependent adaptation of reaching. J Neurophysiol 98:54–62. doi:10.1152/jn.00266.2007
    OpenUrlCrossRefPubMed
  60. ↵
    Tumer EC, Brainard MS (2007) Performance variability enables adaptive plasticity of “crystallized” adult birdsong. Nature 450:1240–1244. doi:10.1038/nature06390 pmid:18097411
    OpenUrlCrossRefPubMed
  61. ↵
    van Beers RJ (2009) Motor learning is optimally tuned to the properties of motor noise. Neuron 63:406–417. doi:10.1016/j.neuron.2009.06.025 pmid:19679079
    OpenUrlCrossRefPubMed
  62. ↵
    van Beers RJ, Haggard P, Wolpert DM (2004) The role of execution noise in movement variability. J Neurophysiol 91:1050–1063. doi:10.1152/jn.00652.2003 pmid:14561687
    OpenUrlCrossRefPubMed
  63. ↵
    van Beers RJ (2012) How does our motor system determine its learning rate? PLoS One 7:e49373. doi:10.1371/journal.pone.0049373 pmid:23152899
    OpenUrlCrossRefPubMed
  64. ↵
    Wagenmakers E-J, Verhagen J, Ly A, Bakker M, Lee MD, Matzke D, Rouder JN, Morey RD (2014) A power fallacy. Behav Res Methods 47:913–917.
    OpenUrl
  65. ↵
    Wei K, Körding K (2010) Uncertainty of feedback and state estimation determines the speed of motor adaptation. Front Comput Neurosci 4:11. doi:10.3389/fncom.2010.00011 pmid:20485466
    OpenUrlCrossRefPubMed
  66. ↵
    Wolpert DM, Ghahramani Z, Jordan MI (1995) An internal model for sensorimotor integration. Science 269:1880–1882. pmid:7569931
    OpenUrlAbstract/FREE Full Text
  67. ↵
    Wu HG, Miyamoto YR, Gonzalez Castro LN, Ölveczky BP, Smith MA (2014) Temporal structure of motor variability is dynamically regulated and predicts motor learning ability. Nat Neurosci 17:312–321. doi:10.1038/nn.3616 pmid:24413700
    OpenUrlCrossRefPubMed

Synthesis

Reviewing Editor: Trevor Drew, University of Montreal

Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below. The following reviewer(s) agreed to reveal their identity: Reza Shadmehr. Note: If this manuscript was transferred from JNeurosci and a decision was made to accept the manuscript without peer review, a brief statement to this effect will instead be what is listed below.

The authors have addressed all of my previous concerns. I only have some minor comments that they may wish to look at. This is a wonderful paper.

On lines 327 and 329, aren't you reporting r here, not β?

Typo: line 292, should say Figure 4F, not 3F

Typo: line 300, should say Figure 4H, 3H

Typo: line 176: probable, not probably

Back to top

In this issue

eneuro: 5 (4)
eNeuro
Vol. 5, Issue 4
July/August 2018
  • Table of Contents
  • Index by author
Email

Thank you for sharing this eNeuro article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Individual Differences in Motor Noise and Adaptation Rate Are Optimally Related
(Your Name) has forwarded a page to you from eNeuro
(Your Name) thought you would be interested in this article in eNeuro.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Individual Differences in Motor Noise and Adaptation Rate Are Optimally Related
Rick van der Vliet, Maarten A. Frens, Linda de Vreede, Zeb D. Jonker, Gerard M. Ribbers, Ruud W. Selles, Jos N. van der Geest, Opher Donchin
eNeuro 16 July 2018, 5 (4) ENEURO.0170-18.2018; DOI: 10.1523/ENEURO.0170-18.2018

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Share
Individual Differences in Motor Noise and Adaptation Rate Are Optimally Related
Rick van der Vliet, Maarten A. Frens, Linda de Vreede, Zeb D. Jonker, Gerard M. Ribbers, Ruud W. Selles, Jos N. van der Geest, Opher Donchin
eNeuro 16 July 2018, 5 (4) ENEURO.0170-18.2018; DOI: 10.1523/ENEURO.0170-18.2018
Reddit logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Significance Statement
    • Introduction
    • Materials and methods
    • Results
    • Discussion
    • Footnotes
    • References
    • Synthesis
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • Bayesian statistics
  • cerebellum
  • motor learning
  • Noise
  • Optimal control
  • Visuomotor adaptation

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

New Research

  • Deciding while acting - Mid-movement decisions are more strongly affected by action probability than reward amount
  • CaMKIIα promoter-controlled circuit manipulations target both pyramidal cells and inhibitory interneurons in cortical networks
  • Gas7 is a novel dendritic spine initiation factor
Show more New Research

Sensory and Motor Systems

  • Different control strategies drive interlimb differences in performance and adaptation during reaching movements in novel dynamics
  • The nasal solitary chemosensory cell signaling pathway triggers mouse avoidance behavior to inhaled nebulized irritants
  • Taste-odor association learning alters the dynamics of intra-oral odor responses in the posterior piriform cortex of awake rats
Show more Sensory and Motor Systems

Subjects

  • Sensory and Motor Systems

  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Latest Articles
  • Issue Archive
  • Blog
  • Browse by Topic

Information

  • For Authors
  • For the Media

About

  • About the Journal
  • Editorial Board
  • Privacy Policy
  • Contact
  • Feedback
(eNeuro logo)
(SfN logo)

Copyright © 2023 by the Society for Neuroscience.
eNeuro eISSN: 2373-2822

The ideas and opinions expressed in eNeuro do not necessarily reflect those of SfN or the eNeuro Editorial Board. Publication of an advertisement or other product mention in eNeuro should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in eNeuro.