Abstract
Although animal research implicates a central role for dopamine in motor skill learning, a direct causal link has yet to be established in neurotypical humans. Here, we tested if a pharmacological manipulation of dopamine alters motor learning, using a paradigm which engaged explicit, goal-directed strategies. Participants (27 females; 11 males; aged 18–29 years) first consumed either 100 mg of levodopa (n = 19), a dopamine precursor that increases dopamine availability, or placebo (n = 19). Then, during training, participants learnt the explicit strategy of aiming away from presented targets by instructed angles of varying sizes. Targets jumped mid-movement by the instructed aiming angle. Task success was thus contingent upon aiming accuracy and not speed. The effect of the dopamine manipulations on skill learning was assessed during training and after an overnight follow-up. Increasing dopamine availability at training improved aiming accuracy and lengthened reaction times, particularly for larger, more difficult aiming angles, both at training and, importantly, at follow-up, despite prominent session-by-session performance improvements in both accuracy and speed. Exogenous dopamine thus seems to result in a learnt, persistent propensity to better adhere to task goals. Results support the proposal that dopamine is important in engagement of instrumental motivation to optimize adherence to task goals, particularly when learning to execute goal-directed strategies in motor skill learning.
Significance Statement
While animal studies show a central role of dopamine in skill learning, such evidence is lacking in neurotypical humans. We provide evidence for a role of dopamine in learning explicit, goal-directed motor strategies in neurotypical humans. Exogenous dopamine at training improved accuracy, as participants traded speed for accuracy in a task that focused on accuracy, not speed. Importantly, this behavior persisted at a no-drug follow-up, suggesting that dopamine resulted in a learnt, persistent propensity to better adhere to task goals. Dopamine influences instrumental motivation to optimize adherence to task goals in motor learning, not only influencing performance at initial learning but also when retrieving such strategies to solve familiar motor problems.
Introduction
Motor skill learning is essential for survival: a hungry bear catching salmon must adapt its paw-strikes to dynamic task parameters, such as the movement and friction of the salmon's body and the forces applied by river waters. Reward is a potent modulator of skilled movement, affecting movement in two primary ways. First, rewards can increase the vigor of movements (Manohar et al., 2015; Carroll et al., 2019; Codol et al., 2019, for a review, see Shadmehr et al. (2019)) even when rewards are not performance contingent (Takikawa et al., 2002). Second, rewards can boost skill learning. For example, rewarding participants for achieving performance criteria can speed up learning (Vassiliadis et al., 2021) or improve retention of motor learning (Abe et al., 2011; Madelain et al., 2011; Galea et al., 2015). The potency of reward in altering behavior and learning is perhaps unsurprising, as animals must learn motor skills to attain resources for survival (Barron et al., 2010).
Experimental manipulations of reward do not, however, always benefit motor learning (Steel et al., 2016; Spampinato et al., 2019). This is perhaps because rewarding participants for achieving some experimenter-determined performance criterion does not directly alter a key learning signal: success or failure at achieving the task goal. Failures to achieve task goals, termed task errors, seem essential to forming memories that improve future motor performance, and such memories take at least two forms: deliberative goal-oriented strategies and automatic stimulus–response associations (Leow et al., 2020).
Task errors can be conceptualized as a form of reward prediction error (Leow et al., 2020), which is a broader term describing discrepancies between predicted rewards and received rewards (Sutton and Barto, 1998). Transient bursts of firing by midbrain dopamine neurons triggered by reward prediction errors have been thought to “stamp-in” associations between a stimulus and its associated response, resulting in the formation of automatic stimulus–response mappings. Increasing evidence also implicates dopamine in the engagement of deliberative, goal-oriented behaviors (Akam and Walton, 2021). Dopamine might therefore play dual roles in skill learning: stamping in stimulus–response associations to facilitate automaticity with training and motivating the engagement of goal-oriented strategies.
While nonhuman animal studies implicate dopamine-dependent circuits in skill learning (Hosp et al., 2011), inter-species differences in the cognitive and neural processes associated with the dopamine system (Khan et al., 1998) limit generalization of findings from animals to humans. Evidence for the role of dopamine in human motor learning comes from studying how dopamine medications alter learning in patients with impaired dopamine function (Kwak et al., 2010), but heterogeneity in disease phenotypes complicates inferences from such work (Cools et al., 2022). Causal evidence for the role of dopamine in skill learning in neurotypical humans is scarce and limited to studies of elementary motor tasks such as repeating simple thumb-abduction movements (Floel et al., 2008; Flöel et al., 2008; Rosser et al., 2008) or tracking on-screen targets (Chen et al., 2020a). Of import, some studies have found no effect of manipulating dopamine on motor learning (Quattrocchi et al., 2018; Palidis et al., 2021). These null results might have resulted from paradigms which predominantly engage implicit learning (Palidis et al., 2021), or which do not dissociate effects of explicit and implicit learning processes (Quattrocchi et al., 2018), which can have mutually compensatory effects on behavior (Albert et al., 2022). Indeed, in reinforcement learning tasks, explicit, goal-directed processes can be more sensitive to effects of dopamine manipulations than implicit, automatic processes (Sharp et al., 2016). Similarly, explicit processes are more sensitive to reward manipulations than implicit processes in motor learning (Codol et al., 2018; Holland et al., 2018). Manipulations of dopamine and reward might thus be more observable in learning tasks driven by explicit knowledge of the task structure and strategies to achieve task goals.
Here, we investigated the role of dopamine in learning goal-directed motor strategies. After consuming the dopamine precursor levodopa or placebo, participants learnt a strategy of aiming away from presented targets by angles of varying sizes (Georgopoulos and Massey, 1987). Participants had to employ strategic aiming to hit targets, as targets would jump mid-movement by the instructed aiming angle. Thus, task success (i.e., target-hitting) was contingent upon successful aiming. Effects of the dopamine manipulations were assessed during training and at a no-drug follow-up session.
Materials and Methods
Participants
Thirty-eight neurotypical young adults (27 females; median age, 21.58 years; range, 18–29 years) were recruited from The University of Queensland community and were reimbursed for participation (AUD$20/h). Participants were screened for neurological and psychiatric conditions and contraindications for levodopa and provided written informed consent. In accordance with the National Health and Medical Research Council's guidelines, this experiment was approved by the human ethics committee at The University of Queensland. No datasets were excluded from the study.
As no published studies on the effects of levodopa on explicit motor skill learning in neurotypical adults existed at the time of planning the study, our sample size was guided by previous studies that examined effects of levodopa on tasks requiring cognitive effort (Vo et al., 2017); however, we did not explicitly conduct a power analysis based on these studies, as their design was dissimilar to ours. Half of the participants (n = 19) were randomly assigned to the levodopa condition (16 females; median age, 20.74 years; range, 18–25 years), and the other half were assigned to the placebo condition (n = 19; 11 females; median age, 22.42 years; range, 19–29 years).
Drug manipulation
The levodopa group consumed the dopamine precursor, a Madopar 125 (levodopa 100 mg/benserazide 25 mg) tablet, whereas the placebo group consumed a crushed multivitamin (Centrum for women). Double blinding was achieved by having an experimenter who was not involved in data collection crush and disperse the tablets in orange juice before consumption by the participant. Behavioral testing of the aiming task began ∼60 min after tablet administration, around the time of peak plasma concentration (Contin and Martinelli, 2010). The drug manipulation was only employed on the training session, and not at follow-up.
Apparatus
A vBot planar robotic manipulandum was used (Howard et al., 2009). This apparatus has a low-mass, two-link carbon fiber arm and measures position with optical encoders sampled at 1,000 Hz. Participants completed the aiming task while seated on a height-adjustable chair at their ideal height for viewing the screen. Visual feedback was presented on a horizontal plane on a 27″ LCD computer monitor (ASUS, VG278H, set at 60 Hz refresh rate) mounted above the vBot and projected to the subject via a mirror in a darkened room. Participants had no direct vision of their hand. Hand position was represented by a red cursor (0.25-cm-radius circle) on a black background. The mirror also allowed visual feedback of the target (a 0.5-cm-radius circle) and the start location (a 0.35-cm-radius circle), which was aligned 10 cm to the right of the participant's midsagittal plane at approximately mid-sternum level.
Procedure
Participants were first given the following instructions: “You are about to perform sets of 80 trials in which you will be asked to move at specific angle relative to the target in clockwise direction. The angle will differ between blocks. We will tell you what the angle is for each block. First, make up your mind and then try to keep your movement straight and fast. Try your best to execute your movements as accurately as possible and hit as many targets as possible. Between each set we will ask you to aim towards the target in a 0° angle, that means you should hit the target directly. These sets will contain 20 trials.” Instructions were followed by a demonstration by the experimenter on how to re-aim by 90° away from the presented target.
Participants were given seven aiming blocks. In each block, participants were asked to re-aim away from the presented target by one of the following angles, 0°, 30°, 60°, 120°, 135°, 150°, and 165°, similar to previous work (Georgopoulos and Massey, 1987; Pellizzer and Georgopoulos, 1993; Bhat and Sanes, 1998; Neely and Heath, 2010; McNeely and Earhart, 2012), where block order was randomized between subjects. Each block contained 10 bins of eight trials each, making a total of 80 trials per block. Each bin contained one presentation of all eight possible target locations (0°, 45°, 90°… 315°), in random order within each bin. All trials commenced by displaying the central start circle, with participants then moving the cursor within 1 cm of the start circle. If participants did not move the cursor to the start after 1 s, the robotic manipulandum moved the participant's cursor to the start circle, using a simulated two-dimensional spring with the spring constant magnitude linearly increasing over time. After the cursor remained within the start circle at a speed below 0.1 cm/s for 200 ms, targets appeared in random order at one of eight locations 9 cm away from the start circle (0°, 45°, 90°… 315°; Fig. 1A).
In trials contained within each aiming block, the target “jumped” mid-movement (i.e., when movement extent reached 4 cm into the 9 cm start-target distance): the size of the target-jump was the same as the instructed aiming angle, such that successful aiming was necessary to attain the task goal of hitting the target (Leow et al., 2020). Target “jumps” were achieved by extinguishing the target and re-displaying the target at the new location as soon as movement extent exceeded 4 cm out of the 9 cm start-target distance (Fig. 1B). Each aiming block was followed by a 20-trial washout block, where the task was to move straight to the presented targets: no target-jumps occurred during these washout blocks. Across all trials, cursor feedback was only given after movement extent exceeded 4 cm and was then extinguished after movement extent exceeded the 9 cm start-target distance. We have previously found this effective in disincentivizing online movement corrections (Leow et al., 2018, 2020, 2021). To economize study duration, a washout block was not provided after the 0° block or the final re-aiming block. Across all trials, including the washout blocks, auditory feedback in the form of a beep sound (coin.wav from Super Mario) was given when the reach angle measured at 4 cm into the 9 cm start-target distance reached an accuracy criterion of being within a +/10° range of the ideal aiming angle. This auditory feedback was implemented to maintain participant engagement in the long test session.
After an overnight delay, participants returned for a follow-up session, scheduled a minimum of 18 h after the first session. The task was identical to the first session. On the second session, all participants consumed a placebo pill dispersed in orange juice. Participant blinding was assessed after both sessions.
Additional assessments
To assess for changes in arousal, heart rate, blood pressure, and mood, we measured these variables at the beginning, after 1 h, and at the end of each session (Vo et al., 2016). Mood was assessed using the Mood Rating Scale (Bond and Lader, 1974) which includes 16 items separated in the following factors: alertness, contentedness, and calmness. The participants could rate each element within a range of 10 points. The three factors were evaluated as a total score.
Individual differences in baseline dopamine function affects responsivity to dopamine medications. Individual dopamine baseline function is partly predicted by impulsivity and working memory. Impulsivity, as measured by the behavioral inhibition scale (BIS-11), is associated with dopamine D2/D3 receptor availability in the midbrain (Buckholtz et al., 2010). BIS scores can predict effects of methylphenidate on learning (Clatworthy et al., 2009). Similarly, working memory span has been associated with dopamine synthesis capacity, with a medium to large effect size for the correlation between listening span scores and dopamine synthesis capacity (Cools et al., 2008). Working memory span can predict effects of pharmacological manipulations of dopamine (Broadway et al., 2018; Fallon et al., 2019). To account for individual differences in baseline dopamine function, we used scores on assessments of impulsivity and working memory capacity as covariates. Working memory was measured with the memory updating task from a validated working memory battery (Lewandowsky et al., 2010). The memory updating task had participants remember a set of digits presented for 1 s in separate on-screen locations and update these digits through arithmetic operations shown on-screen for 1.3 s at corresponding locations. The number of digits presented (i.e., set size) varied from three to five across trials. After a varying number of update operations (between two and six), question mark prompts appeared in each frame. There was no time limit for the recall, and there was no performance feedback. The operations ranged from +7 to −7, excluding 0, and the results from 1 to 9. There were 15 trials in total and two practice trials. Performance on the memory updating task was assessed both at training and at follow-up, but to avoid the influence of practice effects, we used memory updating data from only the training session for analysis.
Data analysis
Our task instructions to “try your best to execute your movements as accurately as possible” and target-jump manipulations (i.e., where to target shifted mid-movement at the required aiming angle, such that target-hitting required accurate aiming) emphasized accuracy of aiming movements, rather than the speed of reaction times nor the speed of the executed movements. Thus, our primary dependent variable of interest was accuracy of aiming at the instructed angle. Aiming accuracy was estimated from the directional error between the ideal aiming angle and the absolute values of the actual reach direction, where reach direction was measured at 20% of the start-target distance (i.e., 1.8 cm into the 9 cm start-target distance).
While our instructions and task structure emphasized aiming accuracy, we were also interested in evaluating changes in the speed of movement planning and movement execution. To this end, we quantified the following: (1) reaction time, defined as the interval from target appearance to movement onset, where movement onset was the time at which the hand speed first exceeded 2 cm/s, and (2) peak velocity. Furthermore, as previous research has implicated reward in movement precision (Manohar et al., 2015), we also estimated precision of aiming by quantifying error variability, defined as the standard deviation of errors calculated for every bin.
To quantify learning and to test whether the drug manipulation altered learning, we measured how performance changed across multiple time scales. Rapid trial-by-trial performance changes were quantified via Session (training, follow-up) × Angle (0°, 30°…165°) × Trial (1…8) × Drug (levodopa, placebo) ANOVAs run on the first eight trials of each aiming block. (2) Bin-by-bin and session-by-session performance changes were quantified via Session (training, follow-up) × Bin (1…10) × Angle (0°, 30°…165°) × Drug (levodopa, placebo) ANOVAs run on bin-averaged data (1 bin = mean of eight trials). Across all analyses, we entered three covariates of no interest, as follows: (1) working memory capacity (estimated via memory updating scores), (2) impulsivity (estimated via BIS total scores), and (3) biological sex. These covariates were considered because working memory (Cools et al., 2008), impulsivity (Cools et al., 2007), and sex (Becker, 1999) can affect responsivity to dopamine drug manipulations. Two participants from the levodopa condition did not complete the 165° block due to experimenter error: we included all other data from these participants in all analyses.
We made inferences using Bayesian statistics, as this allowed us to quantify evidence for the test hypothesis, as well as evidence for the null hypothesis. Inclusion Bayes factors (BFinclusion, BFincl) were determined to estimate the strength of evidence in favor of including an effect relative to models stripped of that effect. Exclusion Bayes factors (BFexclusion, BFexcl) were determined to estimate the strength of evidence in favor of excluding an effect relative to models stripped of that effect. In post hoc tests, where evidence for the null hypothesis is reported as BF01, and evidence for the test hypothesis is reported as BF10, the posterior odds were corrected for multiple testing by fixing to 0.5 the prior probability that the null hypothesis holds across all comparisons (Westfall et al., 1997). Jeffreys's evidence categories for interpretation (Wetzels et al., 2011) were taken as the standard for evaluation of the reported Bayes factors. Specifically, Bayes factors of 1–3 were interpreted as anecdotal, 3–10 as moderate, and >10 as strong evidence for the test hypothesis, whereas Bayes factors of 0.33 or less was taken as evidence for the null hypothesis. Where appropriate, Cohen's d values were used to quantify effect sizes and were reported with 95% confidence intervals of the effect size, as estimated using the webapp estimation statistics (Ho et al., 2019). Bayesian analyses were conducted using JAMOVI version 2.3.28. As JAMOVI does not provide covariate-adjusted estimated marginal means, these were obtained using frequentist ANOVAs using SPSS (version 28.0.1.0), for reporting in the figures. We also make our data available at https://osf.io/wnfxt/.
Results
Control measurements
Pharmacological manipulations of dopamine can elicit undesired side effects such as nausea, resulting in some participants withdrawing from the study (Chen et al., 2020a), and can also change mood state (Vo et al., 2018). None of our participants reported nausea, and all participants completed the study. Levodopa did not alter our participants’ blood pressure or mood, as Time (before drug administration, 1 h after drug administration, 2 h after drug administration) × Drug ANOVAs yielded evidence for the null hypothesis for Time × Drug interactions for blood pressure (diastole, BFexcl = 6.383; systole, BFexcl = 5.683) and mood (BFexcl = 3.681). Heart rate was reduced to a greater extent for participants on placebo (mean reduction in heart rate, 10.211 ± 1.828) than that on levodopa (mean reduction in heart rate, 1.737 ± 1.828), as shown by a Time × Drug interaction (BFincl = 6.519).
Participants were not above chance at accurately guessing the drug condition, as shown by Bayesian binomial tests (BF0+ = 4.685).
Manipulation check
Performance scaled with task difficulty, as larger aiming angles decreased reach accuracy, increased variability and lengthened reaction times (Fig. 2), as supported by main effect of angle for accuracy (BFincl = 2.25E + 117), error variability (BFincl = ∞) and reaction times (BFincl = ∞).
Trial-by-trial performance changes
We provided knowledge of the task structure and explicit strategies to achieve task goals, which typically leads to fast performance improvements over the first few trials (Mazzoni and Krakauer, 2006).
Rapid improvements in performance were evidenced in reductions in reaction times (Fig. 2B; main effect of trial, BFincl = 29,887), specifically from the first trial to all subsequent trials (comparisons with first trial, BF10 ranging from 6.44 to 1.75E + 8). Reaction times over the first eight trials were overall longer for the levodopa group (BF10 = 1,469). However, trial-by-trial reductions in reaction times did not differ between groups, as shown by evidence against the Drug × Trial interaction, BFexcl = 7,518.80. Peak velocity did not change across trials (evidence against a main effect of trial, BFexcl = 14.970), regardless of drug condition (evidence against the Drug × Trial interaction, BFexcl = 1,196.172).
On the first trial, levodopa increased accuracy compared with placebo, as shown by smaller mean errors than placebo [BF10 = 4.37; Cohen's d = 0.86 (95% CI −1.54, −0.163)]. This high level of accuracy was maintained across trials for the levodopa group, as shown by strong evidence for excluding the main effect of trial (BFexcl = 4,048.580) in contrast to the placebo group, who reduced error across trials (BFincl = 3.856). The levodopa group also showed greater accuracy in the first eight trials at follow-up compared with training (main effect of session, BFincl = 11.476; Fig. 2D, pink bars), whereas the placebo group showed similar accuracy in the first eight trials for the training and follow-up sessions (main effect of session, BFexcl = 17.730; Fig. 2D, gray bars). Thus, at least for the first eight trials of each aiming block, levodopa resulted in greater session-to-session gains in accuracy compared with placebo.
Bin-by-bin performance changes
In each session, participants completed 10 bins (80 trials) for each aiming angle. Bin-by-bin performance improvements were evidenced in reduced reaction times and error variability, as shown by strong evidence for the main effect of bin (reaction time, BFincl = 2,199.835; variability, BFincl = 90,621.064; Figure 2F,H), while accuracy and peak velocity stayed constant across bins, as shown by strong evidence for excluding the main effect of bin (accuracy, BFexcl = 90,909.0; peak velocity, BFexcl = 3.46E + 4).
Levodopa did not alter bin-by-bin changes in performance, as shown by strong evidence for excluding Drug × Bin interactions for accuracy (BFexcl = 5,49,176.79), reaction time (BFexcl = 3.59E + 6), peak velocity (BFexcl = 5.57E + 7), and error variability (BFexcl = 37,037.037) and strong evidence for excluding Drug × Angle × Bin interactions (accuracy, BFexcl = 6,134.969; reaction time, BFexcl = 3.45E + 7; peak velocity, BFexcl = 3.937E + 11; error variability, BFexcl = 20,040.080). Thus, while levodopa enhanced performance by increasing accuracy and error variability even within the first eight trials (i.e., the first bin), it did not change bin-by-bin performance improvements (Fig. 3).
Session-by-session performance changes
The follow-up session was completed a minimum of 18 h after the training session, which is sufficient to wash out residual effects of levodopa (Contin and Martinelli, 2010). Session-to-session improvements were prominent (Fig. 4), as shown by very strong evidence for main effects of session for accuracy (BFincl = 2.41e + 18), reaction time (BFincl = 1.86E + 33), peak velocity (BFincl = 3.35E + 190), and error variability (BFincl = 6.49e + 14).
Levodopa increased session-by-session performance improvements for reaction time for the largest, aiming angles (e.g., 165°), as shown by moderate evidence for Drug × Session × Angle interaction, BFincl = 4.175, where the reduction in reaction time from training to follow-up for the largest angle (165°) tended to be larger for the levodopa group [Cohen's d = 0.544 (95% CI: 0.279, 0.86)] than the placebo group [Cohen's d = 0.267 (95% CI: −0.0937, 0.725)]. For peak velocity, there was strong evidence for a Session × Drug × Angle interaction, BFincl = 4,117.00756, where levodopa slowed peak velocity both at training (Drug × Angle interaction, BFincl = 454.835) and follow-up (Drug × Angle interaction, BFincl = 3.29E + 7), but the effect of levodopa appeared more prominent at training (Fig. 4D, left panel) than at follow-up (Fig. 4D, middle panel).
While our trial-by-trial analyses showed that levodopa increased offline gains for accuracy for the first bin (first eight trials), this effect was no longer evident across subsequent aiming bins. Levodopa did not modulate session-by-session performance improvements for accuracy and error variability, as shown by moderate evidence for excluding Session × Drug interactions for error (BFexcl = 4.116) and error variability (BFexcl = 6.7376), and strong evidence for excluding Session × Drug × Angle interactions for error (BFexcl = 3.57E + 3) and error variability (BFexcl = 202.429).
Levodopa increased accuracy while lengthening reaction times under difficult aiming conditions
Figure 4A shows that levodopa administered before training altered aiming performance depending on aiming angle. Aiming accuracy differed depending on aiming angle: for 30°, participants tended to overcompensate (i.e., aim by slightly >30°; Fig. 4A), and this overcompensation was somewhat reduced with levodopa. For the larger, more difficult aiming angles (e.g., 150°, 165°), participants tended to undercompensate (aimed by less than the required angle), but this undercompensation was reduced with levodopa. This pattern of results was supported by Angle × Drug interactions for accuracy (BFincl = 1.67e + 12). Greater accuracy with levodopa co-occurred with lengthened reaction times, particularly with larger aiming angles, as shown by Angle × Drug interactions for reaction time (BFincl = 1.48E + 11; Fig. 4B). Notably, the effect of levodopa in increasing accuracy and lengthening reaction times were prominent both at training and follow-up, despite the fact that effects of levodopa consumed at training would have washed out by 8 h after consumption (Contin and Martinelli, 2010), before the start of training.
Thus, levodopa increased accuracy while lengthening reaction times, particularly with the more difficult, larger aiming angles. In a task where success depended on aiming accuracy, increasing dopamine availability via levodopa during training increased the prioritization of accuracy above speed of movement planning, suggesting increased adherence to task goals. This effect persisted at follow-up, after all levodopa effects were washed out. Exogenous dopamine at training thus appeared to result in a learnt, persistent prioritization of task goals, and this occurred in parallel with prominent session-to-session improvements in accuracy and speed of movement planning and movement execution that were mostly unaffected by exogenous dopamine.
Discussion
Here, we explored the causal role of dopamine in motor learning in neurotypical individuals, by testing how the performance of goal-directed motor strategies was affected by dopamine during training and by testing whether this manipulation altered performance at follow-up, when participants re-encountered the same task. We used a task with varying levels of difficulty and incentivized accurate performance by making task success contingent upon accuracy. Participants were explicitly instructed to use strategies to accurately aim at specified angles from presented targets. Both the task structure and task instructions emphasized accuracy and not speed of movement planning. Here, exogenous dopamine at training increased movement accuracy and lengthened the amount of time taken to plan movements, and this pattern of behavior was evident even within the first eight trials of each aiming angle. This pattern of behavior persisted at follow-up, after all effects of the dopamine medication had been washed out. We posit that this maintained effect reflects some form of learning. While there were marked session-to-session performance improvements, levodopa generally did not increase such offline performance gains, showing only a subtle effect in improving accuracy within the first bin (the first eight trials). Levodopa also did not increase online performance gains. We interpret this result to suggest dopamine altered behavior in an explicit motor learning task, primarily by instilling a learnt propensity to better adhere to task goals (here, to prioritize accuracy over speed of movement planning), not by augmenting the size of online or offline performance gains.
The finding that increasing dopamine availability increased accuracy in an explicit motor learning task that incentivized accuracy supports the view that dopamine influences decisions to engage instrumental motivation in motor learning. These results are consistent with reports that similar exogenous dopamine manipulations can increase deliberative control (Wunderlich et al., 2012; Sharp et al., 2016). Indeed, the focus on explicit motor strategies here might be why dopamine manipulations altered performance here, in contrast to previous null findings (Quattrocchi et al., 2018; Palidis et al., 2021). We provided participants with explicit instructions on how to aim by varying angles from a presented target, where larger aiming angles were associated with larger errors and lengthier reaction times and were more difficult. In contrast, previous work rewarded participants for small, possibly implicit changes in behavior (Palidis et al., 2021), or used traditional paradigms that engaged mutually compensatory implicit and explicit motor learning operations (Quattrocchi et al., 2018). Similarly, pharmacological manipulations of dopamine do not always result in observable effects on learning processes that have been characterized as automatic (Wunderlich et al., 2012; Sharp et al., 2016; Grogan et al., 2017, 2019). Dopamine manipulations often alter the contributions of deliberative, model-based learning processes relative to automatic, model-free learning processes (Wunderlich et al., 2012; Sharp et al., 2016), although exceptions exist (Kroemer et al., 2019). Future studies should dissect how dopamine modulates the relative contributions of deliberative and automatic processes, as automaticity is attained with training during skill learning.
A second feature of this study that differed from previous work (Quattrocchi et al., 2018; Palidis et al., 2021) was the focus on manipulating task success. Instead of manipulating extrinsic rewards (points or money) contingent upon achieving some performance criteria (Galea et al., 2015; Steel et al., 2016; Quattrocchi et al., 2018; Spampinato et al., 2019), we focused on success in achieving the task goal (target-hitting). Participants could only hit targets via accurate aiming, as the targets would move mid-movement by the required aiming angle. Task success thus required accurate performance. Our approach of making task success contingent upon accurate performance might have made our task more sensitive to the effects of exogenous dopamine. In contrast to direct manipulations of task success, which consistently yields clear effects on motor learning (Schaefer et al., 2012; Leow et al., 2018; Kim et al., 2019; Tsay et al., 2022), extrinsic rewards do not always alter motor learning (Steel et al., 2016; Spampinato et al., 2019) or motor performance (Grogan et al., 2020). To investigate how dopamine influences sensitivity to task success and extrinsic rewards, future studies can combine dopamine manipulations with paradigms that dissociate the contributions of task success and extrinsic rewards to performance (Vassiliadis et al., 2021).
Our finding that dopamine increases accuracy under more difficult task conditions is consistent with a view that dopamine guides instrumental motivation in execution of effort to attain accurate task performance. This extends previous research in humans, which has largely focused on the role of dopamine in motivation to select effortful options, as the majority of previous studies have tested how dopamine manipulations alters decisions to select effortful options in tasks where each decision is only associated with a small or random probability of actually executing that response option (for a review, see Lopez-Gamundi et al., 2021). For example, in effort-based decision-making paradigms, methylphenidate, a dopamine and noradrenaline reuptake inhibitor, increases young adults’ willingness to select more difficult working memory conditions (Westbrook et al., 2019). Dopamine denervation in Parkinson's disease patients results in steeper discounting of effort for reward when choosing between effortful options in comparison with controls (Chong et al., 2015), and this deficit is partially remediated by medications which increase dopamine availability (McGuigan et al., 2019). By showing here that dopamine increased accuracy in a task where participants could not opt out of any trial, but could only choose how well to execute each trial, we infer that dopamine not only alters willingness to select effortful options but also increases willingness to execute effort.
How might exogenous dopamine alter the engagement of instrumental motivation during skill acquisition? Influential theories suggest that slow “tonic” dopamine responses signal the background rate of reward and drive motivation (Niv et al., 2007). Fast “phasic” dopamine responses signal unexpected outcomes and help the actor learn to select choices that optimize outcomes (Steinberg et al., 2013). Recent studies however suggest that fast dopamine responses drive both motivation to work for rewards and learning to select options leading to better outcomes (Hamid et al., 2016). Here, we used levodopa, the precursor to dopamine, which increases dopamine availability within the brain. Although we do not fully understand how levodopa affects fast and slow time-scale dopamine responses, animal studies have demonstrated that levodopa increases phasic firing of striatal dopamine cells (Willuhn et al., 2014). We speculate that levodopa employed at training increased accuracy by increasing the amplitude of phasic dopamine cell firing in response to target error feedback. Accuracy improvements that persisted at follow-up might have been associated with training-specific changes in excitatory synaptic transmission in the striatum that were altered by exogenous dopamine at training (Yin et al., 2009).
The increase in accurate aiming with levodopa was accompanied by lengthier deliberation times, suggesting a more cautious mode of response overall—a classic speed/accuracy tradeoff. Intriguingly, increases in response caution have not been demonstrated in previous work examining how pharmacological manipulations of dopamine alters decision-making behavior using evidence accumulation modeling (Winkel et al., 2012; Beste et al., 2018; Rawji et al., 2020). In those studies, dopamine manipulations either had no effect on response caution (Winkel et al., 2012), or increased response errors (Huang et al., 2015), or decreased response caution only in the context of proactive inhibition (Rawji et al., 2020) or reinforcement learning task paradigms (Chakroun et al., 2022). These mixed findings might be due to methodological differences, such as the use of different pharmacological agents (e.g., bromocriptine, a dopamine D2 agonist in Winkel et al., 2012; ropinirole, a dopamine D3 agonist in Rawji et al., 2020; and methylphenidate, a dopamine and noradrenaline reuptake inhibitor in Beste et al., 2018) or the use of different tasks [e.g., simple decision-making tasks vs tasks requiring proactive inhibition in Rawji et al. (2020) or tasks requiring learning from reward feedback (Chakroun et al., 2022)]. Here, in a task that rewarded accuracy and not speed, exogenous dopamine resulted in participants trading off speed for accuracy, particularly for difficult task conditions. It will be important for future studies to investigate how dopamine alters performance in contexts where both speed and accuracy are incentivized (Manohar et al., 2015).
Skill learning is accompanied by a violation of the speed accuracy trade-off, where both speed and accuracy are improved (Krakauer et al., 2019). While both speed and accuracy of aiming movements improved markedly at follow-up compared with training, we did not find evidence that dopamine modulated the size of the improvements in speed and accuracy across sessions. This might be due to the timing of the dopamine manipulation. Consistent with our results, animal studies have shown that dopamine agonists and antagonists applied before learning have no effect on retention 24 h later but did alter retention when applied 3–6 h after initial learning (Bernabeu et al., 1997). Similarly, dopamine medications ingested by Parkinson's disease patients 8–24 h after learning improve subsequent recall when tested a day after drug ingestion (Grogan et al., 2015). Dopamine manipulations might thus have modulated offline consolidation of learning only in a critical post-task period: this possibility awaits investigation from future work that tests how the timing of exogenous dopamine affects consolidation.
While we favor the interpretation that dopamine increased motivation to adhere to task goals, increasing accuracy in a task that incentivized accuracy and not speed, we cannot discount the alternative possibility that dopamine increased accuracy by altering valuation of target hits/target misses relative to the cost of time. Indeed, these two interpretations may not be mutually exclusive. Unexpected outcomes such as target misses might evoke phasic responses from midbrain dopamine neurons, evoking a reward prediction error (Schultz et al., 1997). The dopamine manipulation here might have altered the processing of such reward prediction errors, increasing sensitivity to desired outcomes (i.e., target hits) and/or undesired outcomes (target misses) relative to the cost of time, such that actions prioritized accuracy at the cost of time. To disentangle the relative contributions of goal adherence versus target errors to motor performance, future experiments should systematically manipulate higher-order task goals (e.g., differentially incentivize speed vs accuracy) in contexts with or without target hits/misses.
Because we manipulated task success to incentivize accurate performance, we cannot make direct inferences about how dopamine alters “pure” intrinsic motivation to learn and perform motor skills. Increasing evidence demonstrates a key role of dopamine in learning in the absence of extrinsic rewards and in self-evaluating performance without any extrinsic feedback (Ripollés et al., 2018; Duffy et al., 2022). For example, in zebra finches, dopamine neurons show spontaneous activity that correlates with song performance, even in the absence of extrinsic rewards, cues, or external perturbations (Duffy et al., 2022). Similarly, pharmacological manipulation of dopamine in humans can alter learning of novel words in the absence of explicit reward or feedback, and effects remain prominent on a no-drug follow-up session (Ripollés et al., 2018). To test the role of dopamine in intrinsic motivation in motor learning, future studies can combine pharmacological manipulations of dopamine with behavioral paradigms devoid of any form of extrinsic performance feedback, for example, by removing or limiting visual feedback of movement (Vassiliadis et al., 2021) and/or having target appear only transiently at the start of the trial to preclude target hits or target misses (Taylor and Ivry, 2011).
Although we manipulated task success (i.e., target hits in a target-hitting task), our paradigm does not allow us to dissociate if behavior was motivated by avoidance of losses (target misses) or pursuit of rewards (target hits). A large body of knowledge demonstrates that rewards have potent effects in motivating effort (Takikawa et al., 2002; Xu-Wilson et al., 2009; Manohar et al., 2015, 2017; Codol et al., 2019), even when rewards do not depend on how well we perform (Manohar et al., 2017), or when rewards are presented outside our conscious awareness (Pessiglione et al., 2008). Reward effects are evident even in very early visuomotor responses (Carroll et al., 2019; Codol et al., 2023). Reward also makes us more willing to decide to engage costly cognitive resources, such as choosing to perform difficult working memory tasks, switch task sets, or inhibit prepotent responses (Westbrook and Frank, 2018). While less is known about the motivational effects of punishment avoidance, it is clear that avoidance of negative outcomes is a potent driver of behavior (Chen et al., 2020b). While our evidence is broadly supportive of the proposal that dopamine alters motor performance by altering the brain's sensitivity to rewards and increasing instrumental motivation, future studies should disambiguate if the effects shown here resulted from punishment avoidance or reward pursuit.
Our sample was predominantly female. It is clear that there are sex differences in brain dopamine function (Becker, 1999) and responsivity to pharmacological manipulations of dopamine (Dluzen and Ramirez, 1985). Furthermore, although the menstrual cycle phase alters dopamine receptor availability (Czoty et al., 2009), we did not account for individual differences in the menstrual cycle of our female participants. We note that our results were consistent even after removing male participants from our dataset. However, it is unknown if our results will generalize to male participants.
In conclusion, in a task which incentivized accuracy by making task success reliant on accurate motor performance, we find that pharmacologically increasing dopamine availability increased performance accuracy, while concurrently increasing the amount of time used to deliberate and prepare movements, demonstrating a propensity to better adhere to task goals of prioritizing accuracy over speed. This effect of dopamine was sustained even after an overnight delay, long after the effects of exogenous dopamine had washed out. This persistent propensity to prioritize accuracy at follow-up co-occurred with, but was dissociable from, the prominent session-to-session improvements in accuracy and reductions in movement planning time and variability. We interpret the results to suggest that dopamine plays a key role in decisions to engage instrumental motivation to better adhere to task goals, which not only determines the quality of motor performance at initial learning but also influences the quality of future motor performance when the same motor problem is reencountered. While studies in animals and clinical populations have previously demonstrated a role for dopamine in motor learning (Hosp et al., 2011; Isaias et al., 2011), this is, to the best of our knowledge, one of the first evidence for a direct link between dopamine and motor skill learning in neurotypical humans.
Footnotes
The authors declare no competing financial interests.
The authors would like to thank Laura Fassbender for assistance with data collection. This work was supported by Australian Research Council (DP180103081) and Brain Foundation (RM 2019001768).
This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.