Abstract
When learning a new motor skill, people often must use trial and error to discover which movement is best. In the reinforcement learning framework, this concept is known as exploration and has been linked to increased movement variability in motor tasks. For locomotor tasks, however, increased variability decreases upright stability. As such, exploration during gait may jeopardize balance and safety, making reinforcement learning less effective. Therefore, we set out to determine if humans could acquire and retain a novel locomotor pattern using reinforcement learning alone. Young healthy male and female participants walked on a treadmill and were provided with binary reward feedback (indicated by a green checkmark on the screen) that was tied to a fixed monetary bonus, to learn a novel stepping pattern. We also recruited a comparison group who walked with the same novel stepping pattern but did so by correcting for target error, induced by providing real-time veridical visual feedback of steps and a target. In two experiments, we compared learning, motor variability, and two forms of motor memories between the groups. We found that individuals in the binary reward group did, in fact, acquire the new walking pattern by exploring (increasing motor variability). Additionally, while reinforcement learning did not increase implicit motor memories, it resulted in more accurate explicit motor memories compared with the target error group. Overall, these results demonstrate that humans can acquire new walking patterns with reinforcement learning and retain much of the learning over 24 h.
Significance Statement
Humans can learn some novel movements by independently discovering the actions that lead to success. This discovery process, exploration, requires increased motor variability to determine the best movement. However, in bipedal locomotion especially, increasing motor variability decreases stability, heightening the risk of negative outcomes such as a trip, injury, or fall. Despite this stability constraint, the current study shows that individuals do use exploration to find the most rewarding walking patterns. This form of learning led to improved explicit retention but not implicit aftereffects. Thus, the reinforcement learning framework can explain findings across a wide range of motor and cognitive tasks, including locomotion.
Introduction
In the 2020 Summer Olympics, US athlete Sydney McLaughlin broke the world record and won the gold medal in the 400 m hurdles, aided by her proficiency in jumping over the hurdles equally well with both her dominant and nondominant legs. While reward (clearing the hurdle) is ostensibly a critical learning signal for this and other tasks, its specific role in locomotor learning remains unknown. Rewards influence learning via reward prediction errors, the difference between expected and actual rewards (Rescorla and Wagner, 1972). These error signals form the basis for reinforcement learning theory, a widely used framework describing how animals and machines learn the values of actions (Sutton and Barto, 2017). A core component of reinforcement learning is exploration, defined as sampling from a range of possible actions to update their values. In movement-based reinforcement learning, one sign of exploration is increased motor variability above baseline motor noise as this likely aids in motor learning (Therrien et al., 2016, 2018; Cashaback et al., 2017, 2019; Dhawale et al., 2017). Thus, we refer to exploration in the context of motor learning as an increase in motor variability above baseline motor noise. However, in locomotion, the cost of changing motor variability is much higher, as too much or too little variability in gait patterns reduces stability and could lead to catastrophic consequences like loss of balance, falls, or injuries (Hausdorff et al., 1997, 2001; Maki, 1997; Brach et al., 2005; McAndrew Young and Dingwell, 2012).
During motor tasks, reinforcement learning is often probed by providing only binary reward feedback and has been demonstrated in reaching (Izawa and Shadmehr, 2011), speech production (Parrell, 2021), and saccadic eye movements (Madelain et al., 2011). Since binary reward feedback does not provide information about the direction or distance from a task goal, the only signal by which to learn the correct movement is a reward prediction error, resulting in exploratory behavior. In sharp contrast, learning signals such as target error (i.e., the difference between a task goal and movement feedback) and sensory prediction error (i.e., the difference between the expected and actual sensory consequences of movements) convey both magnitude and directional information. As a result, they produce comparatively less exploration than reward prediction errors (Izawa and Shadmehr, 2011; Therrien et al., 2016; Cashaback et al., 2019). To date, no studies have determined if binary reward feedback can produce learning, and the expected exploratory behavior, during locomotion.
Reinforcement learning also impacts motor memories. Implicit aftereffects, defined as motor memories not under conscious control (Krakauer et al., 2019), are strengthened when reinforcement is combined with sensorimotor adaptation (Huang et al., 2011; Shmuelof et al., 2012; Galea et al., 2015) or use-dependent learning (Mawase et al., 2017; Bao and Lei, 2022; c.f., Tsay et al., 2022). Reinforcement learning may also strengthen explicit retention, defined as the ability to consciously remember and reproduce a previously learned movement (Schmidt and Lee, 2005), potentially because individuals benefit from determining what the successful movement is themselves through exploration, improving engagement in the task, unlike when receiving full visual feedback of movements (Winstein and Schmidt, 1990; Winstein et al., 1994; Hasson et al., 2015). However, previous studies examining explicit retention in reinforcement learning either did not induce learning with purely binary reward feedback (Hasson et al., 2015; Codol et al., 2018) or did not include a comparison group (Holland et al., 2018). Therrien et al. (2016) found retention was better after binary reward feedback than sensory prediction and target errors, but this probe into retention may have also included a strong implicit component. Therefore, the extent to which reinforcement learning alone contributes to implicit and explicit motor memory formation remains unclear.
The primary objective of this study was to determine if individuals can acquire a novel gait pattern using reinforcement learning. We tested this by providing binary reward feedback to participants learning to walk with a longer left step length than baseline. For comparison, we recruited a group that performed the same walking pattern using explicit, target error feedback, which does not induce significant exploration. Prior work has demonstrated that use-dependent learning is the predominant source of implicit aftereffects in this target error paradigm (Wood et al., 2020, 2021), and neither implicit aftereffects nor explicit retention is contaminated by sensory prediction errors. We hypothesized that the reward prediction error group would learn the new walking pattern using greater exploration (a greater increase in motor variability above baseline) and that they would have larger implicit aftereffects and more accurate explicit retention compared with the target error group.
Materials and Methods
Participants
A total of 59 (34 female, 25 male) young, healthy adults were recruited from the University of Delaware community. Within two separate experiments, participants were randomized into one of two groups: a reward prediction error (RPE) group and a target error (TE) group. Participants were excluded if they had any orthopedic or neurologic conditions that might interfere with their ability to learn a new walking pattern. Five participants were removed from the analysis (two due to technical errors, two due to not following instructions, and one who was provided an erroneous verbal reminder of the instructions during learning). Thus, 30 participants were included in experiment 1 and 24 in experiment 2. Before participating, all individuals signed a written, informed consent document. This study was approved by the University of Delaware Institutional Review Board.
Experimental design
Feedback
Participants walked on a dual-belt treadmill (Fig. 1A; Bertec) at a self-selected, comfortable but brisk pace, constrained between 1.0 and 1.2 m/s (Wood et al., 2020, 2021). This allowed each participant to walk at a comfortable pace given their biomechanical constraints, preventing differences in task difficulty due to treadmill speed. The belts moved at the same speed throughout the experiment. Participants wore a ceiling-mounted harness, which did not provide body weight support, and lightly held onto a handrail for safety while walking. A computer monitor placed in front of the treadmill provided real-time feedback of participants’ left step lengths (The MotionMonitor Toolbox, Innovative Sports Training). We chose to focus on step length because it is a robustly modifiable gait parameter that has frequently been the focus of locomotor learning paradigms (Reisman et al., 2005; Roemmich et al., 2016; French et al., 2018, 2021; Wood et al., 2020, 2021) and because both implicit and explicit forms of motor memory have been observed after learning an altered step length pattern (French et al., 2018, 2021; Wood et al., 2020, 2021).
We compared a group that received reward prediction error feedback (RPE group) to a group that received explicit task instructions and visual target errors (TE group). The TE group received real-time visual feedback of their step length relative to a step length target (Fig. 1B, top row). The left step length appeared as a blue bar, starting with a height of “0” and growing vertically during the swing phase, then holding position on the screen at its final height at the time of heel strike, until the next swing phase began. The TE group was instructed in the definition of a step length and that they needed to hit a pink horizontal target line exactly with the blue bar that represented their left step length. They were also informed how the target left step length would change during learning. Although they could see if they hit the target line, no external reward was provided either on the screen or verbally by the experimenter when they successfully hit the target. Thus, since the TE group knew exactly where to step and could correct for any target errors throughout the task, they had no need to explore. Furthermore, since the feedback was veridical, sensory prediction errors could not contaminate either implicit aftereffects or explicit retention.
The RPE group was provided with binary reward feedback (Fig. 1B, bottom row) that was tied to a fixed monetary bonus. This feedback took the form of a large green check mark and money added to a total when a left step length fell within the target window, the same target as the TE group, but here, invisible to participants. If successful, the check mark appeared on the screen at heel strike and remained on the screen until the next swing phase began. When the left step length fell outside of the target window, no check mark appeared, and no money was added to the total. A running tally of the total reward earned was consistently displayed at the top of the screen during the learning phase. The RPE group was told that their goal was to gain as much money as possible. As awareness of the rewarded parameter may be important for movement-based reinforcement learning (Manley et al., 2014), individuals in the RPE group were told that the reward was contingent on their left step length. The experimenter then explained and demonstrated a left step length until the participant could explain the concept themselves. Based on pilot testing, we also told the participants in the RPE group to start off by walking normally so they could achieve some success at the beginning of the learning phase, but critically, no information was provided regarding if or how the rewarded step length would change. Thus, to learn the longer left step length, the RPE group was provided with no other learning signal except a reward prediction error, the difference between the expected and actual reward. Participants in both groups were provided with a brief, verbal reminder of their instructions every 100 strides (TE, “hit the target”; RPE, “gain as much money as possible”).
Participants in the RPE group were told that they would earn a base dollar amount ($10) per visit, plus they had the potential to earn additional “bonus” money, based upon how well they performed the task (maximum $18 bonus). However, in actuality, all participants earned the same amount, which was the base amount plus the maximum bonus amount, regardless of task performance. Participants in the TE group were told that they would earn a base dollar amount per visit, which was equivalent to the amount paid to participants in the RPE group. This payment method was chosen, with IRB consultation, to ensure fairness, that is, to prevent differences in compensation that depended on group assignment or learning abilities.
Experimental schedules
Participants in both groups walked under three different phases: baseline, learning, and postlearning (Fig. 1C). During baseline walking (250 strides), participants were asked to walk normally and no visual feedback was provided. Next, participants performed a learning phase (900 strides), where they gradually learned to lengthen their left step length by 10% from their baseline using the feedback displayed (step length bars and target lines for the TE group or check marks and tally of money earned for the RPE group) on the computer monitor. While the feedback was different between the groups, the learning schedule was the same. For the first 50 strides of the learning phase, the target was centered on each individual's baseline left step length. The target step length then increased by 1% every 10 strides until it reached its final 10% change from baseline where it remained for the rest of learning (strides 141–900). The target maintained a ± 2 cm margin of error throughout the learning phase. Thus, while the center of the target was offset by a portion of each participant's baseline step length, the width of the target was the same for everyone which prevented any differences in difficulty of the task due to baseline step lengths. This target window was well within baseline left step length variability for participants in our sample.
The postlearning phases were used to probe either implicit aftereffects (experiment 1) or explicit retention (experiment 2). Therefore, the visual feedback was turned off during this phase for both experiments, but the instructions were slightly different. In experiment 1, participants completed a washout phase (900 strides) where they were asked to look forward and “walk normally.” Any left step length differences from baseline during this phase represented the implicit aftereffect, a motor memory not under conscious control. In experiment 2, participants completed two explicit retention tests, one immediately after learning (25 strides) and one 24 h later (250 strides). For both tests, participants were instructed to remember how they walked at the end of the learning phase and repeat that walking pattern exactly. Thus, explicit retention measured the ability to consciously remember and reproduce the previously learned step length pattern, exactly. For both experiments, participants in the RPE group were told that no additional money could be earned during these periods.
Data collection
Step length was calculated in real time using The MotionMonitor which received kinematic and kinetic data from Nexus software (Vicon Motion Systems). Kinetic data, sampled at 1,000 Hz, was acquired from the instrumented force plates through each treadmill belt (Bertec), and kinematic data, sampled at 100 Hz, was acquired from an eight-camera Vicon MX40 motion capture system with Nexus software. To capture the kinematic data, we used a custom marker set with seven retroreflective markers, one on each heel, lateral malleolus, and fifth metatarsal head and one on the left first metatarsal. The heel markers were used to determine left step length by calculating the difference along the laboratory's y-axis (anterior–posterior dimension) between the left and the right heel markers at the moment of left heel strike. Left heel strike was determined in real time by detecting when there was a ground reaction force on the left belt and the velocity of the left heel marker first transitioned from positive to negative (Zeni et al., 2008).
Data analyses
After data were collected, we used custom-written MATLAB scripts (v2022a The MathWorks) to calculate and plot the outcome variables. The primary outcome measure was percent left step length change from baseline (ΔLSL) calculated on each stride (s):
We tested the hypothesis that the RPE group could learn a new gait pattern by averaging ΔLSL during late learning. Learning in the RPE group would be reflected in a ΔLSL reliably greater than baseline. In addition, to observe how learning emerged in both groups over time, we calculated two secondary outcome measures during early and late learning. Error was calculated as the absolute difference between ΔLSL and the center of the ΔLSL target (10%), and success was calculated as the percentage of left step lengths that fell within the target step length window.
In stable reward environments (in the current study, once the target reached its plateau), exploration should decrease as the values of actions are learned (Daw et al., 2006; Sutton and Barto, 2017). Based on this theory and our assumption that the TE group required little to no exploration, we made two predictions: (1) During early learning, the RPE group would demonstrate greater exploration than the TE group; and (2) the RPE group would reduce their exploration by a greater magnitude from early to late learning compared with the TE group. Our primary measure of exploration, σΔLSL, was calculated as the standard deviation of ΔLSL, normalized by the ΔLSL standard deviation during the last 50 strides of baseline. Therefore, this measure reflects the increase in motor variability above baseline motor noise. These two predictions served as our test of the hypothesis that the RPE group would demonstrate greater exploration during learning compared with the TE group.
Exploration can also be characterized using a win-stay/lose-shift model. When an individual performs unsuccessful actions, the trial-to-trial changes tend to be more variable than after successful actions (Pekny et al., 2015; Cashaback et al., 2019; Uehara et al., 2019). Therefore, using all of the learning phase data except for the first 50 strides (i.e., before the target started shifting), we performed this trial-to-trial change analysis secondarily to our main exploration analysis. We calculated the difference in ΔLSL values between subsequent strides separately after successful steps and unsuccessful steps. The trial-to-trial change measure was calculated as the standard deviation of these changes, normalized by baseline trial-to-trial change variability (Cashaback et al., 2019; Roth et al., 2023), again to account for increases in this variability relative to baseline. We denote this measure as σtrial-to-trial.
For implicit aftereffects and explicit retention, we controlled for any potential differences in learning by calculating both as a percentage of the learned gait pattern for each individual participant on each stride. For experiment 1, we calculated immediate percent implicit aftereffect as the mean of the first five strides of the washout phase (immediate washout) and early percent implicit aftereffect as the mean of the next 25 strides (early washout). Therefore, we compared the implicit aftereffects at both timepoints between groups to test the hypothesis that the RPE group would show larger implicit aftereffects than the TE group. For experiment 2, we calculated the percent explicit retention at two different timepoints: the mean of all 25 strides during immediate retention and the mean of the first 25 strides during 24 h retention. Here, we considered 25 strides so that individuals had time to achieve a consistent level of performance that reflected their explicit motor memory. For experiment 2, it was also necessary to calculate absolute retention error, the absolute difference between ΔLSL at late learning and ΔLSL at immediate and 24 h retention. This additional measurement of error, or accuracy, was necessary because, unlike for implicit motor memories, we expected that explicit motor memories could either overshoot or undershoot the remembered target location, but the goal was to reproduce the previously learned stepping pattern as accurately as possible. Thus, absolute retention error at both timepoints served as our test for the hypothesis that explicit retention would be more accurate for the RPE group compared with the TE group.
Statistical analysis
We used Bayesian statistics to make inferences regarding differences within and between groups. This allowed us to estimate the specific probability that our hypotheses were supported or not, given the data we collected as opposed to the classical p value, which provides the probably of obtaining the data, assuming the null hypothesis was true. We assumed data were generated from Student's t distributions with a mean and standard deviation that varied based on both between (group) and within (time) subject factors. We estimated the mean and standard deviation parameters by combining our prior assumptions regarding their values with evidence from the data we collected. We set wide and uninformative priors for each of the model parameters similar to the approached described by Kruschke (2014). We calculated the joint posterior probability distribution with the Python package bayes-toolbox (Kim, 2023) which uses the Pymc library (v4.3; Salvatier et al., 2016) in Python (version 4.3.0). The joint posterior distribution was estimated using Markov chain Monte Carlo (MCMC) sampling, with 10,000 draws in four chains (i.e., 40,000 total samples) and 2,000 tuning samples in each chain. We performed diagnostics for each model, ensuring parameter values were consistent across chains and checking posterior estimates for possible errors (Kruschke, 2014; McElreath, 2016).
We made inferences regarding differences between groups based on posterior contrast distributions: the difference between two posterior distributions of interest. When these contrasts represent a between group difference at a specific timepoint, we use the term “group difference,” and when they represent differences across time between groups (analogous to an interaction effect), we use the term “slope difference.” We report the mean difference of each posterior contrast distribution along with the 95% high density interval (HDI), defined as the narrowest span of credible values that contain 95% of the posterior contrast distribution (Kruschke, 2014). We also report the probability of a difference given the data we collected as a percentage of the posterior contrast distribution that lies on one side of 0 (e.g., 85.6% > 0).
Code accessibility
The data analysis code for individual participants and figure generation, as well as the full statistical models and code, including detailed information about the priors and diagnostics are available online at the Open Science Framework (https://osf.io/tcwu6/).
Results
In experiment 1, 15 individuals (9 females; mean age ± 1 SD, 22.8 ± 4.2 years) were randomized into the RPE group and 15 (10 females; 22.5 ± 3.7 years) were randomized into the TE group. In experiment 2, a new sample of individuals was recruited, with 12 (6 females; 23.5 ± 4.6 years) randomized into the RPE group and 12 (7 females; 22.3 ± 4.2 years) randomized into the TE group.
Experiment 1
The RPE group learned the novel locomotor pattern
First, in Figure 2, we display individual learning data. Individual participants displayed in this figure represent their respective groups’ 10th, 50th, and 90th percentile values for our measure of exploration (σΔLSL) during early learning. Individuals in the TE group performed consistently with one another across the learning phase, regardless of the percentile (Fig. 2, top row), which is not surprising given they received accurate visual feedback of their step length and the target. However, individuals in the RPE group demonstrated a much wider range of idiosyncratic behavior during the learning phase across percentiles (Fig. 2, bottom row). The individual representing the 10th percentile performed similar to the individuals in the TE group (Fig. 2, bottom left), while the two other individuals show higher variability overall, along with what appears to be different patterns of exploration as learning progressed.
Despite the idiosyncratic behavior of the RPE group, overall, they learned the new walking pattern by the end of the 900-stride learning phase (Fig. 3A). The RPE group's ΔLSL was much larger than baseline (Fig. 3B; RPE late learning, mean [95% HDI] = 9.16% [8.36 9.95], 100% > 0), providing a high degree of certainty that the RPE group learned the new walking pattern. The TE group's ΔLSL at late learning was slightly greater by comparison (group difference = −0.94% [−1.99 0.02], 96.8% < 0). Additionally, the RPE group's performance was greatly improved from early to late learning compared with the TE group (Fig. 3C,D). Specifically, compared with the TE group, the RPE group had much larger improvements from early to late learning in terms of error (slope difference = 1.22% [−0.07 2.47], 97.1% > 0) and percent success (slope difference = −12.0% [−23.6 0.8], 97.1% < 0), an effect due to the much larger error in the RPE group in early learning. Together, these findings indicate that the RPE group learned the novel gait pattern, though not as completely, or as quickly, as the TE group.
The RPE group explored to learn the novel gait pattern
The results supported the hypothesis that individuals in the RPE group explored to learn the new gait pattern. We found differences in σΔLSL between the RPE and TE groups during the learning phase (Fig. 4A,B). Specifically, σΔLSL was much greater during early learning for individuals in the RPE group compared with that for individuals in the TE group (group difference = 1.20 [−0.02 2.57], 97.8% > 0), and the σΔLSL decreased more from early to late learning for the RPE group compared with that for the TE group (slope difference = 0.99 [−0.27 2.31], 94.1% > 0). These two results provide strong evidence for the hypothesis that the RPE group explored to learn the new walking pattern.
In addition to exploration at the two specific timepoints, we also performed a secondary analysis of exploration, measured as the variability of trial-to-trial changes after unsuccessful trials across most of the learning phase (Fig. 4C). The RPE group had slightly greater σtrial-to-trial after unsuccessful steps compared with the TE group (group difference = 0.37 [−0.17 1.01], 90.8% > 0). And the difference between σtrial-to-trial for successful versus unsuccessful steps was also slightly larger for the RPE group compared with that for the TE group (slope difference = 0.31 [−0.25 0.97], 84.6% > 0). Overall, while less reliable than the σΔLSL differences (probabilities of ∼91 and 85%), these findings provide additional support for the hypothesis that the RPE group was exploring during locomotion.
Reinforcement learning did not increase the magnitude of implicit aftereffects
We probed implicit aftereffects by asking individuals to walk normally during the no-feedback washout phase. Implicit aftereffects were greater for the TE group during initial washout (group difference = −10.7% [−23.8 2.8], 94.9% < 0) and during early washout (group difference = −9.6% [−17.4 −1.9], 98.9% < 0). Therefore, contrary to our hypothesis, learning from reward prediction errors did not increase implicit aftereffects more than learning from target error; rather, the evidence strongly supported the opposite. This finding could reflect a reduced amount of use-dependent learning for the RPE group caused by the increased variability (Wood et al., 2020, 2021).
Experiment 2
Learning from reward prediction errors versus target errors resulted in unique signatures of explicit retention
In experiment 2, we compared explicit retention after learning from reward prediction errors and target errors. This was done with a new sample of individuals who experienced a different postlearning phase but the same baseline and learning phases. We replicated the main findings from experiment 1 (see code for the remainder of the analyses). The RPE group learned (RPE ΔLSL at late learning = 8.74% [7.82 9.66], 100% > 0), and they did so by exploring (group σΔLSL difference at early learning = 0.87 [0.07 1.65], 98.6% > 0; σΔLSL slope difference = 0.54 [−0.24 1.37], 89.9% > 0). After learning, both groups performed a test of explicit retention where they were instructed to repeat the same walking pattern they performed at the end of learning without feedback, both immediately after learning and 24 h later (Fig. 5A).
First, we assessed the accuracy of the explicit motor memories by comparing the absolute error between the ΔLSL behavior at the end of learning and during the retention phases (Fig. 5B). Recall that this is the primary outcome measure of explicit retention because it reflects the true accuracy of retention or how far each individual's steps were from the remembered target, regardless of direction. The RPE group demonstrated less absolute retention error at immediate retention compared with the TE group (group difference = −1.36% [−2.94 0.20], 95.8% < 0), but after 24 h, the RPE group was very similar to the TE group (group difference = −0.13% [−2.01 1.68], 56.5% < 0). Thus, the experiment 2 results revealed that learning a new gait pattern using reward prediction errors produces a unique pattern of explicit retention compared with correcting for target error that is, at least initially, more accurate.
We found large between group differences in the specific walking pattern that was remembered both immediately and 24 h after learning (Fig. 5C). Specifically, percent explicit retention was much larger for the TE group at both immediate (group difference = −44.4% [−63.3 −24.9], 100% < 0) and 24 h retention (group difference = −45.8% [−71.4 −20.0], 99.8% < 0) compared with the RPE group. Recall that here, “larger” retention does not equate to “better” retention because we asked individuals to walk “exactly” like they did at the end of learning. Rather, this result reflects the fact that the TE group performed a step length that was longer than what they performed at the end of learning (i.e., overstepping), resulting in percent retention values mostly >100% (9/12 and 8/12 participants for immediate and 24 h retention, respectively). On the other hand, the RPE group performed a step length that was shorter than what they performed at the end of learning (i.e., understepping), creating in percent retention values mostly <100% (8/12 and 10/12 participants for immediate and 24 h retention, respectively).
Explicit retention resulted from differences in declarative knowledge of the successful step length
Why did the RPE and TE groups have such distinct patterns of percent explicit retention? One possibility is that declarative knowledge of the successful step length impacted explicit retention. To investigate this, we used our postsession interviews after the 24 h retention phase to probe declarative knowledge of successful step lengths. We asked each individual to rate the length of the successful step length compared with their baseline left step into one of three categories: “not longer,” “slightly longer,” or “longer.” The proportions of individuals in each category were as follows: not longer, TE 0%, RPE 25%; slightly longer, TE 50%, RPE 50%; and longer, TE 50%, RPE 25%. On average, the TE group rated the successful step length as longer (by about one category) than the RPE group (group difference = −1.00 [−2.00 −0.02], 97.7% < 0). Thus, the different feedback and instructions led to differences in declarative knowledge of the successful step length.
Discussion
This works shows, for the first time, that despite the stability constraints inherent in gait, young healthy individuals can learn a new walking pattern using reward prediction errors. Consistent with the reinforcement learning framework, individuals in the RPE group learned the most valuable action by initially exploring the action space, assigning values to possible step lengths, and then as learning progressed, exploiting the learned action/value pairs with more consistent, correct step lengths. We also found that reinforcement learning produced smaller magnitudes of implicit aftereffects, but more accurate explicit retention, than individuals provided with target errors and a specific task strategy, an effect that waned by 24 h.
Reinforcement learning during locomotion is accomplished by exploring
Prior studies of locomotor learning have not isolated reward prediction errors by providing only binary reward feedback. Rather, they either added rewards to other teaching signals like target error (Sato et al., 2022) or sensory prediction errors (Bakkum and Marigold, 2022), or used graded reward feedback, increasing the reward as movements became closer to a target (Hasson et al., 2015). While graded reward feedback provides a reward prediction error, having any knowledge about distance or direction from the target adds another learning signal similar to target error. We therefore provided only binary reward feedback in the current study to isolate reinforcement learning as much as possible. The RPE group learned the new walking pattern, a finding consistent with prior studies in reaching (Izawa and Shadmehr, 2011; Therrien et al., 2016; Cashaback et al., 2017; Holland et al., 2018), eye movements (Madelain et al., 2011), and speech production (Parrell, 2021). Unlike a prior study in reaching (Wu et al., 2014), baseline variability was not related to the rate of learning for the RPE groups across both experiments (slope = 0.5 [−2.8 3.8], 63.9% > 0), but our ability to assess the rate of learning could be hindered by the gradual nature of the perturbation in the current study.
To learn the new walking pattern, the RPE group explored the action space. We observed a classic signature of reinforcement learning behavior, the exploration/exploitation trade-off, in the RPE group's step length variability measurement, σΔLSL. Interestingly, prior reinforcement-based motor learning studies in human reaching have not observed a reduction in exploratory behavior over time in stable reward environments. We attribute this to the relatively short (360–400 trials; Pekny et al., 2015; Uehara et al., 2019), or absent (Izawa and Shadmehr, 2011; Therrien et al., 2016; Cashaback et al., 2017) steady-state learning phases in those studies. However, in the current study, steady-state learning was 760 strides which allowed the RPE group adequate time to learn the reward landscape well enough to exploit the rewarding step length.
While parsimonious, motor variability is not the only way to characterize exploration in movement-based reinforcement learning experiments. Some methods have characterized exploration as a random walk (Haith and Krakauer, 2014; Roth et al., 2023), but most methods have focused on the magnitude or variability of trial-to-trial of changes after successful versus unsuccessful movements (i.e., win-stay/lose-shift methods; Pekny et al., 2015; Therrien et al., 2016; Cashaback et al., 2019; Uehara et al., 2019; van Mastrigt et al., 2020, van Mastrigt et al., 2021). However, no studies have compared the latter measure of exploration while learning from reward prediction error versus correcting for target errors. While we found that the RPE group had slightly greater trial-to-trial variability after unsuccessful steps compared with the TE group, the effect was relatively small compared with the effect in baseline-normalized motor variability. It is possible other models of exploration, for example, random walk behavior (Roth et al., 2023), may better distinguish between reward prediction error learning and target error correction. Or perhaps these methods of exploration could be task dependent. Future work could elucidate different methods of movement-based exploration similar to the distinct modes of exploration described in decision-making tasks (Wilson et al., 2014).
Reinforcement learning does not increase the magnitude of implicit aftereffects
Reinforcement learning did not increase implicit aftereffects as we expected, and the RPE group actually produced smaller implicit aftereffects compared with the TE group which was not provided with any overt reward. It is possible this difference was due to the increased percent success for the TE group. We performed a Bayesian regression analysis to determine if percent success, measured during the learning phase while the target was at its 10% plateau, could predict immediate implicit aftereffects. We found that the 95% HDIs surrounded a slope value of zero (RPE group slope mean [95% HDI] = 0.32 [−1.56 2.15]; target group = 0.81 [−0.57 2.27]). Thus, while percent success could have had a positive effect on implicit aftereffects for the target group, these results are unreliable, a finding that is consistent with our prior work (Wood et al., 2020, 2021).
We suspect the implicit aftereffects in the TE group were caused by use-dependent learning because the feedback was similar to prior studies in our lab and the implicit aftereffect persisted throughout the 900-stride washout phase (Wood et al., 2020, 2021). It is possible that implicit aftereffects in the RPE group were also caused by use-dependent learning as suggested previously (Codol et al., 2018; Holland et al., 2018). If so, the current results contrast somewhat with findings in sensorimotor adaptation (Huang et al., 2011; Shmuelof et al., 2012; Galea et al., 2015; Therrien et al., 2016) and use-dependent learning (Floel et al., 2008; Mawase et al., 2017) in which reward boosts implicit aftereffects caused by other learning processes. However, the current results align with work suggesting that increased motor variability attenuates use-dependent learning (Verstynen and Sabes, 2011; Marinovic et al., 2017; Wood et al., 2021; Tsay et al., 2022). Indeed, if we match individuals across groups based on non-normalized, total motor variability (calculated across all strides of the learning phase when the target was plateaued at 10%), the difference in implicit aftereffects disappears (n = 6 per group; initial washout difference = −2.3% [−62.1 61.0], 53.9% < 0; early washout difference = −6.1% [−59.0 47.8], 60.5% < 0). Therefore, we suggest implicit aftereffects in both groups were due to use-dependent learning since, after controlling for motor variability, both groups demonstrated a similar magnitude of implicit aftereffects.
Reinforcement learning improves the accuracy of explicit retention
Despite less explicit knowledge of the task itself, including the mapping between step length and target position, the RPE group had more accurate immediate explicit retention compared with the TE group. This increased accuracy could have been caused by the reward itself (money and green checkmark) by solidifying the memory of the walking pattern better than simply hitting a target without overt reward. This explanation is similar to Thorndike's law of effect (Thorndike, 1927) and is consistent with studies showing that rewarding successful movements during skill learning and sensorimotor adaptation result in better motor memories than without reward (Abe et al., 2011; Shmuelof et al., 2012; Therrien et al., 2016; Mawase et al., 2017). However, we did not find evidence of a relationship between task success and absolute retention accuracy at immediate retention for either group (RPE = 0.05 [−0.06 0.15]; TE = 0.03 [−0.16 0.20]). Another possibility is that the act of exploring produces a clearer picture of the successful action(s) needed within the workspace, which then leads to more accurate explicit retention. Studies across various learning tasks suggest that greater variability produces worse performance during training but better memories/generalization (Raviv et al., 2022). A recent study in a grid-sailing, key press task suggested that more variable training leads to better knowledge of novel action outcome pair mappings (Velázquez-Vargas et al., 2023). Therefore, it is possible that by exploring the space, individuals in the RPE group developed better knowledge of the action outcome pair.
Reinforcement learning and correcting for target errors produce distinct explicit motor memories
Interestingly, we found consistent and unique biases in explicit motor memories between groups when assessed as percent explicit retention, a measure that takes the direction of error into account. During both explicit retention phases, the RPE group was biased to step shorter than they did at the end of the learning phase, while the TE group was biased to step longer than they did at the end of the learning phase. Participants’ declarative memories of the successful step length followed a similar pattern (i.e., more individuals in the TE group reported stepping “longer”), suggesting their bias was likely cognitive in nature. While the link between declarative knowledge and different forms of motor learning and memory has been made previously (Shanks and Johnstone, 1999; Keisler and Shadmehr, 2010; Stanley and Krakauer, 2013; Haith et al., 2015; Wong et al., 2015; McDougle et al., 2022), we are not aware of previous studies diplaying such stark biases in declarative motor memories. We speculate that these biases were caused by heuristics, computationally inexpensive mental shortcuts, which not only produce biases in cognitive decisions (Tversky and Kahneman, 1974) but can also explain biases in visual and auditory perception (Raviv et al., 2012; Rahnev and Denison, 2018; Gardner, 2019; Branch et al., 2022).
Conclusion
Here, we determined that reinforcement learning can drive the acquisition and retention of new gait patterns. Consistent with the reinforcement learning framework, individuals do so by exploring the action space until the most rewarding movement is discovered. Then, the most rewarding action is exploited by repetition. Thus, the reinforcement learning framework can describe performance across a wide range of species and behavior, and bipedal human locomotion is no exception, as our environment is rife with risks (e.g., a trip over a cracked sidewalk) and rewards (e.g., clearing a hurdle on the way to winning a race) that we must learn to pair with specific actions to successfully navigate in the world.
Footnotes
The authors declare no competing financial interests.
We thank Saunders Penn for his help with data collections and making Figure 1A and Maci Keaton for her help with data collections.
This work was supported by the National Institutes of Health, S10-RR028114 (S.M.M.), K12 HD055931 (H.E.K.), and the University of Delaware Graduate College through the University Dissertation Fellowship (J.M.W.).
This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.