A popular model of human sensorimotor learning suggests that a fast process and a slow process work in parallel to produce the canonical learning curve (Smith et al., 2006). Recent evidence supports the subdivision of sensorimotor learning into explicit and implicit processes that simultaneously subserve task performance (Taylor et al., 2014). We set out to test whether these two accounts of learning processes are homologous. Using a recently developed method to assay explicit and implicit learning directly in a sensorimotor task, along with a computational modeling analysis, we show that the fast process closely resembles explicit learning and the slow process approximates implicit learning. In addition, we provide evidence for a subdivision of the slow/implicit process into distinct manifestations of motor memory. We conclude that the two-state model of motor learning is a close approximation of sensorimotor learning, but it is unable to describe adequately the various implicit learning operations that forge the learning curve. Our results suggest that a wider net be cast in the search for the putative psychological mechanisms and neural substrates underlying the multiplicity of processes involved in motor learning.
Computations underlying cognitive strategies in human motor learning are poorly understood. Here we investigate such strategies in a common sensorimotor transformation task. We show that strategies assume two forms, likely reflecting distinct working memory representations: discrete caching of stimulus-response contingencies, and time-consuming parametric computations. Reaction times and errors suggest that both strategies are employed during learning, and trade off based on task complexity. Experiments using pressured preparation time further support dissociable strategies: In response caching, time pressure elicits multi-modal distributions of movements; during parametric computations, time pressure elicits a shifting distribution of movements between visual targets and distal goals, consistent with analog re-computing of a movement plan. A generalization experiment reveals that discrete and parametric strategies produce, respectively, more localized or more global transfer effects. These results describe how qualitatively distinct cognitive representations are leveraged for motor learning and produce downstream consequences for behavioral flexibility.
Sensorimotor adaptation tasks have been used to characterize processes responsible for calibrating the mapping between desired outcomes and motor commands. Research has focused on how this form of error-based learning occurs in an implicit and automatic manner. However, recent work has revealed the operation of multiple learning processes, even in this simple form of learning. This review focuses on the contribution of cognitive strategies and heuristics to sensorimotor learning, and how these processes enable humans to rapidly explore and evaluate novel solutions to enable flexible, goal-oriented behavior. This new work points to limitations in current computational models, and how these must be updated to describe the conjoint impact of multiple processes in sensorimotor learning.
Generalization is a fundamental aspect of behavior, allowing for the transfer of knowledge from one context to another. The details of this transfer are thought to reveal how the brain represents what it learns. Generalization has been a central focus in studies of sensorimotor adaptation, and its pattern has been well characterized: Learning of new dynamic and kinematic transformations in one region of space tapers off in a Gaussian-like fashion to neighboring untrained regions, echoing tuned population codes in the brain. In contrast to common allusions to generalization in cognitive science, generalization in visually guided reaching is usually framed as a passive consequence of neural tuning functions rather than a cognitive feature of learning. While previous research has presumed that maximum generalization occurs at the instructed task goal or the actual movement direction, recent work suggests that maximum generalization may occur at the location of an explicitly accessible movement plan. Here we provide further support for plan-based generalization, formalize this theory in an updated model of adaptation, and test several unexpected implications of the model. First, we employ a generalization paradigm to parameterize the generalization function and ascertain its maximum point. We then apply the derived generalization function to our model and successfully simulate and fit the time course of implicit adaptation across three behavioral experiments. We find that dynamics predicted by plan-based generalization are borne out in the data, are contrary to what traditional models predict, and lead to surprising implications for the behavioral, computational, and neural characteristics of sensorimotor adaptation. The pattern of generalization is thought to reveal how the motor system represents learned actions. Recent work has made the intriguing suggestion that maximum generalization in sensorimotor adaptation tasks occurs at the location of the learned movement plan. Here we support this interpretation, develop a novel model of motor adaptation that incorporates plan-based generalization, and use the model to successfully predict surprising dynamics in the time course of adaptation across several conditions.
When a person fails to obtain an expected reward from an object in the environment, they face a credit assignment problem: Did the absence of reward reflect an extrinsic property of the environment or an intrinsic error in motor execution? To explore this problem, we modified a popular decision-making task used in studies of reinforcement learning, the two-armed bandit task. We compared a version in which choices were indicated by key presses, the standard response in such tasks, to a version in which the choices were indicated by reaching movements, which affords execution failures. In the key press condition, participants exhibited a strong risk aversion bias; strikingly, this bias reversed in the reaching condition. This result can be explained by a reinforcement model wherein movement errors influence decision-making, either by gating reward prediction errors or by modifying an implicit representation of motor competence. Two further experiments support the gating hypothesis. First, we used a condition in which we provided visual cues indicative of movement errors but informed the participants that trial outcomes were independent of their actual movements. The main result was replicated, indicating that the gating process is independent of participants' explicit sense of control. Second, individuals with cerebellar degeneration failed to modulate their behavior between the key press and reach conditions, providing converging evidence of an implicit influence of movement error signals on reinforcement learning. These results provide a mechanistically tractable solution to the credit assignment problem.decision-making | reinforcement learning | sensory prediction error | reward prediction error | cerebellum W hen a diner reaches across the table and knocks over her coffee, the absence of anticipated reward should be attributed to a failure of coordination rather than diminish her love of coffee. Although this attribution is intuitive, current models of decision-making lack a mechanistic explanation for this seemingly simple computation. We set out to ask if, and how, selection processes in decision-making incorporate information specific to action execution and thus solve the credit assignment problem that arises when an expected reward is not obtained because of a failure in motor execution.Humans are highly capable of tracking the value of stimuli, varying their behavior on the basis of reinforcement history (1, 2), and exhibiting sensitivity to intrinsic motor noise when reward outcomes depend on movement accuracy (3-5). In real-world behavior, the underlying cause of unrewarded events is often ambiguous: A lost point in tennis could occur because the player made a poor choice about where to hit the ball or failed to properly execute the stroke. However, in laboratory studies of reinforcement learning, the underlying cause of unrewarded events is typically unambiguous, either solely dependent on properties of the stimulus or on motor noise. Thus, it remains unclear how people assign credit to either extrins...
To survive, animals must learn to control their movements with millisecond-level precision, and adjust the kinematics if conditions, or task requirements, change. Here, we examine adaptive timing of motor output in mice, using a simple eyelid conditioning task. Mice were trained to blink in response to a light stimulus that was always followed by a corneal air-puff at a constant time interval. Different mice were trained with different intervals of time separating the onset of the light and the air-puff. As in previous work in other animal species, mice learned to control the speed of the blink, such that the time of maximum eyelid closure matched the interval used during training. However, we found that the time of maximum eyelid speed was always in the first 100 ms after movement onset and did not scale with the training interval, indicating that adaptive timing is not accomplished by slowing down (or speeding up) the eyelid movement uniformly throughout the duration of the blink. A new analysis, specifically designed to examine the kinematics of blinks in single trials, revealed that the underlying control signal responsible for the eyelid movement is made up of oscillatory bursts that are time-locked to the light stimulus at the beginning of the blink, becoming desynchronized later on. Furthermore, mice learn to blink at different speeds and time the movement appropriately by adjusting the amplitude, but not the frequency of the bursts in the eyelid oscillation.
One persistent curiosity in visuomotor adaptation tasks is that participants often do not reach maximal performance. This incomplete asymptote has been explained as a consequence of obligatory computations within the implicit adaptation system, such as an equilibrium between learning and forgetting. A body of recent work has shown that in standard adaptation tasks, cognitive strategies operate alongside implicit learning. We reasoned that incomplete learning in adaptation tasks may primarily reflect a speed-accuracy tradeoff on time-consuming motor planning. Across three experiments, we find evidence supporting this hypothesis, showing that hastened motor planning may primarily lead to under-compensation. When an obligatory waiting period was administered before movement start, participants were able to fully counteract imposed perturbations (Experiment 1). Inserting the same delay between trials – rather than during movement planning – did not induce full compensation, suggesting that the motor planning interval influences the learning asymptote (Experiment 2). In the last experiment (Experiment 3), we asked participants to continuously report their movement intent. We show that emphasizing explicit re-aiming strategies (and concomitantly increasing planning time) also lead to complete asymptotic learning. Findings from all experiments support the hypothesis that incomplete adaptation is, in part, the result of an intrinsic speed-accuracy tradeoff, perhaps related to cognitive strategies that require parametric attentional reorienting from the visual target to the goal.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.