skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Inter-individual variability in motor learning due to differences in effective learning rates between generalist and specialist memory stores
Abstract Humans exhibit large interindividual differences in motor learning ability. However, most previous studies have examined properties common across populations, with less emphasis on interindividual differences. We hypothesized here, based on our previous experimental and computational motor adaptation studies, that individual differences in effective learning rates between a generalist memory module that assumes environmental continuity and specialist modules that are responsive to trial-by-trial environmental changes could explain both large population-wise and individual-wise differences in dual tasks adaptation under block and random schedules. Participants adapted to two opposing force fields, either sequentially with alternating training blocks or simultaneously with random sequences. As previously reported, in the block training schedule, all participants adapted to the force field presented in a block but showed large interference in the subsequent opposing force field blocks, such that adapting to the two force fields was impossible. In contrast, in the random training schedule, participants could adapt to the two conflicting tasks simultaneously as a group; however, large interindividual variability was observed. A modified MOSAIC computational model of motor learning equipped with one generalist module and two specialist modules explained the observed behavior and variability for wide parameter ranges: when the predictions errors were large and consistent as in block schedules, the generalist module was selected to adapt quickly. In contrast, the specialist modules were selected when they more accurately predicted the changing environment than the generalist, as during random schedules; this resulted in consolidated memory specialized to each environment, but only when the ratio of learning rates of the generalist to specialists was relatively small. This dynamic selection process plays a crucial role in explaining the individual differences observed in motor learning abilities.  more » « less
Award ID(s):
2216344
PAR ID:
10571144
Author(s) / Creator(s):
; ; ; ;
Publisher / Repository:
bioRxiv
Date Published:
Format(s):
Medium: X
Institution:
bioRxiv
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract Previous research has demonstrated significant inter-individual variability in the recruitment of the fast-explicit and slow-implicit processes during motor adaptation. In addition, we previously identified qualitative individual differences in adaptation linked to the formation and updating of new memory processes. Here, we investigated quantitative and qualitative differences in visuomotor adaptation with a design incorporating repeated learning and forgetting blocks, allowing for precise estimation of individual learning and forgetting rates in fast-slow adaptation models. Participants engaged in a two-day online visuomotor adaptation task. They first adapted to a 30-degree perturbation to eight targets in three blocks separated by short blocks of no feedback trials. Approximately 24 hours later, they performed a no-feedback retention block and a relearning block. We clustered the participants into strong and weak learners based on adaptation levels at the end of day one and fitted a fast-slow system to the adaptation data. Strong learners exhibited a strong negative correlation between the estimated slow and fast processes, which predicted 24-hour retention and savings, respectively, supporting the engagement of a fast-slow system. The pronounced individual differences in the recruitment of the two processes were attributed to wide ranges of estimated learning rates. Conversely, weak learners exhibited a positive correlation between the two estimated processes, as well as retention but no savings, supporting the engagement of a single slow system. Finally, both during baseline and adaptation, reaction times were shorter for weak learners. Our findings thus revealed two distinct ways to learn in visuomotor adaptation and highlight the necessity of considering both quantitative and qualitative individual differences in studies of motor learning. 
    more » « less
  2. This paper presents a novel zero-shot learning approach towards personalized speech enhancement through the use of a sparsely active ensemble model. Optimizing speech denoising systems towards a particular test-time speaker can improve performance and reduce run-time complexity. However, test-time model adaptation may be challenging if collecting data from the test-time speaker is not possible. To this end, we propose using an ensemble model wherein each specialist module denoises noisy utterances from a distinct partition of training set speakers. The gating module inexpensively estimates test-time speaker characteristics in the form of an embedding vector and selects the most appropriate specialist module for denoising the test signal. Grouping the training set speakers into non-overlapping semantically similar groups is non-trivial and ill-defined. To do this, we first train a Siamese network using noisy speech pairs to maximize or minimize the similarity of its output vectors depending on whether the utterances derive from the same speaker or not. Next, we perform k-means clustering on the latent space formed by the averaged embedding vectors per training set speaker. In this way, we designate speaker groups and train specialist modules optimized around partitions of the complete training set. Our experiments show that ensemble models made up of low-capacity specialists can outperform high-capacity generalist models with greater efficiency and improved adaptation towards unseen test-time speakers. 
    more » « less
  3. The human sensorimotor system can adapt to various changes in the environmental dynamics by updating motor commands to improve performance after repeated exposure to the same task. However, the characteristics and mechanisms of the adaptation process remain unknown for dexterous manipulation, a unique motor task in which the body physically interacts with the environment with multiple effectors, i.e., digits, in parallel. We addressed this gap by using robotic manipulanda to investigate the changes in the digit force coordination following mechanical perturbation of an object held by tripod grasps. As the participants gradually adapted to lifting the object under perturbations, we quantified two components of digit force coordination. One is the direction-specific manipulation moment that directly counteracts the perturbation, whereas the other one is the direction-independent internal moment that supports the stability and stiffness of the grasp. We found that trial-to-trial improvement of task performance was associated with increased manipulation moment and a gradual decrease of the internal moment. These two moments were characterized by different rates of adaptation. We also examined how these two force coordination components respond to changes in perturbation directions. Importantly, we found that the manipulation moment was sensitive to the extent of repetitive exposure to the previous context that has an opposite perturbation direction, whereas the internal moment did not. However, the internal moment was sensitive to whether the postchange perturbation direction was previously experienced. Our results reveal, for the first time, that two distinct processes underlie the adaptation of multidigit force coordination for dexterous manipulation. NEW & NOTEWORTHY Changes in digit force coordination in multidigit object manipulation were quantified with a novel experimental design in which human participants adapted to mechanical perturbations applied to the object. Our results show that the adaptation of digit force coordination can be characterized by two distinct components that operate at different timescales. We further show that these two components respond to changes in perturbation direction differently. 
    more » « less
  4. Ko, Hanseok (Ed.)
    Backpropagation (BP) has been a successful optimization technique for deep learning models. However, its limitations, such as backward- and update-locking, and its biological implausibility, hinder the concurrent updating of layers and do not mimic the local learning processes observed in the human brain. To address these issues, recent research has suggested using local error signals to asynchronously train network blocks. However, this approach often involves extensive trial-and-error iterations to determine the best configuration for local training. This includes decisions on how to decouple network blocks and which auxiliary networks to use for each block. In our work, we introduce a novel BP-free approach: a block-wise BP-free (BWBPF) neural network that leverages local error signals to optimize distinct sub-neural networks separately, where the global loss is only responsible for updating the output layer. The local error signals used in the BP-free model can be computed in parallel, enabling a potential speed-up in the weight update process through parallel implementation. Our experimental results consistently show that this approach can identify transferable decoupled architectures for VGG and ResNet variations, outperforming models trained with end-to-end backpropagation and other state-of-the-art block-wise learning techniques on datasets such as CIFAR-10 and Tiny-ImageNet. The code is released at https://github.com/Belis0811/BWBPF. 
    more » « less
  5. Large-scale machine learning (ML) models are increasingly being used in critical domains like education, lending, recruitment, healthcare, criminal justice, etc. However, the training, deployment, and utilization of these models demand substantial computational resources. To decrease computation and memory costs, machine learning models with sparse weight matrices are widely used in the literature. Among sparse models, those with special sparse structures (e.g., models with block-wise sparse weight matrices) fit better with the hardware accelerators and can decrease the memory and computation costs during the inference. Unfortunately, while there are several efficient training methods, none of them are designed to train a block-wise sparse model efficiently. As a result, the current methods for training block-wise sparse models start with full and dense models leading to inefficient training. In this work, we focus on training models with block-wise sparse matrices and propose an efficient training algorithm to decrease both computation and memory costs during training and inference. In addition, we will show that our proposed method enables us to efficiently find the right block size for the sparsity pattern during the training process. Our extensive empirical and theoretical analyses show that our algorithms can decrease the computation and memory costs significantly without a performance drop compared to baselines. 
    more » « less