skip to main content

Attention:

The NSF Public Access Repository (PAR) system and access will be unavailable from 11:00 PM ET on Thursday, February 13 until 2:00 AM ET on Friday, February 14 due to maintenance. We apologize for the inconvenience.


Title: Multiple traces and altered signal-to-noise in systems consolidation: Evidence from the 7T fMRI Natural Scenes Dataset
The brain mechanisms of memory consolidation remain elusive. Here, we examine blood-oxygen-level-dependent (BOLD) correlates of image recognition through the scope of multiple influential systems consolidation theories. We utilize the longitudinal Natural Scenes Dataset, a 7-Tesla functional magnetic resonance imaging human study in which ∼135,000 trials of image recognition were conducted over the span of a year among eight subjects. We find that early- and late-stage image recognition associates with both medial temporal lobe (MTL) and visual cortex when evaluating regional activations and a multivariate classifier. Supporting multiple-trace theory (MTT), parts of the MTL activation time course show remarkable fit to a 20-y-old MTT time-dynamical model predicting early trace intensity increases and slight subsequent interference ( R 2 > 0.90). These findings contrast a simplistic, yet common, view that memory traces are transferred from MTL to cortex. Next, we test the hypothesis that the MTL trace signature of memory consolidation should also reflect synaptic “desaturation,” as evidenced by an increased signal-to-noise ratio. We find that the magnitude of relative BOLD enhancement among surviving memories is positively linked to the rate of removal (i.e., forgetting) of competing traces. Moreover, an image-feature and time interaction of MTL and visual cortex functional connectivity suggests that consolidation mechanisms improve the specificity of a distributed trace. These neurobiological effects do not replicate on a shorter timescale (within a session), implicating a prolonged, offline process. While recognition can potentially involve cognitive processes outside of memory retrieval (e.g., re-encoding), our work largely favors MTT and desaturation as perhaps complementary consolidative memory mechanisms.  more » « less
Award ID(s):
1822929 1822683
PAR ID:
10390004
Author(s) / Creator(s):
; ; ; ; ; ; ;
Date Published:
Journal Name:
Proceedings of the National Academy of Sciences
Volume:
119
Issue:
44
ISSN:
0027-8424
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. The human medial temporal lobe (MTL) plays a crucial role in recognizing visual objects, a key cognitive function that relies on the formation of semantic representations. Nonetheless, it remains unknown how visual information of general objects is translated into semantic representations in the MTL. Furthermore, the debate about whether the human MTL is involved in perception has endured for a long time. To address these questions, we investigated three distinct models of neural object coding—semantic coding, axis-based feature coding, and region-based feature coding—in each subregion of the MTL, using high-resolution fMRI in two male and six female participants. Our findings revealed the presence of semantic coding throughout the MTL, with a higher prevalence observed in the parahippocampal cortex (PHC) and perirhinal cortex (PRC), while axis coding and region coding were primarily observed in the earlier regions of the MTL. Moreover, we demonstrated that voxels exhibiting axis coding supported the transition to region coding and contained information relevant to semantic coding. Together, by providing a detailed characterization of neural object coding schemes and offering a comprehensive summary of visual coding information for each MTL subregion, our results not only emphasize a clear role of the MTL in perceptual processing but also shed light on the translation of perception-driven representations of visual features into memory-driven representations of semantics along the MTL processing pathway.

    Significance StatementIn this study, we delved into the mechanisms underlying visual object recognition within the human medial temporal lobe (MTL), a pivotal region known for its role in the formation of semantic representations crucial for memory. In particular, the translation of visual information into semantic representations within the MTL has remained unclear, and the enduring debate regarding the involvement of the human MTL in perception has persisted. To address these questions, we comprehensively examined distinct neural object coding models across each subregion of the MTL, leveraging high-resolution fMRI. We also showed transition of information between object coding models and across MTL subregions. Our findings significantly contributes to advancing our understanding of the intricate pathway involved in visual object coding.

     
    more » « less
  2. Abstract

    According to a “Swiss Army Knife” model of the brain, cognitive functions such as episodic memory and face perception map onto distinct neural substrates. In contrast, representational accounts propose that each brain region is best explained not by which specialized function it performs, but by the type of information it represents with its neural firing. In a functional magnetic resonance imaging study, we asked whether the neural signals supporting recognition memory fall mandatorily within the medial temporal lobes (MTL), traditionally thought the seat of declarative memory, or whether these signals shift within cortex according to the content of the memory. Participants studied objects and scenes that were unique conjunctions of pre-defined visual features. Next, we tested recognition memory in a task that required mnemonic discrimination of both simple features and complex conjunctions. Feature memory signals were strongest in posterior visual regions, declining with anterior progression toward the MTL, while conjunction memory signals followed the opposite pattern. Moreover, feature memory signals correlated with feature memory discrimination performance most strongly in posterior visual regions, whereas conjunction memory signals correlated with conjunction memory discrimination most strongly in anterior sites. Thus, recognition memory signals shifted with changes in memory content, in line with representational accounts.

     
    more » « less
  3. Neuroimaging studies of human memory have consistently found that univariate responses in parietal cortex track episodic experience with stimuli (whether stimuli are 'old' or 'new'). More recently, pattern-based fMRI studies have shown that parietal cortex also carries information about the semantic content of remembered experiences. However, it is not well understood how memory-based and content-based signals are integrated within parietal cortex. Here, in humans (males and females), we used voxel-wise encoding models and a recognition memory task to predict the fMRI activity patterns evoked by complex natural scene images based on (1) the episodic history and (2) the semantic content of each image. Models were generated and compared across distinct subregions of parietal cortex and for occipitotemporal cortex. We show that parietal and occipitotemporal regions each encode memory and content information, but they differ in how they combine this information. Among parietal subregions, angular gyrus was characterized by robust and overlapping effects of memory and content. Moreover, subject-specific semantic tuning functions revealed that successful recognition shifted the amplitude of tuning functions in angular gyrus but did not change the selectivity of tuning. In other words, effects of memory and content were additive in angular gyrus. This pattern of data contrasted with occipitotemporal cortex where memory and content effects were interactive: memory effects were preferentially expressed by voxels tuned to the content of a remembered image. Collectively, these findings provide unique insight into how parietal cortex combines information about episodic memory and semantic content.

    SIGNIFICANCE STATEMENTNeuroimaging studies of human memory have identified multiple brain regions that not only carry information about “whether” a visual stimulus is successfully recognized but also “what” the content of that stimulus includes. However, a fundamental and open question concerns how the brain integrates these two types of information (memory and content). Here, using a powerful combination of fMRI analysis methods, we show that parietal cortex, particularly the angular gyrus, robustly combines memory- and content-related information, but these two forms of information are represented via additive, independent signals. In contrast, memory effects in high-level visual cortex critically depend on (and interact with) content representations. Together, these findings reveal multiple and distinct ways in which the brain combines memory- and content-related information.

     
    more » « less
  4. Memories of fearful events can last a lifetime. The prelimbic (PL) cortex, a subregion of prefrontal cortex, plays a critical role in fear memory retrieval over time. Most studies have focused on acquisition, consolidation, and retrieval of recent memories, but much less is known about the neural mechanisms of remote memory. Using a new knock-in mouse for activity-dependent genetic labeling (TRAP2), we demonstrate that neuronal ensembles in the PL cortex are dynamic. PL neurons TRAPed during later memory retrievals are more likely to be reactivated and make larger behavioral contributions to remote memory retrieval compared to those TRAPed during learning or early memory retrieval. PL activity during learning is required to initiate this time-dependent reorganization in PL ensembles underlying memory retrieval. Finally, while neurons TRAPed during earlier and later retrievals have similar broad projections throughout the brain, PL neurons TRAPed later have a stronger functional recruitment of cortical targets. 
    more » « less
  5. To fluidly engage with the world, our brains must simultaneously represent both the scene in front of us and our memory of the immediate surrounding environment (i.e., local visuospatial context). How does the brain's functional architecture enable sensory and mnemonic representations to closely interface while also avoiding sensory-mnemonic interference? Here, we asked this question using first-person, head-mounted virtual reality and fMRI. Using virtual reality, human participants of both sexes learned a set of immersive, real-world visuospatial environments in which we systematically manipulated the extent of visuospatial context associated with a scene image in memory across three learning conditions, spanning from a single FOV to a city street. We used individualized, within-subject fMRI to determine which brain areas support memory of the visuospatial context associated with a scene during recall (Experiment 1) and recognition (Experiment 2). Across the whole brain, activity in three patches of cortex was modulated by the amount of known visuospatial context, each located immediately anterior to one of the three scene perception areas of high-level visual cortex. Individual subject analyses revealed that these anterior patches corresponded to three functionally defined place memory areas, which selectively respond when visually recalling personally familiar places. In addition to showing activity levels that were modulated by the amount of visuospatial context, multivariate analyses showed that these anterior areas represented the identity of the specific environment being recalled. Together, these results suggest a convergence zone for scene perception and memory of the local visuospatial context at the anterior edge of high-level visual cortex.

    SIGNIFICANCE STATEMENTAs we move through the world, the visual scene around us is integrated with our memory of the wider visuospatial context. Here, we sought to understand how the functional architecture of the brain enables coexisting representations of the current visual scene and memory of the surrounding environment. Using a combination of immersive virtual reality and fMRI, we show that memory of visuospatial context outside the current FOV is represented in a distinct set of brain areas immediately anterior and adjacent to the perceptually oriented scene-selective areas of high-level visual cortex. This functional architecture would allow efficient interaction between immediately adjacent mnemonic and perceptual areas while also minimizing interference between mnemonic and perceptual representations.

     
    more » « less