1
|
Forming cognitive maps for abstract spaces: the roles of the human hippocampus and orbitofrontal cortex. Commun Biol 2024; 7:517. [PMID: 38693344 PMCID: PMC11063219 DOI: 10.1038/s42003-024-06214-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2023] [Accepted: 04/18/2024] [Indexed: 05/03/2024] Open
Abstract
How does the human brain construct cognitive maps for decision-making and inference? Here, we conduct an fMRI study on a navigation task in multidimensional abstract spaces. Using a deep neural network model, we assess learning levels and categorized paths into exploration and exploitation stages. Univariate analyses show higher activation in the bilateral hippocampus and lateral prefrontal cortex during exploration, positively associated with learning level and response accuracy. Conversely, the bilateral orbitofrontal cortex (OFC) and retrosplenial cortex show higher activation during exploitation, negatively associated with learning level and response accuracy. Representational similarity analysis show that the hippocampus, entorhinal cortex, and OFC more accurately represent destinations in exploitation than exploration stages. These findings highlight the collaboration between the medial temporal lobe and prefrontal cortex in learning abstract space structures. The hippocampus may be involved in spatial memory formation and representation, while the OFC integrates sensory information for decision-making in multidimensional abstract spaces.
Collapse
|
2
|
An analysis of information segregation in parallel streams of a multi-stream convolutional neural network. Sci Rep 2024; 14:9097. [PMID: 38643326 PMCID: PMC11032341 DOI: 10.1038/s41598-024-59930-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 04/16/2024] [Indexed: 04/22/2024] Open
Abstract
Visual information is processed in hierarchically organized parallel streams in the primate brain. In the present study, information segregation in parallel streams was examined by constructing a convolutional neural network with parallel architecture in all of the convolutional layers. Although filter weights for convolution were initially set to random values, color information was segregated from shape information in most model instances after training. Deletion of the color-related stream decreased recognition accuracy of animate images, whereas deletion of the shape-related stream decreased recognition accuracy of both animate and inanimate images. The results suggest that properties of filters and functions of a stream are spontaneously segregated in parallel streams of neural networks.
Collapse
|
3
|
Assessing Spontaneous Categorical Processing of Visual Shapes via Frequency-Tagging EEG. J Neurosci 2024; 44:e1346232024. [PMID: 38423762 PMCID: PMC11026363 DOI: 10.1523/jneurosci.1346-23.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 01/25/2024] [Accepted: 02/20/2024] [Indexed: 03/02/2024] Open
Abstract
Categorization is an essential cognitive and perceptual process, which happens spontaneously. However, earlier research often neglected the spontaneous nature of this process by mainly adopting explicit tasks in behavioral or neuroimaging paradigms. Here, we use frequency-tagging (FT) during electroencephalography (EEG) in 22 healthy human participants (both male and female) as a direct approach to pinpoint spontaneous visual categorical processing. Starting from schematic natural visual stimuli, we created morph sequences comprising 11 equal steps. Mirroring a behavioral categorical perception discrimination paradigm, we administered a FT-EEG oddball paradigm, assessing neural sensitivity for equally sized differences within and between stimulus categories. Likewise, mirroring a behavioral category classification paradigm, we administered a sweep FT-EEG oddball paradigm, sweeping from one end of the morph sequence to the other, thereby allowing us to objectively pinpoint the neural category boundary. We found that FT-EEG can implicitly measure categorical processing and discrimination. More specifically, we could derive an objective neural index of the required level to differentiate between the two categories, and this neural index showed the typical marker of categorical perception (i.e., stronger discrimination across as compared with within categories). The neural findings of the implicit paradigms were also validated using an explicit behavioral task. These results provide evidence that FT-EEG can be used as an objective tool to measure discrimination and categorization and that the human brain inherently and spontaneously (without any conscious or decisional processes) uses higher-level meaningful categorization information to interpret ambiguous (morph) shapes.
Collapse
|
4
|
Use of functional magnetic resonance imaging to identify cortical loci for lower limb movements and their efficacy for individuals after stroke. J Neuroeng Rehabil 2024; 21:58. [PMID: 38627779 PMCID: PMC11020805 DOI: 10.1186/s12984-024-01319-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Accepted: 01/29/2024] [Indexed: 04/19/2024] Open
Abstract
BACKGROUND Identification of cortical loci for lower limb movements for stroke rehabilitation is crucial for better rehabilitation outcomes via noninvasive brain stimulation by targeting the fine-grained cortical loci of the movements. However, identification of the cortical loci for lower limb movements using functional MRI (fMRI) is challenging due to head motion and difficulty in isolating different types of movement. Therefore, we developed a custom-made MR-compatible footplate and leg cushion to identify the cortical loci for lower limb movements and conducted multivariate analysis on the fMRI data. We evaluated the validity of the identified loci using both fMRI and behavioral data, obtained from healthy participants as well as individuals after stroke. METHODS We recruited 33 healthy participants who performed four different lower limb movements (ankle dorsiflexion, ankle rotation, knee extension, and toe flexion) using our custom-built equipment while fMRI data were acquired. A subgroup of these participants (Dataset 1; n = 21) was used to identify the cortical loci associated with each lower limb movement in the paracentral lobule (PCL) using multivoxel pattern analysis and representational similarity analysis. The identified cortical loci were then evaluated using the remaining healthy participants (Dataset 2; n = 11), for whom the laterality index (LI) was calculated for each lower limb movement using the cortical loci identified for the left and right lower limbs. In addition, we acquired a dataset from 15 individuals with chronic stroke for regression analysis using the LI and the Fugl-Meyer Assessment (FMA) scale. RESULTS The cortical loci associated with the lower limb movements were hierarchically organized in the medial wall of the PCL following the cortical homunculus. The LI was clearer using the identified cortical loci than using the PCL. The healthy participants (mean ± standard deviation: 0.12 ± 0.30; range: - 0.63 to 0.91) exhibited a higher contralateral LI than the individuals after stroke (0.07 ± 0.47; - 0.83 to 0.97). The corresponding LI scores for individuals after stroke showed a significant positive correlation with the FMA scale for paretic side movement in ankle dorsiflexion (R2 = 0.33, p = 0.025) and toe flexion (R2 = 0.37, p = 0.016). CONCLUSIONS The cortical loci associated with lower limb movements in the PCL identified in healthy participants were validated using independent groups of healthy participants and individuals after stroke. Our findings suggest that these cortical loci may be beneficial for the neurorehabilitation of lower limb movement in individuals after stroke, such as in developing effective rehabilitation interventions guided by the LI scores obtained for neuronal activations calculated from the identified cortical loci across the paretic and non-paretic sides of the brain.
Collapse
|
5
|
Similarity in evoked responses does not imply similarity in macroscopic network states. Netw Neurosci 2024; 8:335-354. [PMID: 38711543 PMCID: PMC11073549 DOI: 10.1162/netn_a_00354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Accepted: 11/17/2023] [Indexed: 05/08/2024] Open
Abstract
It is commonplace in neuroscience to assume that if two tasks activate the same brain areas in the same way, then they are recruiting the same underlying networks. Yet computational theory has shown that the same pattern of activity can emerge from many different underlying network representations. Here we evaluated whether similarity in activation necessarily implies similarity in network architecture by comparing region-wise activation patterns and functional correlation profiles from a large sample of healthy subjects (N = 242). Participants performed two executive control tasks known to recruit nearly identical brain areas, the color-word Stroop task and the Multi-Source Interference Task (MSIT). Using a measure of instantaneous functional correlations, based on edge time series, we estimated the task-related networks that differed between incongruent and congruent conditions. We found that the two tasks were much more different in their network profiles than in their evoked activity patterns at different analytical levels, as well as for a wide range of methodological pipelines. Our results reject the notion that having the same activation patterns means two tasks engage the same underlying representations, suggesting that task representations should be independently evaluated at both node and edge (connectivity) levels.
Collapse
|
6
|
Representational formats of human memory traces. Brain Struct Funct 2024; 229:513-529. [PMID: 37022435 PMCID: PMC10978732 DOI: 10.1007/s00429-023-02636-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2022] [Accepted: 03/28/2023] [Indexed: 04/07/2023]
Abstract
Neural representations are internal brain states that constitute the brain's model of the external world or some of its features. In the presence of sensory input, a representation may reflect various properties of this input. When perceptual information is no longer available, the brain can still activate representations of previously experienced episodes due to the formation of memory traces. In this review, we aim at characterizing the nature of neural memory representations and how they can be assessed with cognitive neuroscience methods, mainly focusing on neuroimaging. We discuss how multivariate analysis techniques such as representational similarity analysis (RSA) and deep neural networks (DNNs) can be leveraged to gain insights into the structure of neural representations and their different representational formats. We provide several examples of recent studies which demonstrate that we are able to not only measure memory representations using RSA but are also able to investigate their multiple formats using DNNs. We demonstrate that in addition to slow generalization during consolidation, memory representations are subject to semantization already during short-term memory, by revealing a shift from visual to semantic format. In addition to perceptual and conceptual formats, we describe the impact of affective evaluations as an additional dimension of episodic memories. Overall, these studies illustrate how the analysis of neural representations may help us gain a deeper understanding of the nature of human memory.
Collapse
|
7
|
Alignment of brain embeddings and artificial contextual embeddings in natural language points to common geometric patterns. Nat Commun 2024; 15:2768. [PMID: 38553456 PMCID: PMC10980748 DOI: 10.1038/s41467-024-46631-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Accepted: 03/04/2024] [Indexed: 04/02/2024] Open
Abstract
Contextual embeddings, derived from deep language models (DLMs), provide a continuous vectorial representation of language. This embedding space differs fundamentally from the symbolic representations posited by traditional psycholinguistics. We hypothesize that language areas in the human brain, similar to DLMs, rely on a continuous embedding space to represent language. To test this hypothesis, we densely record the neural activity patterns in the inferior frontal gyrus (IFG) of three participants using dense intracranial arrays while they listened to a 30-minute podcast. From these fine-grained spatiotemporal neural recordings, we derive a continuous vectorial representation for each word (i.e., a brain embedding) in each patient. Using stringent zero-shot mapping we demonstrate that brain embeddings in the IFG and the DLM contextual embedding space have common geometric patterns. The common geometric patterns allow us to predict the brain embedding in IFG of a given left-out word based solely on its geometrical relationship to other non-overlapping words in the podcast. Furthermore, we show that contextual embeddings capture the geometry of IFG embeddings better than static word embeddings. The continuous brain embedding space exposes a vector-based neural code for natural language processing in the human brain.
Collapse
|
8
|
Representational maps in the brain: concepts, approaches, and applications. Front Cell Neurosci 2024; 18:1366200. [PMID: 38584779 PMCID: PMC10995314 DOI: 10.3389/fncel.2024.1366200] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Accepted: 03/08/2024] [Indexed: 04/09/2024] Open
Abstract
Neural systems have evolved to process sensory stimuli in a way that allows for efficient and adaptive behavior in a complex environment. Recent technological advances enable us to investigate sensory processing in animal models by simultaneously recording the activity of large populations of neurons with single-cell resolution, yielding high-dimensional datasets. In this review, we discuss concepts and approaches for assessing the population-level representation of sensory stimuli in the form of a representational map. In such a map, not only are the identities of stimuli distinctly represented, but their relational similarity is also mapped onto the space of neuronal activity. We highlight example studies in which the structure of representational maps in the brain are estimated from recordings in humans as well as animals and compare their methodological approaches. Finally, we integrate these aspects and provide an outlook for how the concept of representational maps could be applied to various fields in basic and clinical neuroscience.
Collapse
|
9
|
Human EEG and artificial neural networks reveal disentangled representations of object real-world size in natural images. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.08.19.553999. [PMID: 37662197 PMCID: PMC10473678 DOI: 10.1101/2023.08.19.553999] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/05/2023]
Abstract
Remarkably, human brains have the ability to accurately perceive and process the real-world size of objects, despite vast differences in distance and perspective. While previous studies have delved into this phenomenon, distinguishing this ability from other visual perceptions, like depth, has been challenging. Using the THINGS EEG2 dataset with high time-resolution human brain recordings and more ecologically valid naturalistic stimuli, our study uses an innovative approach to disentangle neural representations of object real-world size from retinal size and perceived real-world depth in a way that was not previously possible. Leveraging this state-of-the-art dataset, our EEG representational similarity results reveal a pure representation of object real-world size in human brains. We report a representational timeline of visual object processing: object real-world depth appeared first, then retinal size, and finally, real-world size. Additionally, we input both these naturalistic images and object-only images without natural background into artificial neural networks. Consistent with the human EEG findings, we also successfully disentangled representation of object real-world size from retinal size and real-world depth in all three types of artificial neural networks (visual-only ResNet, visual-language CLIP, and language-only Word2Vec). Moreover, our multi-modal representational comparison framework across human EEG and artificial neural networks reveals real-world size as a stable and higher-level dimension in object space incorporating both visual and semantic information. Our research provides a detailed and clear characterization of the object processing process, which offers further advances and insights into our understanding of object space and the construction of more brain-like visual models.
Collapse
|
10
|
Emotional state dynamics impacts temporal memory. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.07.25.550412. [PMID: 38464043 PMCID: PMC10925226 DOI: 10.1101/2023.07.25.550412] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/12/2024]
Abstract
Emotional fluctuations are ubiquitous in everyday life, but precisely how they sculpt the temporal organization of memories remains unclear. Here, we designed a novel task-the Emotion Boundary Task-wherein participants viewed sequences of negative and neutral images surrounded by a color border. We manipulated perceptual context (border color), emotional valence, as well as the direction of emotional-valence shifts (i.e., shifts from neutral-to-negative and negative-to-neutral events) to create encoding events comprised of image sequences with a shared perceptual and/or emotional context. We measured memory for temporal order and subjectively remembered temporal distances for images processed within and across events. Negative images processed within events were remembered as closer in time compared to neutral ones. In contrast, temporal distance was remembered as longer for images spanning neutral-to-negative shifts-suggesting temporal dilation in memory with the onset of a negative event following a previously-neutral state. The extent of this negative-picture induced temporal dilation in memory correlated with dispositional negativity across individuals. Lastly, temporal order memory was enhanced for recently presented negative (compared to neutral) images. These findings suggest that emotional-state dynamics matters when considering emotion-temporal memory interactions: While persistent negative events may compress subjectively remembered time, dynamic shifts from neutral to negative events produce temporal dilation in memory, which may be relevant for adaptive emotional functioning.
Collapse
|
11
|
Distributed representations of behaviorally-relevant object dimensions in the human visual system. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.08.23.553812. [PMID: 37662312 PMCID: PMC10473665 DOI: 10.1101/2023.08.23.553812] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/05/2023]
Abstract
Object vision is commonly thought to involve a hierarchy of brain regions processing increasingly complex image features, with high-level visual cortex supporting object recognition and categorization. However, object vision supports diverse behavioral goals, suggesting basic limitations of this category-centric framework. To address these limitations, we mapped a series of behaviorally-relevant dimensions derived from a large-scale analysis of human similarity judgments directly onto the brain. Our results reveal broadly distributed representations of behaviorally-relevant information, demonstrating selectivity to a wide variety of novel dimensions while capturing known selectivities for visual features and categories. Behaviorally-relevant dimensions were superior to categories at predicting brain responses, yielding mixed selectivity in much of visual cortex and sparse selectivity in category-selective clusters. This framework reconciles seemingly disparate findings regarding regional specialization, explaining category selectivity as a special case of sparse response profiles among representational dimensions, suggesting a more expansive view on visual processing in the human brain.
Collapse
|
12
|
Characterizing the discriminability of visual categorical information in strongly connected voxels. Neuropsychologia 2024; 195:108815. [PMID: 38311112 DOI: 10.1016/j.neuropsychologia.2024.108815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Revised: 01/06/2024] [Accepted: 02/01/2024] [Indexed: 02/06/2024]
Abstract
Functional brain responses are strongly influenced by connectivity. Recently, we demonstrated a major example of this: category discriminability within occipitotemporal cortex (OTC) is enhanced for voxel sets that share strong functional connectivity to distal brain areas, relative to those that share lesser connectivity. That is, within OTC regions, sets of 'most-connected' voxels show improved multivoxel pattern discriminability for tool-, face-, and place stimuli relative to voxels with weaker connectivity to the wider brain. However, understanding whether these effects generalize to other domains (e.g. body perception network), and across different levels of the visual processing streams (e.g. dorsal as well as ventral stream areas) is an important extension of this work. Here, we show that this so-called connectivity-guided decoding (CGD) effect broadly generalizes across a wide range of categories (tools, faces, bodies, hands, places). This effect is robust across dorsal stream areas, but less consistent in earlier ventral stream areas. In the latter regions, category discriminability is generally very high, suggesting that extraction of category-relevant visual properties is less reliant on connectivity to downstream areas. Further, CGD effects are primarily expressed in a category-specific manner: For example, within the network of tool regions, discriminability of tool information is greater than non-tool information. The connectivity-guided decoding approach shown here provides a novel demonstration of the crucial relationship between wider brain connectivity and complex local-level functional responses at different levels of the visual processing streams. Further, this approach generates testable new hypotheses about the relationships between connectivity and local selectivity.
Collapse
|
13
|
Improved modeling of human vision by incorporating robustness to blur in convolutional neural networks. Nat Commun 2024; 15:1989. [PMID: 38443349 PMCID: PMC10915141 DOI: 10.1038/s41467-024-45679-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2023] [Accepted: 01/30/2024] [Indexed: 03/07/2024] Open
Abstract
Whenever a visual scene is cast onto the retina, much of it will appear degraded due to poor resolution in the periphery; moreover, optical defocus can cause blur in central vision. However, the pervasiveness of blurry or degraded input is typically overlooked in the training of convolutional neural networks (CNNs). We hypothesized that the absence of blurry training inputs may cause CNNs to rely excessively on high spatial frequency information for object recognition, thereby causing systematic deviations from biological vision. We evaluated this hypothesis by comparing standard CNNs with CNNs trained on a combination of clear and blurry images. We show that blur-trained CNNs outperform standard CNNs at predicting neural responses to objects across a variety of viewing conditions. Moreover, blur-trained CNNs acquire increased sensitivity to shape information and greater robustness to multiple forms of visual noise, leading to improved correspondence with human perception. Our results provide multi-faceted neurocomputational evidence that blurry visual experiences may be critical for conferring robustness to biological visual systems.
Collapse
|
14
|
The development of visual categorization based on high-level cues. Child Dev 2024; 95:e122-e138. [PMID: 37787438 DOI: 10.1111/cdev.14015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/04/2023]
Abstract
This study investigated the development of rapid visual object categorization. N = 20 adults (Experiment 1), N = 21 five to six-year-old children (Experiment 2), and N = 140 four-, seven-, and eleven-month-old infants (Experiment 3; all predominantly White, 81 females, data collected in 2013-2020) participated in a fast periodic visual stimulation electroencephalographic task. Similar categorization of animal and furniture stimuli emerged in children and adults, with responses much reduced by phase-scrambling (R2 = .34-.73). Categorization was observed from 4 months, but only at 11 months, high-level cues enhanced performance (R2 = .11). Thus, first signs of rapid categorization were evident from 4 months, but similar categorization patterns as in adults were recorded only from 11 months on.
Collapse
|
15
|
Decoding Semantics from Dynamic Brain Activation Patterns: From Trials to Task in EEG/MEG Source Space. eNeuro 2024; 11:ENEURO.0277-23.2023. [PMID: 38320767 PMCID: PMC10913025 DOI: 10.1523/eneuro.0277-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2023] [Revised: 10/21/2023] [Accepted: 11/30/2023] [Indexed: 03/06/2024] Open
Abstract
The temporal dynamics within the semantic brain network and its dependence on stimulus and task parameters are still not well understood. Here, we addressed this by decoding task as well as stimulus information from source-estimated EEG/MEG human data. We presented the same visual word stimuli in a lexical decision (LD) and three semantic decision (SD) tasks. The meanings of the presented words varied across five semantic categories. Source space decoding was applied over time in five ROIs in the left hemisphere (anterior and posterior temporal lobe, inferior frontal gyrus, primary visual areas, and angular gyrus) and one in the right hemisphere (anterior temporal lobe). Task decoding produced sustained significant effects in all ROIs from 50 to 100 ms, both when categorizing tasks with different semantic demands (LD-SD) as well as for similar semantic tasks (SD-SD). In contrast, a semantic word category could only be decoded in lATL, rATL, PTC, and IFG, between 250 and 500 ms. Furthermore, we compared two approaches to source space decoding: conventional ROI-by-ROI decoding and combined-ROI decoding with back-projected activation patterns. The former produced more reliable results for word category decoding while the latter was more informative for task decoding. This indicates that task effects are distributed across the whole semantic network while stimulus effects are more focal. Our results demonstrate that the semantic network is widely distributed but that bilateral anterior temporal lobes together with control regions are particularly relevant for the processing of semantic information.
Collapse
|
16
|
Decoding face recognition abilities in the human brain. PNAS NEXUS 2024; 3:pgae095. [PMID: 38516275 PMCID: PMC10957238 DOI: 10.1093/pnasnexus/pgae095] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/19/2023] [Accepted: 02/20/2024] [Indexed: 03/23/2024]
Abstract
Why are some individuals better at recognizing faces? Uncovering the neural mechanisms supporting face recognition ability has proven elusive. To tackle this challenge, we used a multimodal data-driven approach combining neuroimaging, computational modeling, and behavioral tests. We recorded the high-density electroencephalographic brain activity of individuals with extraordinary face recognition abilities-super-recognizers-and typical recognizers in response to diverse visual stimuli. Using multivariate pattern analyses, we decoded face recognition abilities from 1 s of brain activity with up to 80% accuracy. To better understand the mechanisms subtending this decoding, we compared representations in the brains of our participants with those in artificial neural network models of vision and semantics, as well as with those involved in human judgments of shape and meaning similarity. Compared to typical recognizers, we found stronger associations between early brain representations of super-recognizers and midlevel representations of vision models as well as shape similarity judgments. Moreover, we found stronger associations between late brain representations of super-recognizers and representations of the artificial semantic model as well as meaning similarity judgments. Overall, these results indicate that important individual variations in brain processing, including neural computations extending beyond purely visual processes, support differences in face recognition abilities. They provide the first empirical evidence for an association between semantic computations and face recognition abilities. We believe that such multimodal data-driven approaches will likely play a critical role in further revealing the complex nature of idiosyncratic face recognition in the human brain.
Collapse
|
17
|
Scene-selectivity in CA1/subicular complex: Multivoxel pattern analysis at 7T. Neuropsychologia 2024; 194:108783. [PMID: 38161052 DOI: 10.1016/j.neuropsychologia.2023.108783] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2023] [Revised: 12/21/2023] [Accepted: 12/27/2023] [Indexed: 01/03/2024]
Abstract
Prior univariate functional magnetic resonance imaging (fMRI) studies in humans suggest that the anteromedial subicular complex of the hippocampus is a hub for scene-based cognition. However, it is possible that univariate approaches were not sufficiently sensitive to detect scene-related activity in other subfields that have been implicated in spatial processing (e.g., CA1). Further, as connectivity-based functional gradients in the hippocampus do not respect classical subfield boundary definitions, category selectivity may be distributed across anatomical subfields. Region-of-interest approaches, therefore, may limit our ability to observe category selectivity across discrete subfield boundaries. To address these issues, we applied searchlight multivariate pattern analysis to 7T fMRI data of healthy adults who undertook a simultaneous visual odd-one-out discrimination task for scene and non-scene (including face) visual stimuli, hypothesising that scene classification would be possible in multiple hippocampal regions within, but not constrained to, anteromedial subicular complex and CA1. Indeed, we found that the scene-selective searchlight map overlapped not only with anteromedial subicular complex (distal subiculum, pre/para subiculum), but also inferior CA1, alongside posteromedial (including retrosplenial) and parahippocampal cortices. Probabilistic overlap maps revealed gradients of scene category selectivity, with the strongest overlap located in the medial hippocampus, converging with searchlight findings. This was contrasted with gradients of face category selectivity, which had stronger overlap in more lateral hippocampus, supporting ideas of parallel processing streams for these two categories. Our work helps to map the scene, in contrast to, face processing networks within, and connected to, the human hippocampus.
Collapse
|
18
|
Grid-like entorhinal representation of an abstract value space during prospective decision making. Nat Commun 2024; 15:1198. [PMID: 38336756 PMCID: PMC10858181 DOI: 10.1038/s41467-024-45127-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2023] [Accepted: 01/16/2024] [Indexed: 02/12/2024] Open
Abstract
How valuable a choice option is often changes over time, making the prediction of value changes an important challenge for decision making. Prior studies identified a cognitive map in the hippocampal-entorhinal system that encodes relationships between states and enables prediction of future states, but does not inherently convey value during prospective decision making. In this fMRI study, participants predicted changing values of choice options in a sequence, forming a trajectory through an abstract two-dimensional value space. During this task, the entorhinal cortex exhibited a grid-like representation with an orientation aligned to the axis through the value space most informative for choices. A network of brain regions, including ventromedial prefrontal cortex, tracked the prospective value difference between options. These findings suggest that the entorhinal grid system supports the prediction of future values by representing a cognitive map, which might be used to generate lower-dimensional value signals to guide prospective decision making.
Collapse
|
19
|
Category-based attention facilitates memory search. eNeuro 2024; 11:ENEURO.0012-24.2024. [PMID: 38331577 PMCID: PMC10897531 DOI: 10.1523/eneuro.0012-24.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Accepted: 01/16/2024] [Indexed: 02/10/2024] Open
Abstract
We often need to decide whether the object we look at is also the object we look for. When we look for one specific object, this process can be facilitated by feature-based attention. However, when we look for many objects at the same time (e.g., the products on our shopping list) such a strategy may no longer be possible, as research has shown that we can actively prepare to detect only one or two objects at a time. Therefore, looking for multiple objects additionally requires long-term memory search, slowing down decision making. Interestingly, however, previous research has shown that distractor objects can be efficiently rejected during memory search when they are from a different category than the items in the memory set. Here, using EEG, we show that this efficiency is supported by top-down attention at the category level. In Experiment 1, human participants (both sexes) performed a memory search task on individually presented objects from different categories, most of which were distractors. We observed category-level attentional modulation of distractor processing from ∼150 ms after stimulus onset, expressed both as an evoked response modulation and as an increase in decoding accuracy of same-category distractors. In Experiment 2, memory search was performed on two concurrently presented objects. When both objects were distractors, spatial attention (indexed by the N2pc component) was directed to the object that was of the same category as the objects in the memory set. Together, these results demonstrate how top-down attention can facilitate memory search.Significance statement When we are in the supermarket, we repeatedly decide whether a product we look at (e.g., a banana) is on our memorized shopping list (e.g., apples, oranges, kiwis). This requires searching our memory, which takes time. However, when the product is of an entirely different category (e.g., dairy instead of fruit), the decision can be made quickly. Here, we used EEG to show that this between-category advantage in memory search tasks is supported by top-down attentional modulation of visual processing: The visual response evoked by distractor objects was modulated by category membership, and spatial attention was quickly directed to the location of within-category (vs. between-category) distractors. These results demonstrate a close link between attention and memory.
Collapse
|
20
|
Face adaptation induces duration distortion of subsequent face stimuli in a face category-specific manner. J Vis 2024; 24:7. [PMID: 38386341 PMCID: PMC10896233 DOI: 10.1167/jov.24.2.7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/23/2024] Open
Abstract
Studies have shown that duration perception depends on several visual processes. However, the stages of visual processes that contribute to duration perception remain unclear. This study examined the effects of categorical differences in face adaptation on perceived duration. In all the experiments, we compared the perceived durations of human, monkey, and cat faces (comparison stimuli) after adapting to a human face. Results revealed that the human comparison stimuli were perceived shorter than the monkey and cat comparison stimuli (categorical face adaptation on duration perception [CFAD]). The difference between the face categories disappeared when the adapting stimulus was rendered unrecognizable by phase scrambling, indicating that adaptation to low-level visual properties cannot fully account for the CFAD effect. Furthermore, CFAD was preserved but attenuated when the adapting stimulus was inverted or a 1,000-ms interval was inserted before the comparison stimuli, which implied that CFAD occurred as long as the adapting stimulus was perceived as a face and not simply based on conceptual category processes. These findings indicate that face adaptation affects perceived duration in a category-specific manner (the CFAD effect) and highlights the involvement of visual categorical processes in duration perception.
Collapse
|
21
|
Identifying content-invariant neural signatures of perceptual vividness. PNAS NEXUS 2024; 3:pgae061. [PMID: 38415219 PMCID: PMC10898512 DOI: 10.1093/pnasnexus/pgae061] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 01/31/2024] [Indexed: 02/29/2024]
Abstract
Some conscious experiences are more vivid than others. Although perceptual vividness is a key component of human consciousness, how variation in this magnitude property is registered by the human brain is unknown. A striking feature of neural codes for magnitude in other psychological domains, such as number or reward, is that the magnitude property is represented independently of its sensory features. To test whether perceptual vividness also covaries with neural codes that are invariant to sensory content, we reanalyzed existing magnetoencephalography and functional MRI data from two distinct studies which quantified perceptual vividness via subjective ratings of awareness and visibility. Using representational similarity and decoding analyses, we find evidence for content-invariant neural signatures of perceptual vividness distributed across visual, parietal, and frontal cortices. Our findings indicate that the neural correlates of subjective vividness may share similar properties to magnitude codes in other cognitive domains.
Collapse
|
22
|
Patterns of saliency and semantic features distinguish gaze of expert and novice viewers of surveillance footage. Psychon Bull Rev 2024:10.3758/s13423-024-02454-y. [PMID: 38273144 DOI: 10.3758/s13423-024-02454-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/04/2024] [Indexed: 01/27/2024]
Abstract
When viewing the actions of others, we not only see patterns of body movements, but we also "see" the intentions and social relations of people. Experienced forensic examiners - Closed Circuit Television (CCTV) operators - have been shown to convey superior performance in identifying and predicting hostile intentions from surveillance footage than novices. However, it remains largely unknown what visual content CCTV operators actively attend to, and whether CCTV operators develop different strategies for active information seeking from what novices do. Here, we conducted computational analysis for the gaze-centered stimuli captured by experienced CCTV operators and novices' eye movements when viewing the same surveillance footage. Low-level image features were extracted by a visual saliency model, whereas object-level semantic features were extracted by a deep convolutional neural network (DCNN), AlexNet, from gaze-centered regions. We found that the looking behavior of CCTV operators differs from novices by actively attending to visual contents with different patterns of saliency and semantic features. Expertise in selectively utilizing informative features at different levels of visual hierarchy may play an important role in facilitating the efficient detection of social relationships between agents and the prediction of harmful intentions.
Collapse
|
23
|
The Geometry of Low- and High-Level Perceptual Spaces. J Neurosci 2024; 44:e1460232023. [PMID: 38267235 PMCID: PMC10860617 DOI: 10.1523/jneurosci.1460-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 11/27/2023] [Accepted: 11/28/2023] [Indexed: 01/26/2024] Open
Abstract
Low-level features are typically continuous (e.g., the gamut between two colors), but semantic information is often categorical (there is no corresponding gradient between dog and turtle) and hierarchical (animals live in land, water, or air). To determine the impact of these differences on cognitive representations, we characterized the geometry of perceptual spaces of five domains: a domain dominated by semantic information (animal names presented as words), a domain dominated by low-level features (colored textures), and three intermediate domains (animal images, lightly texturized animal images that were easy to recognize, and heavily texturized animal images that were difficult to recognize). Each domain had 37 stimuli derived from the same animal names. From 13 participants (9F), we gathered similarity judgments in each domain via an efficient psychophysical ranking paradigm. We then built geometric models of each domain for each participant, in which distances between stimuli accounted for participants' similarity judgments and intrinsic uncertainty. Remarkably, the five domains had similar global properties: each required 5-7 dimensions, and a modest amount of spherical curvature provided the best fit. However, the arrangement of the stimuli within these embeddings depended on the level of semantic information: dendrograms derived from semantic domains (word, image, and lightly texturized images) were more "tree-like" than those from feature-dominated domains (heavily texturized images and textures). Thus, the perceptual spaces of domains along this feature-dominated to semantic-dominated gradient shift to a tree-like organization when semantic information dominates, while retaining a similar global geometry.
Collapse
|
24
|
A Chinese verb semantic feature dataset (CVFD). Behav Res Methods 2024; 56:342-361. [PMID: 36622559 DOI: 10.3758/s13428-022-02047-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/01/2022] [Indexed: 01/10/2023]
Abstract
Language is an advanced cognitive function of humans, and verbs play a crucial role in language. To understand how the human brain represents verbs, it is critical to analyze what knowledge humans have about verbs. Thus, several verb feature datasets have been developed in different languages such as English, Spanish, and German. However, there is still a lack of a dataset of Chinese verbs. In this study, we developed a semantic feature dataset of 1140 Chinese Mandarin verbs (CVFD) with 11 dimensions including verb familiarity, agentive subject, patient, action effector, perceptual modality, instrumentality, emotional valence, action imageability, action complexity, action intensity, and the usage scenario of action. We calculated the semantic features of each verb and the correlation between dimensions. We also compared the difference between action, mental, and other verbs and gave some examples about how to use CVFD to classify verbs according to different dimensions. Finally, we discussed the potential applications of CVFD in the fields of neuroscience, psycholinguistics, cultural differences, and artificial intelligence. All the data can be found at https://osf.io/pv29z/ .
Collapse
|
25
|
Neural sensitivity to translational self- and object-motion velocities. Hum Brain Mapp 2024; 45:e26571. [PMID: 38224544 PMCID: PMC10785198 DOI: 10.1002/hbm.26571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Revised: 12/04/2023] [Accepted: 12/07/2023] [Indexed: 01/17/2024] Open
Abstract
The ability to detect and assess world-relative object-motion is a critical computation performed by the visual system. This computation, however, is greatly complicated by the observer's movements, which generate a global pattern of motion on the observer's retina. How the visual system implements this computation is poorly understood. Since we are potentially able to detect a moving object if its motion differs in velocity (or direction) from the expected optic flow generated by our own motion, here we manipulated the relative motion velocity between the observer and the object within a stationary scene as a strategy to test how the brain accomplishes object-motion detection. Specifically, we tested the neural sensitivity of brain regions that are known to respond to egomotion-compatible visual motion (i.e., egomotion areas: cingulate sulcus visual area, posterior cingulate sulcus area, posterior insular cortex [PIC], V6+, V3A, IPSmot/VIP, and MT+) to a combination of different velocities of visually induced translational self- and object-motion within a virtual scene while participants were instructed to detect object-motion. To this aim, we combined individual surface-based brain mapping, task-evoked activity by functional magnetic resonance imaging, and parametric and representational similarity analyses. We found that all the egomotion regions (except area PIC) responded to all the possible combinations of self- and object-motion and were modulated by the self-motion velocity. Interestingly, we found that, among all the egomotion areas, only MT+, V6+, and V3A were further modulated by object-motion velocities, hence reflecting their possible role in discriminating between distinct velocities of self- and object-motion. We suggest that these egomotion regions may be involved in the complex computation required for detecting scene-relative object-motion during self-motion.
Collapse
|
26
|
Hippocampal contributions to novel spatial learning are both age-related and age-invariant. Proc Natl Acad Sci U S A 2023; 120:e2307884120. [PMID: 38055735 PMCID: PMC10723126 DOI: 10.1073/pnas.2307884120] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Accepted: 10/30/2023] [Indexed: 12/08/2023] Open
Abstract
Older adults show declines in spatial memory, although the extent of these alterations is not uniform across the healthy older population. Here, we investigate the stability of neural representations for the same and different spatial environments in a sample of younger and older adults using high-resolution functional MRI of the medial temporal lobes. Older adults showed, on average, lower neural pattern similarity for retrieving the same environment and more variable neural patterns compared to young adults. We also found a positive association between spatial distance discrimination and the distinctiveness of neural patterns between environments. Our analyses suggested that one source for this association was the extent of informational connectivity to CA1 from other subfields, which was dependent on age, while another source was the fidelity of signals within CA1 itself, which was independent of age. Together, our findings suggest both age-dependent and independent neural contributions to spatial memory performance.
Collapse
|
27
|
Deep neural networks are not a single hypothesis but a language for expressing computational hypotheses. Behav Brain Sci 2023; 46:e392. [PMID: 38054329 DOI: 10.1017/s0140525x23001553] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
An ideal vision model accounts for behavior and neurophysiology in both naturalistic conditions and designed lab experiments. Unlike psychological theories, artificial neural networks (ANNs) actually perform visual tasks and generate testable predictions for arbitrary inputs. These advantages enable ANNs to engage the entire spectrum of the evidence. Failures of particular models drive progress in a vibrant ANN research program of human vision.
Collapse
|
28
|
Distinct Neural Components of Visually Guided Grasping during Planning and Execution. J Neurosci 2023; 43:8504-8514. [PMID: 37848285 PMCID: PMC10711727 DOI: 10.1523/jneurosci.0335-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Revised: 07/18/2023] [Accepted: 09/06/2023] [Indexed: 10/19/2023] Open
Abstract
Selecting suitable grasps on three-dimensional objects is a challenging visuomotor computation, which involves combining information about an object (e.g., its shape, size, and mass) with information about the actor's body (e.g., the optimal grasp aperture and hand posture for comfortable manipulation). Here, we used functional magnetic resonance imaging to investigate brain networks associated with these distinct aspects during grasp planning and execution. Human participants of either sex viewed and then executed preselected grasps on L-shaped objects made of wood and/or brass. By leveraging a computational approach that accurately predicts human grasp locations, we selected grasp points that disentangled the role of multiple grasp-relevant factors, that is, grasp axis, grasp size, and object mass. Representational Similarity Analysis revealed that grasp axis was encoded along dorsal-stream regions during grasp planning. Grasp size was first encoded in ventral stream areas during grasp planning then in premotor regions during grasp execution. Object mass was encoded in ventral stream and (pre)motor regions only during grasp execution. Premotor regions further encoded visual predictions of grasp comfort, whereas the ventral stream encoded grasp comfort during execution, suggesting its involvement in haptic evaluation. These shifts in neural representations thus capture the sensorimotor transformations that allow humans to grasp objects.SIGNIFICANCE STATEMENT Grasping requires integrating object properties with constraints on hand and arm postures. Using a computational approach that accurately predicts human grasp locations by combining such constraints, we selected grasps on objects that disentangled the relative contributions of object mass, grasp size, and grasp axis during grasp planning and execution in a neuroimaging study. Our findings reveal a greater role of dorsal-stream visuomotor areas during grasp planning, and, surprisingly, increasing ventral stream engagement during execution. We propose that during planning, visuomotor representations initially encode grasp axis and size. Perceptual representations of object material properties become more relevant instead as the hand approaches the object and motor programs are refined with estimates of the grip forces required to successfully lift the object.
Collapse
|
29
|
The Representation of Observed Actions at the Subordinate, Basic, and Superordinate Level. J Neurosci 2023; 43:8219-8230. [PMID: 37798129 PMCID: PMC10697398 DOI: 10.1523/jneurosci.0700-22.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2022] [Revised: 08/08/2023] [Accepted: 09/06/2023] [Indexed: 10/07/2023] Open
Abstract
Actions can be planned and recognized at different hierarchical levels, ranging from very specific (e.g., to swim backstroke) to very broad (e.g., locomotion). Understanding the corresponding neural representation is an important prerequisite to reveal how our brain flexibly assigns meaning to the world around us. To address this question, we conducted an event-related fMRI study in male and female human participants in which we examined distinct representations of observed actions at the subordinate, basic and superordinate level. Using multiple regression representational similarity analysis (RSA) in predefined regions of interest, we found that the three different taxonomic levels were best captured by patterns of activations in bilateral lateral occipitotemporal cortex (LOTC), showing the highest similarity with the basic level model. A whole-brain multiple regression RSA revealed that information unique to the basic level was captured by patterns of activation in dorsal and ventral portions of the LOTC and in parietal regions. By contrast, the unique information for the subordinate level was limited to bilateral occipitotemporal cortex, while no single cluster was obtained that captured unique information for the superordinate level. The behaviorally established action space was best captured by patterns of activation in the LOTC and superior parietal cortex, and the corresponding neural patterns of activation showed the highest similarity with patterns of activation corresponding to the basic level model. Together, our results suggest that occipitotemporal cortex shows a preference for the basic level model, with flexible access across the subordinate and the basic level.SIGNIFICANCE STATEMENT The human brain captures information at varying levels of abstraction. It is debated which brain regions host representations across different hierarchical levels, with some studies emphasizing parietal and premotor regions, while other studies highlight the role of the lateral occipitotemporal cortex (LOTC). To shed light on this debate, here we examined the representation of observed actions at the three taxonomic levels suggested by Rosch et al. (1976) Our results highlight the role of the LOTC, which hosts a shared representation across the subordinate and the basic level, with the highest similarity with the basic level model. These results shed new light on the hierarchical organization of observed actions and provide insights into the neural basis underlying the basic level advantage.
Collapse
|
30
|
Hippocampal contributions to novel spatial learning are both age-related and age-invariant. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.06.28.546918. [PMID: 37425879 PMCID: PMC10326977 DOI: 10.1101/2023.06.28.546918] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/11/2023]
Abstract
Older adults show declines in spatial memory, although the extent of these alterations is not uniform across the healthy older population. Here, we investigate the stability of neural representations for the same and different spatial environments in a sample of younger and older adults using high-resolution functional magnetic resonance imaging (fMRI) of the medial temporal lobe. Older adults showed, on average, lower neural pattern similarity for retrieving the same environment and more variable neural patterns compared to young adults. We also found a positive association between spatial distance discrimination and the distinctiveness of neural patterns between environments. Our analyses suggested that one source for this association was the extent of informational connectivity to CA1 from other subfields, which was dependent on age, while another source was the fidelity of signals within CA1 itself, which was independent of age. Together, our findings suggest both age-dependent and independent neural contributions to spatial memory performance.
Collapse
|
31
|
Parietal-driven visual working memory representation in occipito-temporal cortex. Curr Biol 2023; 33:4516-4523.e5. [PMID: 37741281 PMCID: PMC10615870 DOI: 10.1016/j.cub.2023.08.080] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Revised: 07/24/2023] [Accepted: 08/25/2023] [Indexed: 09/25/2023]
Abstract
Human fMRI studies have documented extensively that the content of visual working memory (VWM) can be reliably decoded from fMRI voxel response patterns during the delay period in both the occipito-temporal cortex (OTC), including early visual areas (EVC), and the posterior parietal cortex (PPC).1,2,3,4 Further work has revealed that VWM signal in OTC is largely sustained by feedback from associative areas such as prefrontal cortex (PFC) and PPC.4,5,6,7,8,9 It is unclear, however, if feedback during VWM simply restores sensory representations initially formed in OTC or if it can reshape the representational content of OTC during VWM delay. Taking advantage of a recent finding showing that object representational geometry differs between OTC and PPC in perception,10 here we find that, during VWM delay, the object representational geometry in OTC becomes more aligned with that of PPC during perception than with itself during perception. This finding supports the role of feedback in shaping the content of VWM in OTC, with the VWM content of OTC more determined by information retained in PPC than by the sensory information initially encoded in OTC.
Collapse
|
32
|
Brain Representation of Animal and Non-Animal Images in Patients with Mild Cognitive Impairment and Alzheimer's Disease. J Alzheimers Dis Rep 2023; 7:1133-1152. [PMID: 38025804 PMCID: PMC10657719 DOI: 10.3233/adr-230132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 09/06/2023] [Indexed: 12/01/2023] Open
Abstract
Background In early Alzheimer's disease (AD), high-level visual functions and processing speed are impacted. Few functional magnetic resonance imaging (fMRI) studies have investigated high-level visual deficits in AD, yet none have explored brain activity patterns during rapid animal/non-animal categorization tasks. Objective To address this, we utilized the previously known Integrated Cognitive Assessment (ICA) to collect fMRI data and compare healthy controls (HC) to individuals with mild cognitive impairment (MCI) and mild AD. Methods The ICA encompasses a rapid visual categorization task that involves distinguishing between animals and non-animals within natural scenes. To comprehensively explore variations in brain activity levels and patterns, we conducted both univariate and multivariate analyses of fMRI data. Results The ICA task elicited activation across a range of brain regions, encompassing the temporal, parietal, occipital, and frontal lobes. Univariate analysis, which compared responses to animal versus non-animal stimuli, showed no significant differences in the regions of interest (ROIs) across all groups, with the exception of the left anterior supramarginal gyrus in the HC group. In contrast, multivariate analysis revealed that in both HC and MCI groups, several regions could differentiate between animals and non-animals based on distinct patterns of activity. Notably, such differentiation was absent within the mild AD group. Conclusions Our study highlights the ICA task's potential as a valuable cognitive assessment tool designed for MCI and AD. Additionally, our use of fMRI pattern analysis provides valuable insights into the complex changes in brain function associated with AD. This approach holds promise for enhancing our understanding of the disease's progression.
Collapse
|
33
|
Proximity to boundaries reveals spatial context representation in human hippocampal CA1. Neuropsychologia 2023; 189:108656. [PMID: 37541615 DOI: 10.1016/j.neuropsychologia.2023.108656] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2023] [Revised: 06/30/2023] [Accepted: 08/01/2023] [Indexed: 08/06/2023]
Abstract
Recollection of real-world events is often accompanied by a sense of being in the place where the event transpired. Convergent evidence suggests the hippocampus plays a key role in supporting episodic memory by associating information with the time and place it was originally encountered. This representation is reinstated during memory retrieval. However, little is known about the roles of different subfields of the human hippocampus in this process. Research in humans and non-human animal models has suggested that spatial environmental boundaries have a powerful influence on spatial and episodic memory, as well as hippocampal representations of contexts and events. Here, we used high-resolution fMRI to investigate how boundaries influence hippocampal activity patterns during the recollection of objects encountered in different spatial contexts. During the encoding phase, participants viewed objects once in a naturalistic virtual reality task in which they passively explored two rooms in one of two houses. Following the encoding phase, participants were scanned while they recollected items in the absence of any spatial contextual information. Our behavioral results demonstrated that spatial context memory was enhanced for objects encountered near a boundary. Activity patterns in CA1 carried information about the spatial context associated with each of these boundary items. Exploratory analyses revealed that recollection performance was correlated with the fidelity of retrieved spatial context representations in anterior parahippocampal cortex and subiculum. Our results highlight the privileged role of boundaries in CA1 and suggest more generally a close relationship between memory for spatial contexts and representations in the hippocampus and parahippocampal region.
Collapse
|
34
|
Modality-specific brain representations during automatic processing of face, voice and body expressions. Front Neurosci 2023; 17:1132088. [PMID: 37869514 PMCID: PMC10587395 DOI: 10.3389/fnins.2023.1132088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2022] [Accepted: 09/05/2023] [Indexed: 10/24/2023] Open
Abstract
A central question in affective science and one that is relevant for its clinical applications is how emotions provided by different stimuli are experienced and represented in the brain. Following the traditional view emotional signals are recognized with the help of emotion concepts that are typically used in descriptions of mental states and emotional experiences, irrespective of the sensory modality. This perspective motivated the search for abstract representations of emotions in the brain, shared across variations in stimulus type (face, body, voice) and sensory origin (visual, auditory). On the other hand, emotion signals like for example an aggressive gesture, trigger rapid automatic behavioral responses and this may take place before or independently of full abstract representation of the emotion. This pleads in favor specific emotion signals that may trigger rapid adaptative behavior only by mobilizing modality and stimulus specific brain representations without relying on higher order abstract emotion categories. To test this hypothesis, we presented participants with naturalistic dynamic emotion expressions of the face, the whole body, or the voice in a functional magnetic resonance (fMRI) study. To focus on automatic emotion processing and sidestep explicit concept-based emotion recognition, participants performed an unrelated target detection task presented in a different sensory modality than the stimulus. By using multivariate analyses to assess neural activity patterns in response to the different stimulus types, we reveal a stimulus category and modality specific brain organization of affective signals. Our findings are consistent with the notion that under ecological conditions emotion expressions of the face, body and voice may have different functional roles in triggering rapid adaptive behavior, even if when viewed from an abstract conceptual vantage point, they may all exemplify the same emotion. This has implications for a neuroethologically grounded emotion research program that should start from detailed behavioral observations of how face, body, and voice expressions function in naturalistic contexts.
Collapse
|
35
|
Effects of avatar shape and motion on mirror neuron system activity. Front Hum Neurosci 2023; 17:1173185. [PMID: 37859767 PMCID: PMC10582709 DOI: 10.3389/fnhum.2023.1173185] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2023] [Accepted: 09/21/2023] [Indexed: 10/21/2023] Open
Abstract
Humanness is an important characteristic for facilitating interpersonal communication, particularly through avatars in the metaverse. In this study, we explored the mirror neuron system (MNS) as a potential neural basis for perceiving humanness in avatars. Although previous research suggests that the MNS may be influenced by human-like shape and motion, the results have been inconsistent due to the diversity and complexity of the MNS investigation. Therefore, this study aims to investigate the effects of shape and motion humanness in avatars on MNS activity. Participants viewed videos of avatars with four different shapes (HumanShape, AngularShape, AbbreviatedShape, and ScatteredShape) and two types of motion (HumanMotion and LinearMotion), and their μ-wave attenuation in the electroencephalogram was evaluated. Results from a questionnaire indicated that HumanMotion was perceived as human-like, while AbbreviatedShape and ScatteredShape were seen as non-human-like. AngularShape's humanity was indefinite. The MNS was activated as expected for avatars with human-like shapes and/or motions. However, for non-human-like motions, there were differences in activity trends depending on the avatar shape. Specifically, avatars with HumanShape and ScatteredShape in LinearMotion activated the MNS, but the MNS was indifferent to AngularShape and AbbreviatedShape. These findings suggest that when avatars make non-human-like motions, the MNS is activated not only for human-like appearance but also for the scattered and exaggerated appearance of the human body in the avatar shape. These findings could enhance inter-avatar communication by considering brain activity.
Collapse
|
36
|
Abstract perceptual choice signals during action-linked decisions in the human brain. PLoS Biol 2023; 21:e3002324. [PMID: 37816222 PMCID: PMC10564462 DOI: 10.1371/journal.pbio.3002324] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2022] [Accepted: 09/04/2023] [Indexed: 10/12/2023] Open
Abstract
Humans can make abstract choices independent of motor actions. However, in laboratory tasks, choices are typically reported with an associated action. Consequentially, knowledge about the neural representation of abstract choices is sparse, and choices are often thought to evolve as motor intentions. Here, we show that in the human brain, perceptual choices are represented in an abstract, motor-independent manner, even when they are directly linked to an action. We measured MEG signals while participants made choices with known or unknown motor response mapping. Using multivariate decoding, we quantified stimulus, perceptual choice, and motor response information with distinct cortical distributions. Choice representations were invariant to whether the response mapping was known during stimulus presentation, and they occupied a distinct representational space from motor signals. As expected from an internal decision variable, they were informed by the stimuli, and their strength predicted decision confidence and accuracy. Our results demonstrate abstract neural choice signals that generalize to action-linked decisions, suggesting a general role of an abstract choice stage in human decision-making.
Collapse
|
37
|
The representation of occluded image regions in area V1 of monkeys and humans. Curr Biol 2023; 33:3865-3871.e3. [PMID: 37643620 DOI: 10.1016/j.cub.2023.08.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Revised: 07/04/2023] [Accepted: 08/02/2023] [Indexed: 08/31/2023]
Abstract
Neuronal activity in the primary visual cortex (V1) is driven by feedforward input from within the neurons' receptive fields (RFs) and modulated by contextual information in regions surrounding the RF. The effect of contextual information on spiking activity occurs rapidly and is therefore challenging to dissociate from feedforward input. To address this challenge, we recorded the spiking activity of V1 neurons in monkeys viewing either natural scenes or scenes where the information in the RF was occluded, effectively removing the feedforward input. We found that V1 neurons responded rapidly and selectively to occluded scenes. V1 responses elicited by occluded stimuli could be used to decode individual scenes and could be predicted from those elicited by non-occluded images, indicating that there is an overlap between visually driven and contextual responses. We used representational similarity analysis to show that the structure of V1 representations of occluded scenes measured with electrophysiology in monkeys correlates strongly with the representations of the same scenes in humans measured with functional magnetic resonance imaging (fMRI). Our results reveal that contextual influences rapidly alter V1 spiking activity in monkeys over distances of several degrees in the visual field, carry information about individual scenes, and resemble those in human V1. VIDEO ABSTRACT.
Collapse
|
38
|
High-dimensional topographic organization of visual features in the primate temporal lobe. Nat Commun 2023; 14:5931. [PMID: 37739988 PMCID: PMC10517140 DOI: 10.1038/s41467-023-41584-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2023] [Accepted: 09/07/2023] [Indexed: 09/24/2023] Open
Abstract
The inferotemporal cortex supports our supreme object recognition ability. Numerous studies have been conducted to elucidate the functional organization of this brain area, but there are still important questions that remain unanswered, including how this organization differs between humans and non-human primates. Here, we use deep neural networks trained on object categorization to construct a 25-dimensional space of visual features, and systematically measure the spatial organization of feature preference in both male monkey brains and human brains using fMRI. These feature maps allow us to predict the selectivity of a previously unknown region in monkey brains, which is corroborated by additional fMRI and electrophysiology experiments. These maps also enable quantitative analyses of the topographic organization of the temporal lobe, demonstrating the existence of a pair of orthogonal gradients that differ in spatial scale and revealing significant differences in the functional organization of high-level visual areas between monkey and human brains.
Collapse
|
39
|
Visual Representations: Insights from Neural Decoding. Annu Rev Vis Sci 2023; 9:313-335. [PMID: 36889254 DOI: 10.1146/annurev-vision-100120-025301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/10/2023]
Abstract
Patterns of brain activity contain meaningful information about the perceived world. Recent decades have welcomed a new era in neural analyses, with computational techniques from machine learning applied to neural data to decode information represented in the brain. In this article, we review how decoding approaches have advanced our understanding of visual representations and discuss efforts to characterize both the complexity and the behavioral relevance of these representations. We outline the current consensus regarding the spatiotemporal structure of visual representations and review recent findings that suggest that visual representations are at once robust to perturbations, yet sensitive to different mental states. Beyond representations of the physical world, recent decoding work has shone a light on how the brain instantiates internally generated states, for example, during imagery and prediction. Going forward, decoding has remarkable potential to assess the functional relevance of visual representations for human behavior, reveal how representations change across development and during aging, and uncover their presentation in various mental disorders.
Collapse
|
40
|
Abstract
Perception and memory are traditionally thought of as separate cognitive functions, supported by distinct brain regions. The canonical perspective is that perceptual processing of visual information is supported by the ventral visual stream, whereas long-term declarative memory is supported by the medial temporal lobe. However, this modular framework cannot account for the increasingly large body of evidence that reveals a role for early visual areas in long-term recognition memory and a role for medial temporal lobe structures in high-level perceptual processing. In this article, we review relevant research conducted in humans, nonhuman primates, and rodents. We conclude that the evidence is largely inconsistent with theoretical proposals that draw sharp functional boundaries between perceptual and memory systems in the brain. Instead, the weight of the empirical findings is best captured by a representational-hierarchical model that emphasizes differences in content, rather than in cognitive processes within the ventral visual stream and medial temporal lobe.
Collapse
|
41
|
Neural and behavioral signatures of the multidimensionality of manipulable object processing. Commun Biol 2023; 6:940. [PMID: 37709924 PMCID: PMC10502059 DOI: 10.1038/s42003-023-05323-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Accepted: 09/04/2023] [Indexed: 09/16/2023] Open
Abstract
Understanding how we recognize objects requires unravelling the variables that govern the way we think about objects and the neural organization of object representations. A tenable hypothesis is that the organization of object knowledge follows key object-related dimensions. Here, we explored, behaviorally and neurally, the multidimensionality of object processing. We focused on within-domain object information as a proxy for the decisions we typically engage in our daily lives - e.g., identifying a hammer in the context of other tools. We extracted object-related dimensions from subjective human judgments on a set of manipulable objects. We show that the extracted dimensions are cognitively interpretable and relevant - i.e., participants are able to consistently label them, and these dimensions can guide object categorization; and are important for the neural organization of knowledge - i.e., they predict neural signals elicited by manipulable objects. This shows that multidimensionality is a hallmark of the organization of manipulable object knowledge.
Collapse
|
42
|
Multidimensional object properties are dynamically represented in the human brain. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.09.08.556679. [PMID: 37745325 PMCID: PMC10515754 DOI: 10.1101/2023.09.08.556679] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/26/2023]
Abstract
Our visual world consists of an immense number of unique objects and yet, we are easily able to identify, distinguish, interact, and reason about the things we see within several hundred milliseconds. This requires that we flexibly integrate and focus on different object properties to support specific behavioral goals. In the current study, we examined how these rich object representations unfold in the human brain by modelling time-resolved MEG signals evoked by viewing thousands of objects. Using millions of behavioral judgments to guide our understanding of the neural representation of the object space, we find distinct temporal profiles across the object dimensions. These profiles fell into two broad types with either a distinct and early peak (~150 ms) or a slow rise to a late peak (~300 ms). Further, the early effects are stable across participants in contrast to later effects which show more variability across people. This highlights that early peaks may carry stimulus-specific and later peaks subject-specific information. Given that the dimensions with early peaks seem to be primarily visual dimensions and those with later peaks more conceptual, our results suggest that conceptual processing is more variable across people. Together, these data provide a comprehensive account of how a variety of object properties unfold in the human brain and contribute to the rich nature of object vision.
Collapse
|
43
|
The locus of recognition memory signals in human cortex depends on the complexity of the memory representations. Cereb Cortex 2023; 33:9835-9849. [PMID: 37401000 DOI: 10.1093/cercor/bhad248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2022] [Revised: 06/12/2023] [Accepted: 06/14/2023] [Indexed: 07/05/2023] Open
Abstract
According to a "Swiss Army Knife" model of the brain, cognitive functions such as episodic memory and face perception map onto distinct neural substrates. In contrast, representational accounts propose that each brain region is best explained not by which specialized function it performs, but by the type of information it represents with its neural firing. In a functional magnetic resonance imaging study, we asked whether the neural signals supporting recognition memory fall mandatorily within the medial temporal lobes (MTL), traditionally thought the seat of declarative memory, or whether these signals shift within cortex according to the content of the memory. Participants studied objects and scenes that were unique conjunctions of pre-defined visual features. Next, we tested recognition memory in a task that required mnemonic discrimination of both simple features and complex conjunctions. Feature memory signals were strongest in posterior visual regions, declining with anterior progression toward the MTL, while conjunction memory signals followed the opposite pattern. Moreover, feature memory signals correlated with feature memory discrimination performance most strongly in posterior visual regions, whereas conjunction memory signals correlated with conjunction memory discrimination most strongly in anterior sites. Thus, recognition memory signals shifted with changes in memory content, in line with representational accounts.
Collapse
|
44
|
Statistical inference on representational geometries. eLife 2023; 12:e82566. [PMID: 37610302 PMCID: PMC10446828 DOI: 10.7554/elife.82566] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2022] [Accepted: 08/07/2023] [Indexed: 08/24/2023] Open
Abstract
Neuroscience has recently made much progress, expanding the complexity of both neural activity measurements and brain-computational models. However, we lack robust methods for connecting theory and experiment by evaluating our new big models with our new big data. Here, we introduce new inference methods enabling researchers to evaluate and compare models based on the accuracy of their predictions of representational geometries: A good model should accurately predict the distances among the neural population representations (e.g. of a set of stimuli). Our inference methods combine novel 2-factor extensions of crossvalidation (to prevent overfitting to either subjects or conditions from inflating our estimates of model accuracy) and bootstrapping (to enable inferential model comparison with simultaneous generalization to both new subjects and new conditions). We validate the inference methods on data where the ground-truth model is known, by simulating data with deep neural networks and by resampling of calcium-imaging and functional MRI data. Results demonstrate that the methods are valid and conclusions generalize correctly. These data analysis methods are available in an open-source Python toolbox (rsatoolbox.readthedocs.io).
Collapse
|
45
|
Improved modeling of human vision by incorporating robustness to blur in convolutional neural networks. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.07.29.551089. [PMID: 37577646 PMCID: PMC10418076 DOI: 10.1101/2023.07.29.551089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/15/2023]
Abstract
Whenever a visual scene is cast onto the retina, much of it will appear degraded due to poor resolution in the periphery; moreover, optical defocus can cause blur in central vision. However, the pervasiveness of blurry or degraded input is typically overlooked in the training of convolutional neural networks (CNNs). We hypothesized that the absence of blurry training inputs may cause CNNs to rely excessively on high spatial frequency information for object recognition, thereby causing systematic deviations from biological vision. We evaluated this hypothesis by comparing standard CNNs with CNNs trained on a combination of clear and blurry images. We show that blur-trained CNNs outperform standard CNNs at predicting neural responses to objects across a variety of viewing conditions. Moreover, blur-trained CNNs acquire increased sensitivity to shape information and greater robustness to multiple forms of visual noise, leading to improved correspondence with human perception. Our results provide novel neurocomputational evidence that blurry visual experiences are very important for conferring robustness to biological visual systems.
Collapse
|
46
|
Distinct ventral stream and prefrontal cortex representational dynamics during sustained conscious visual perception. Cell Rep 2023; 42:112752. [PMID: 37422763 PMCID: PMC10530642 DOI: 10.1016/j.celrep.2023.112752] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 05/12/2023] [Accepted: 06/20/2023] [Indexed: 07/11/2023] Open
Abstract
Instances of sustained stationary sensory input are ubiquitous. However, previous work focused almost exclusively on transient onset responses. This presents a critical challenge for neural theories of consciousness, which should account for the full temporal extent of experience. To address this question, we use intracranial recordings from ten human patients with epilepsy to view diverse images of multiple durations. We reveal that, in sensory regions, despite dramatic changes in activation magnitude, the distributed representation of categories and exemplars remains sustained and stable. In contrast, in frontoparietal regions, we find transient content representation at stimulus onset. Our results highlight the connection between the anatomical and temporal correlates of experience. To the extent perception is sustained, it may rely on sensory representations and to the extent perception is discrete, centered on perceptual updating, it may rely on frontoparietal representations.
Collapse
|
47
|
Adaptive coding of stimulus information in human frontoparietal cortex during visual classification. Neuroimage 2023; 274:120150. [PMID: 37191656 DOI: 10.1016/j.neuroimage.2023.120150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Revised: 04/19/2023] [Accepted: 04/29/2023] [Indexed: 05/17/2023] Open
Abstract
The neural mechanisms of how frontal and parietal brain regions support flexible adaptation of behavior remain poorly understood. Here, we used functional magnetic resonance imaging (fMRI) and representational similarity analysis (RSA) to investigate frontoparietal representations of stimulus information during visual classification under varying task demands. Based on prior research, we predicted that increasing perceptual task difficulty should lead to adaptive changes in stimulus coding: task-relevant category information should be stronger, while task-irrelevant exemplar-level stimulus information should become weaker, reflecting a focus on the behaviorally relevant category information. Counter to our expectations, however, we found no evidence for adaptive changes in category coding. We did find weakened coding at the exemplar-level within categories however, demonstrating that task-irrelevant information is de-emphasized in frontoparietal cortex. These findings reveal adaptive coding of stimulus information at the exemplar-level, highlighting how frontoparietal regions might support behavior even under challenging conditions.
Collapse
|
48
|
Functionally analogous body- and animacy-responsive areas are present in the dog (Canis familiaris) and human occipito-temporal lobe. Commun Biol 2023; 6:645. [PMID: 37369804 PMCID: PMC10300132 DOI: 10.1038/s42003-023-05014-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2022] [Accepted: 06/05/2023] [Indexed: 06/29/2023] Open
Abstract
Comparing the neural correlates of socio-cognitive skills across species provides insights into the evolution of the social brain and has revealed face- and body-sensitive regions in the primate temporal lobe. Although from a different lineage, dogs share convergent visuo-cognitive skills with humans and a temporal lobe which evolved independently in carnivorans. We investigated the neural correlates of face and body perception in dogs (N = 15) and humans (N = 40) using functional MRI. Combining univariate and multivariate analysis approaches, we found functionally analogous occipito-temporal regions involved in the perception of animate entities and bodies in both species and face-sensitive regions in humans. Though unpredicted, we also observed neural representations of faces compared to inanimate objects, and dog compared to human bodies in dog olfactory regions. These findings shed light on the evolutionary foundations of human and dog social cognition and the predominant role of the temporal lobe.
Collapse
|
49
|
Cortical topographic motifs emerge in a self-organized map of object space. SCIENCE ADVANCES 2023; 9:eade8187. [PMID: 37343093 DOI: 10.1126/sciadv.ade8187] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2022] [Accepted: 05/17/2023] [Indexed: 06/23/2023]
Abstract
The human ventral visual stream has a highly systematic organization of object information, but the causal pressures driving these topographic motifs are highly debated. Here, we use self-organizing principles to learn a topographic representation of the data manifold of a deep neural network representational space. We find that a smooth mapping of this representational space showed many brain-like motifs, with a large-scale organization by animacy and real-world object size, supported by mid-level feature tuning, with naturally emerging face- and scene-selective regions. While some theories of the object-selective cortex posit that these differently tuned regions of the brain reflect a collection of distinctly specified functional modules, the present work provides computational support for an alternate hypothesis that the tuning and topography of the object-selective cortex reflect a smooth mapping of a unified representational space.
Collapse
|
50
|
Spikiness and animacy as potential organizing principles of human ventral visual cortex. Cereb Cortex 2023; 33:8194-8217. [PMID: 36958809 PMCID: PMC10321104 DOI: 10.1093/cercor/bhad108] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2022] [Revised: 03/05/2023] [Accepted: 03/06/2023] [Indexed: 03/25/2023] Open
Abstract
Considerable research has been devoted to understanding the fundamental organizing principles of the ventral visual pathway. A recent study revealed a series of 3-4 topographical maps arranged along the macaque inferotemporal (IT) cortex. The maps articulated a two-dimensional space based on the spikiness and animacy of visual objects, with "inanimate-spiky" and "inanimate-stubby" regions of the maps constituting two previously unidentified cortical networks. The goal of our study was to determine whether a similar functional organization might exist in human IT. To address this question, we presented the same object stimuli and images from "classic" object categories (bodies, faces, houses) to humans while recording fMRI activity at 7 Tesla. Contrasts designed to reveal the spikiness-animacy object space evoked extensive significant activation across human IT. However, unlike the macaque, we did not observe a clear sequence of complete maps, and selectivity for the spikiness-animacy space was deeply and mutually entangled with category-selectivity. Instead, we observed multiple new stimulus preferences in category-selective regions, including functional sub-structure related to object spikiness in scene-selective cortex. Taken together, these findings highlight spikiness as a promising organizing principle of human IT and provide new insights into the role of category-selective regions in visual object processing.
Collapse
|