Early detection and intervention in schizophrenia requires mechanism-based biomarkers that capture neural circuitry dysfunction, allowing better patient stratification, monitoring of disease progression and treatment. In prefrontal cortex and blood of redox dysregulated mice (Gclm-KO ± GBR), oxidative stress induces miR-137 upregulation, leading to decreased COX6A2 and mitophagy markers (NIX, Fundc1, and LC3B) and to accumulation of damaged mitochondria, further exacerbating oxidative stress and parvalbumin interneurons (PVI) impairment. MitoQ, a mitochondria-targeted antioxidant, rescued all these processes. Translating to early psychosis patients (EPP), blood exosomal miR-137 increases and COX6A2 decreases, combined with mitophagy markers alterations, suggest that observations made centrally and peripherally in animal model were reflected in patients’ blood. Higher exosomal miR-137 and lower COX6A2 levels were associated with a reduction of ASSR gamma oscillations in EEG. As ASSR requires proper PVI-related networks, alterations in miR-137/COX6A2 plasma exosome levels may represent a proxy marker of PVI cortical microcircuit impairment. EPP can be stratified in two subgroups: (a) a patients’ group with mitochondrial dysfunction “Psy-D”, having high miR-137 and low COX6A2 levels in exosomes, and (b) a “Psy-ND” subgroup with no/low mitochondrial impairment, including patients having miR-137 and COX6A2 levels in the range of controls. Psy-D patients exhibited more impaired ASSR responses in association with worse psychopathological status, neurocognitive performance, and global and social functioning, suggesting that impairment of PVI mitochondria leads to more severe disease profiles. This stratification would allow, with high selectivity and specificity, the selection of patients for treatments targeting brain mitochondria dysregulation and capture the clinical and functional efficacy of future clinical trials.
Real-world environments are nearly always multisensory in nature. Processing in such situations confers perceptual advantages, but its automaticity remains poorly understood. Automaticity has been invoked to explain the activation of visual cortices by laterally-presented sounds. This has been observed even when the sounds were task-irrelevant and spatially uninformative about subsequenttargets. An auditory-evoked contralateral occipital positivity (ACOP) at ~250ms post-sound onset has been postulated as the event-related potential (ERP) correlate of this cross-modal effect. However, the spatial dimension of the stimuli was nevertheless relevant in all prior studies where the ACOP was observed. By manipulating the implicit predictability of the location of lateralised sounds in a passive auditory paradigm, we tested the automaticity of cross-modal activations of visual cortices.128-channel ERP data from healthy participants were analysed within an electrical neuroimaging framework. The timing, topography, and localisation resembled previous characterisations of the ACOP. However, the cross-modal activations of visual cortices by sounds were critically dependent on whether the sound location was (un)predictable. Our results are the first direct evidence that this particular cross-modal process is not (fully) automatic; instead, it is context-contingent. More generally, the present findings provide novel insights into the importance of context-related factors in controlling information processing across the senses, and call for a revision of current models of automaticity in cognitive sciences.
In real-world environments, information is typically multisensory, and objects are a primary unit of information processing. Object recognition and action necessitate attentional selection of task-relevant from among task-irrelevant objects. However, the brain and cognitive mechanisms governing these processes remain not well understood. Here, we demonstrate that attentional selection of visual objects is controlled by integrated top–down audiovisual object representations (“attentional templates”) while revealing a new brain mechanism through which they can operate. In multistimulus (visual) arrays, attentional selection of objects in humans and animal models is traditionally quantified via “the N2pc component”: spatially selective enhancements of neural processing of objects within ventral visual cortices at approximately 150–300 msec poststimulus. In our adaptation of Folk et al.'s [Folk, C. L., Remington, R. W., & Johnston, J. C. Involuntary covert orienting is contingent on attentional control settings. Journal of Experimental Psychology: Human Perception and Performance, 18, 1030–1044, 1992] spatial cueing paradigm, visual cues elicited weaker behavioral attention capture and an attenuated N2pc during audiovisual versus visual search. To provide direct evidence for the brain, and so, cognitive, mechanisms underlying top–down control in multisensory search, we analyzed global features of the electrical field at the scalp across our N2pcs. In the N2pc time window (170–270 msec), color cues elicited brain responses differing in strength and their topography. This latter finding is indicative of changes in active brain sources. Thus, in multisensory environments, attentional selection is controlled via integrated top–down object representations, and so not only by separate sensory-specific top–down feature templates (as suggested by traditional N2pc analyses). We discuss how the electrical neuroimaging approach can aid research on top–down attentional control in naturalistic, multisensory settings and on other neurocognitive functions in the growing area of real-world neuroscience.
Sensory impairments constitute core dysfunctions in schizophrenia. In the auditory modality, impaired mismatch negativity (MMN) has been observed in chronic schizophrenia and may reflect N-methyl-d-aspartate (NMDA) hypo-function, consistent with models of schizophrenia based on oxidative stress. Moreover, a recent study demonstrated deficits in the N100 component of the auditory evoked potential (AEP) in early psychosis patients. Previous work has shown that add-on administration of the glutathione precursor N-acetyl-cysteine (NAC) improves the MMN and clinical symptoms in chronic schizophrenia. To date, it remains unknown whether NAC also improves general low-level auditory processing and if its efficacy would extend to early-phase psychosis. We addressed these issues with a randomized, double-blind study of a small sample (N=15) of early psychosis (EP) patients and 18 healthy controls from whom AEPs were recorded during an active, auditory oddball task. Patients were recorded twice: once prior to NAC/placebo administration and once after six months of treatment. The N100 component was significantly smaller in patients before NAC administration versus controls. Critically, NAC administration improved this AEP deficit. Source estimations revealed increased activity in the left temporo-parietal lobe in patients after NAC administration. Overall, the data from this pilot study, which call for replication in a larger sample, indicate that NAC improves low-level auditory processing in early psychosis.
Everyday vision includes the detection of stimuli, figure-ground segregation, as well as object localization and recognition. Such processes must often surmount impoverished or noisy conditions; borders are perceived despite occlusion or absent contrast gradients. These illusory contours (ICs) are an example of so-called mid-level vision, with an event-related potential (ERP) correlate at ∼100-150 ms post-stimulus onset and originating within lateral-occipital cortices (the IC). Presently, visual completion processes supporting IC perception are considered exclusively visual; any influence from other sensory modalities is currently unknown. It is now well-established that multisensory processes can influence both low-level vision (e.g. detection) as well as higher-level object recognition. By contrast, it is unknown if mid-level vision exhibits multisensory benefits and, if so, through what mechanisms. We hypothesized that sounds would impact the IC. We recorded 128-channel ERPs from 17 healthy, sighted participants who viewed ICs or no-contour (NC) counterparts either in the presence or absence of task-irrelevant sounds. The IC was enhanced by sounds and resulted in the recruitment of a distinct configuration of active brain areas over the 70-170 ms post-stimulus period. IC-related source-level activity within the lateral occipital cortex (LOC), inferior parietal lobe (IPL), as well as primary visual cortex (V1) were enhanced by sounds. Moreover, the activity in these regions was correlated when sounds were present, but not when absent. Results from a control experiment, which employed amodal variants of the stimuli, suggested that sounds impact the perceived brightness of the IC rather than shape formation per se. We provide the first demonstration that multisensory processes augment mid-level vision and everyday visual completion processes, and that one of the mechanisms is brightness enhancement. These results have important implications for the design of treatments and/or visual aids for low-vision patients.
Highlights By age 7, children show adult-like task-set contingent attentional capture in behavior (top-down visual attentional control). Children showed no behavioral evidence for multisensory enhancement of attention capture by visual objects paired with sounds. But 9-year-olds adult-like EEG topographic patterns, differing when elicited by multisensory vs. purely visual distractors. Traditional N2pc analyses showed no N2pc component in any of the children groups, and no multisensory modulations in adults. Electrical neuroimaging of well-known ERP components is more sensitive to developmental change in neurocognitive processes.
Distinct anatomical and functional pathways are postulated for analysing a sound's object-related ('what') and space-related ('where') information. It remains unresolved to which extent distinct or overlapping neural resources subserve specific object-related dimensions (i.e. who is speaking and what is being said can both be derived from the same acoustic input). To address this issue, we recorded high-density auditory evoked potentials (AEPs) while participants selectively attended and discriminated sounds according to their pitch, speaker identity, uttered syllable ('what' dimensions) or their location ('where'). Sound acoustics were held constant across blocks; the only manipulation involved the sound dimension that participants had to attend to. The task-relevant dimension was varied across blocks. AEPs from healthy participants were analysed within an electrical neuroimaging framework to differentiate modulations in response strength from modulations in response topography; the latter of which forcibly follow from changes in the configuration of underlying sources. There were no behavioural differences in discrimination of sounds across the 4 feature dimensions. As early as 90ms post-stimulus onset, AEP topographies differed across 'what' conditions, supporting a functional sub-segregation within the auditory 'what' pathway. This study characterises the spatio-temporal dynamics of segregated, yet parallel, processing of multiple sound object-related feature dimensions when selective attention is directed to them.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.