Lateral occipital cortical areas are involved in the perception of objects, but it is not clear how these areas interact with first tier visual areas. Using synthetic images portraying a simple texture-defined figure and an electrophysiological paradigm that allows us to monitor cortical responses to figure and background regions separately, we found distinct neuronal networks responsible for the processing of each region. The figure region of our displays was tagged with one temporal frequency (3.0 Hz) and the background region with another (3.6 Hz). Spectral analysis was used to separate the responses to the two regions during their simultaneous presentation. Distributed source reconstructions were made by using the minimum norm method, and cortical current density was measured in a set of visual areas defined on retinotopic and functional criteria with the use of functional magnetic resonance imaging. The results of the main experiments, combined with a set of control experiments, indicate that the figure region, but not the background, was routed preferentially to lateral cortex. A separate network extending from first tier through more dorsal areas responded preferentially to the background region. The figure-related responses were mostly invariant with respect to the texture types used to define the figure, did not depend on its spatial location or size, and mostly were unaffected by attentional instructions. Because of the emergent nature of a segmented figure in our displays, feedback from higher cortical areas is a likely candidate for the selection mechanism by which the figure region is routed to lateral occipital cortex.Key words: visual cortex; object processing; figure/ground; cue invariance; lateral occipital complex; source imaging IntroductionObject recognition mechanisms must be able to extract shape independently of the surface cues that are present. Local estimates of surface cues such as texture grain or orientation, although necessary as inputs to the recognition process, convey little sense of object shape. Rather it is the pattern of cue similarity across regions and cue discontinuity across borders that must be integrated to recover object shape.The process of cue-invariant shape processing begins at an early stage of visual cortex and extends deep into extrastriate cortex and the temporal lobe. Cue invariance first is seen as early as V2, where some cells have a similar orientation or direction tuning for borders defined by different feature discontinuities (Leventhal et al., 1998;Marcar et al., 2000;Ramsden et al., 2001;Zhan and Baker, 2006). At higher levels of the visual system, such as inferotemporal cortex (Sary et al., 1993) and medial superior temporal area (Geesaman and Andersen, 1996), cells show shape selectivity that is mostly independent of the defining cues and spatial position. Functional magnetic resonance imaging (fMRI) studies in humans have implicated homologous extrastriate regions, in particular the lateral occipital complex (LOC), as sites of category-specific, cue-...
Heterogeneity of the posterior alpha rhythm (AR) is a widely assumed but rarely tested phenomenon. We decomposed the posterior AR in the cortical source space with a 3-way PARAFAC technique, taking into account the spatial, frequency, and temporal aspects of mid-density EEG. We found a multicomponent AR structure in 90% of a group of 29 healthy adults. The typical resting-state structure consisted of a high-frequency occipito-parietal component of the AR (ARC1) and a low-frequency occipito-temporal component (ARC2), characterized by individual dynamics in time. In a few cases, we found a 3-component structure, with two ARC1s and one ARC2. The AR structures were stable in their frequency and spatial features over weeks to months, thus representing individual EEG alpha phenotypes. Cortical topography, individual stability, and similarity to the primate AR organization link ARC1 to the dorsal visual stream and ARC2 to the ventral one. Understanding how many and what kind of posterior AR components contribute to the EEG is essential for clinical neuroscience as an objective basis for AR segmentation and for interpreting AR dynamics under various conditions, both normal and pathological, which can selectively affect individual components.
Discontinuities in feature maps serve as important cues for the location of object boundaries. Here we used multi-input nonlinear analysis methods and EEG source imaging to assess the role of several different boundary cues in visual scene segmentation. Synthetic figure/ground displays portraying a circular figure region were defined solely by differences in the temporal frequency of the figure and background regions in the limiting case and by the addition of orientation or relative alignment cues in other cases. The use of distinct temporal frequencies made it possible to separately record responses arising from each region and to characterize the nature of nonlinear interactions between the two regions as measured in a set of retinotopically and functionally defined cortical areas. Figure/background interactions were prominent in retinotopic areas, and in an extra-striate region lying dorsal and anterior to area MT+. Figure/background interaction was greatly diminished by the elimination of orientation cues, the introduction of small gaps between the two regions, or by the presence of a constant second-order border between regions. Nonlinear figure/background interactions therefore carry spatially precise, time-locked information about the continuity/discontinuity of oriented texture fields. This information is widely distributed throughout occipital areas, including areas that do not display strong retinotopy.
Symmetry is a highly salient feature of animals, plants, and the constructed environment. Although the perceptual phenomenology of symmetry processing is well understood, little is known about the underlying neural mechanisms. Here we use visual evoked potentials to measure the time course of neural events associated with the extraction of symmetry in random dot fields. We presented sparse random dot patterns that were symmetric about both the vertical and horizontal axes. Symmetric patterns were alternated with random patterns of the same density every 500 msec, using new exemplars of symmetric and random patterns on each image update. Random/random exchanges were used as a control. The response to updates of random patterns was multiphasic, consisting of P65, N90, P110, N140 and P220 peaks. The response to symmetric/random sequences was indistinguishable from that for random/random sequences up to about 220 msec, after which the response to symmetric patterns became relatively more negative. Symmetry in random dot patterns thus appears to be extracted after an initial response phase that is indifferent to configuration. These results are consistent with the hypothesis (Lee, Mumford, Romero, & Lamme, 1998; Tyler & Baseler, 1998) that the symmetry property is extracted by processing in extrastriate cortex.
Vision researchers rely on visual display technology for the presentation of stimuli to human and nonhuman observers. Verifying that the desired and displayed visual patterns match along dimensions such as luminance, spectrum, and spatial and temporal frequency is an essential part of developing controlled experiments. With cathode-ray tubes (CRTs) becoming virtually unavailable on the commercial market, it is useful to determine the characteristics of newly available displays based on organic light emitting diode (OLED) panels to determine how well they may serve to produce visual stimuli. This report describes a series of measurements summarizing the properties of images displayed on two commercially available OLED displays: the Sony Trimaster EL BVM-F250 and PVM-2541. The results show that the OLED displays have large contrast ratios, wide color gamuts, and precise, well-behaved temporal responses. Correct adjustment of the settings on both models produced luminance nonlinearities that were well predicted by a power function ("gamma correction"). Both displays have adjustable pixel independence and can be set to have little to no spatial pixel interactions. OLED displays appear to be a suitable, or even preferable, option for many vision research applications.
Glass patterns are a type of moiré created when a random-dot field is overlaid with a rotated, translated or dilated copy. The overall form of the moiré cannot be detected using local processing mechanisms, and because of this, Glass patterns are useful probes of global form processing. Here, we use event-related potentials to show that certain global organizations (concentric structure created by rotation and radial structure produced by dilation) produce much larger brain responses than others (linear structure created by translation). The results are consistent with the existence of specialized form processing mechanisms in the extrastriate cortex.
EEG has been central to investigations of the time course of various neural functions underpinning visual word recognition. Recently the steady-state visual evoked potential (SSVEP) paradigm has been increasingly adopted for word recognition studies due to its high signal-to-noise ratio. Such studies, however, have been typically framed around a single source in the left ventral occipitotemporal cortex (vOT). Here, we combine SSVEP recorded from 16 adult native English speakers with a data-driven spatial filtering approach—Reliable Components Analysis (RCA)—to elucidate distinct functional sources with overlapping yet separable time courses and topographies that emerge when contrasting words with pseudofont visual controls. The first component topography was maximal over left vOT regions with a shorter latency (approximately 180 ms). A second component was maximal over more dorsal parietal regions with a longer latency (approximately 260 ms). Both components consistently emerged across a range of parameter manipulations including changes in the spatial overlap between successive stimuli, and changes in both base and deviation frequency. We then contrasted word-in-nonword and word-in-pseudoword to test the hierarchical processing mechanisms underlying visual word recognition. Results suggest that these hierarchical contrasts fail to evoke a unitary component that might be reasonably associated with lexical access.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.