Peripheral vision provides a less faithful representation of the visual input than foveal vision. Nonetheless, we can gain a lot of information about the world from our peripheral vision, for example in order to plan eye movements. The phenomenon of crowding shows that the reduction of information available in the periphery is not merely the result of reduced resolution. Crowding refers to visual phenomena in which identification of a target stimulus is significantly impaired by the presence of nearby stimuli, or flankers. What information is available in the periphery? We propose that the visual system locally represents peripheral stimuli by the joint statistics of responses of cells sensitive to different position, phase, orientation, and scale. This “textural” representation by summary statistics predicts the subjective “jumble” of features often associated with crowding. We show that the difficulty of performing an identification task within a single pooling region using this representation of the stimuli is correlated with peripheral identification performance under conditions of crowding. Furthermore, for a simple stimulus with no flankers, this representation can be adequate to specify the stimulus with some position invariance. This provides evidence that a unified neuronal mechanism may underlie peripheral vision, ordinary pattern recognition in central vision, and texture perception. A key component of our methodology involves creating visualizations of the information available in the summary statistics of a stimulus. We call these visualizations “mongrels” and show that they are highly useful in examining how the early visual system represents the visual input. Mongrels enable one to study the “equivalence classes” of our model, i.e., the sets of stimuli that map to the same representation according to the model.
Infant face processing becomes more selective during the first year of life as a function of varying experience with distinct face categories defined by species, race, and age. Given that any individual face belongs to many such categories (e.g. A young Caucasian man's face) we asked how the neural selectivity for one aspect of facial appearance was affected by category membership along another dimension of variability. 6-month-old infants were shown upright and inverted pictures of either their own mother or a stranger while event-related potentials (ERPs) were recorded. We found that the amplitude of the P400 (a face-sensitive ERP component) was only sensitive to the orientation of the mother's face, suggesting that “tuning” of the neural response to faces is realized jointly across multiple dimensions of face appearance.
Vision is an active process: we repeatedly move our eyes to seek out objects of interest and explore our environment. Visual search experiments capture aspects of this process, by having subjects look for a target within a background of distractors. Search speed often correlates with target-distractor discriminability; search is faster when the target and distractors look quite different. However, there are notable exceptions. A given discriminability can yield efficient searches (where the target seems to “pop-out”) as well as inefficient ones (where additional distractors make search significantly slower and more difficult). Search is often more difficult when finding the target requires distinguishing a particular configuration or conjunction of features. Search asymmetries abound. These puzzling results have fueled three decades of theoretical and experimental studies. We argue that the key issue in search is the processing of image patches in the periphery, where visual representation is characterized by summary statistics computed over a sizable pooling region. By quantifying these statistics, we predict a set of classic search results, as well as peripheral discriminability of crowded patches such as those found in search displays.
Traditionally, texture perception has been studied using artificial textures made of random dots or repeated shapes. At the same time, computer algorithms for natural texture synthesis have improved dramatically. We seek to unify these two fields through a psychophysical assessment of a particular computational model, providing insight into which statistics are most vital for natural texture perception. We employ Portilla and Simoncelli's texture synthesis algorithm, a parametric model that mimics computations carried out in human vision. We find an intriguing interaction between texture type (periodic, structured, or 3-D textures) and image statistics (autocorrelation function and filter magnitude correlations), suggesting different representations may be employed for these texture families under pre-attentive viewing.
Over the last ten years, Oosterhof and Todorov's valence-dominance model has emerged as the most prominent account of how people evaluate faces on social dimensions. In this model, two dimensions (valence and dominance) underpin social judgments of faces. Because this model has primarily been developed and tested in Western regions, it is unclear whether these findings apply to other regions. We addressed this question by replicating Oosterhof and Todorov's methodology across 11 world regions, 41 countries, and 11,570 participants. When we used Oosterhof and Todorov's original analysis strategy, the valence-dominance model generalized across regions. When we used an alternative methodology to allow for correlated dimensions we observed much less generalization. Collectively, these results suggest that, while the valence-dominance model generalizes very well across regions when dimensions are forced to be orthogonal, regional differences are revealed when we use different extraction methods, correlate and rotate the dimension reduction solution.
How does the remarkable human ability for face recognition arise over development? Competing theories have proposed either late maturity (beyond 10 years) or early maturity (before 5 years), but have not distinguished between perceptual and memory aspects of face recognition. Here, we demonstrate a perception-memory dissociation. We compare rate of development for (adult, human) faces versus other social stimuli (bodies), other discrete objects (cars), and other categories processed in discrete brain regions (scenes, bodies), from 5 years to adulthood. For perceptual discrimination, performance improved with age at the same rate for faces and all other categories, indicating no domain-specific development. In contrast, face memory increased more strongly than non-face memory, indicating domain-specific development. The results imply that each theory is partly true: the late maturity theory holds for face memory, and the early maturity theory for face perception.
Adult observers generally find it difficult to recognize and distinguish faces that belong to categories with which they have limited visual experience. One aspect of this phenomenon is commonly known as the "Other-Race Effect" (ORE) since this behavior is typically highly evident in the perception of faces belonging to ethnic or racial groups other than that of the observer. This acquired disadvantage in face recognition likely results from highly specific "tuning" of the underlying representation of facial appearance, leading to efficient processing of commonly-seen faces at the expense of poor generalization to other face categories. In the current study we used electrophysiological (eventrelated potentials or ERPs) and behavioral measures of performance to characterize face processing in racial categories defined by dissociable shape and pigmentation information. Our goal was to examine the specificity of the representation of facial appearance in more detail by investigating how race-specific face shape and pigmentation separately modulated neural responses previously implicated in face processing, the N170 and N250 components. We found that both components were modulated by skin color, independent of face shape, but that only the N250 exhibited sensitivity to face shape. Moreover, the N250 appears to only respond differentially to the skin color of upright faces, showing a lack of color sensitivity for inverted faces.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.