Many visual displays, such as movies and television, rely on sampling in the time domain. We derive the spatiotemporal-frequency spectra for some simple moving images and illustrate how these spectra are altered by sampling in the time domain. We construct a simple model of the human perceiver that predicts the critical sample rate required to render sampled and continuous moving images indistinguishable. The rate is shown to depend on the spatial and the temporal acuity of the observer and on the velocity and spatial-frequency content of the image. Several predictions of this model are tested and confirmed. The model is offered as an explanation of many of the phenomena known as apparent motion. Finally, the implications of the model for computer-generated imagery are discussed.
Sequential alternation between same-shaped stimuli differing in size (size ratio s) and orientation (angular difference v) produced a visual illusion of translation in depth and concurrent rotation. The minimum stimulus-onset asynchrony required for the appearance of a rigidly moving object was approximately a linearly increasing function of (s-1)/(s+1) for simple translation in depth and a linearly increasing function of v for simple rotation. The extrapolated zero intercept was lower for translation than for rotation, but estimated transformation times were additive in combined transformations. The results suggest that (a) the processes of apparent translation in depth and apparent rotation are individually sequential-additive in structure, and (b) apparent translations and rotations are combined by fine-grained alternation of steps of apparent translation and steps of apparent rotation. Similar principles account for recent data on imagined spatial transformations of visual size and orientation.
Apparent rotational motion was investigated in polygonal shapes ranging in rotational symmetry from random to self-identical under 180 degrees rotation. Observers adjusted the rate of alternation between two computer-displayed orientations of any given polygon to determine the point of breakdown of perceived rigid rotation between those two orientations. For asymmetric polygons, the minimum stimulus-onset asynchrony yielding apparent rigid rotation increased approximately linearly with orientational disparity, as anticipated on the basis of the extension of Korte's third law to rotational motion by Shepard and Judd. For nearly symmetric polygons, however, the critical time increased markedly as the disparities approached 180 degrees, owing to the availability of a shorter, nonrigid rotation in the opposite direction. The results demonstrate the existence of competing mental tendencies to preserve the rigid structure of an object and to traverse a minimum transformational path.
We describe a simulation of the complete image processing pipeline of a digital camera, beginning with a radiometric description of the scene captured by the camera and ending with a radiometric description of the image rendered on a display. We show that there is a good correspondence between measured and simulated sensor performance. Through the use of simulation, we can quantify the effects of individual digital camera components on system performance and image quality. This computational approach can be helpful for both camera design and image quality assessment.
The Image Systems Evaluation Toolkit (ISET) is an integrated suite of software routines that simulate the capture and processing of visual scenes. ISET includes a graphical user interface (GUI) for users to control the physical characteristics of the scene and many parameters of the optics, sensor electronics and image processing-pipeline. ISET also includes color tools and metrics based on international standards (chromaticity coordinates, CIELAB and others) that assist the engineer in evaluating the color accuracy and quality of the rendered image.
A single line was presented in a succession of orientations, each orientation separated by a fixed angle and by a fixed interval of time, and subjects reported the number of successive lines that appeared to rotate together. The perceived number of rotating lines increased linearly with the rate of stimulus presentation, with a slope that was proportional to the spatial separation. The linear functions obtained in this first experiment predicted the results of a second experiment in which subjects adjusted the spatial and temporal variables to a discrimination threshold for seeing two rotating lines. If the slope of the linear functions is considered to be an estimate of the duration of visible persistence, then these results suggest that the visible persistence of a briefly presented stimulus increases with the distance separating that stimulus from other stimuli.
Vision researchers rely on visual display technology for the presentation of stimuli to human and nonhuman observers. Verifying that the desired and displayed visual patterns match along dimensions such as luminance, spectrum, and spatial and temporal frequency is an essential part of developing controlled experiments. With cathode-ray tubes (CRTs) becoming virtually unavailable on the commercial market, it is useful to determine the characteristics of newly available displays based on organic light emitting diode (OLED) panels to determine how well they may serve to produce visual stimuli. This report describes a series of measurements summarizing the properties of images displayed on two commercially available OLED displays: the Sony Trimaster EL BVM-F250 and PVM-2541. The results show that the OLED displays have large contrast ratios, wide color gamuts, and precise, well-behaved temporal responses. Correct adjustment of the settings on both models produced luminance nonlinearities that were well predicted by a power function ("gamma correction"). Both displays have adjustable pixel independence and can be set to have little to no spatial pixel interactions. OLED displays appear to be a suitable, or even preferable, option for many vision research applications.
Abstract-This paper describes the design and performance of an image capture simulator. The general model underlying the simulator assumes that the image capture device contains multiple classes of sensors with different spectral sensitivities and that each sensor responds in a known way to irradiance over most of its operating range. The input to the simulator is a set of narrow-band images of the scene taken with a custom-designed hyperspectral camera system. The parameters for the simulator are the number of sensor classes, the sensor spectral sensitivities, the noise statistics and number of quantization levels for each sensor class, the spatial arrangement of the sensors and the exposure duration. The output of the simulator is the raw image data that would have been acquired by the simulated image capture device.To test the simulator, we acquired images of the same scene both with the hyperspectral camera and with a calibrated Kodak DCS-200 digital color camera. We used the simulator to predict the DCS-200 output from the hyperspectral data. The agreement between simulated and acquired images validated the image capture response model and our simulator implementation. We believe the simulator will provide a useful tool for understanding the effect of varying the design parameters of an image capture device.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.