“…The effect is mostly due to the RT decrease for upright scenes, as no effect of SOA on RT was found for the rotated scenes. This RT reduction with upright scenes is expected at longer SOAs, because many choice models that incorporate temporal evidence accumulation (e.g., Hick, 1952;Perrett et al, 1998;Usher, Olami, & McClelland, 2002) predict shorter RTs when the decision process has more information available. Bacon-Mace, Mace, Fabre-Thorpe, and Thorpe (2005) found a similar effect by using masked presentations of natural scenes.…”
The visual system rapidly extracts information about objects from the cluttered natural environment. In 5 experiments, the authors quantified the influence of orientation and semantics on the classification speed of objects in natural scenes, particularly with regard to object-context interactions. Natural scene photographs were presented in an object-discrimination task and pattern masked with various scene-tomask stimulus-onset asynchronies (SOAs). Full psychometric functions and reaction times (RTs) were measured. The authors found that (a) rotating the full scenes increased threshold SOA at intermediate rotation angles but not for inversion; (b) rotating object or context degraded classification performance in a similar manner; (c) semantically congruent contexts had negligible facilitatory effects on object classification compared with meaningless baseline contexts with a matching contrast structure, but incongruent contexts severely degraded performance; (d) any object-context incongruence (orientation or semantic) increased RTs at longer SOAs, indicating dependent processing of object and context; and (e) facilitatory effects of context emerged only when the context shortly preceded the object. The authors conclude that the effects of natural scene context on object classification are primarily inhibitory and discuss possible reasons.
“…The effect is mostly due to the RT decrease for upright scenes, as no effect of SOA on RT was found for the rotated scenes. This RT reduction with upright scenes is expected at longer SOAs, because many choice models that incorporate temporal evidence accumulation (e.g., Hick, 1952;Perrett et al, 1998;Usher, Olami, & McClelland, 2002) predict shorter RTs when the decision process has more information available. Bacon-Mace, Mace, Fabre-Thorpe, and Thorpe (2005) found a similar effect by using masked presentations of natural scenes.…”
The visual system rapidly extracts information about objects from the cluttered natural environment. In 5 experiments, the authors quantified the influence of orientation and semantics on the classification speed of objects in natural scenes, particularly with regard to object-context interactions. Natural scene photographs were presented in an object-discrimination task and pattern masked with various scene-tomask stimulus-onset asynchronies (SOAs). Full psychometric functions and reaction times (RTs) were measured. The authors found that (a) rotating the full scenes increased threshold SOA at intermediate rotation angles but not for inversion; (b) rotating object or context degraded classification performance in a similar manner; (c) semantically congruent contexts had negligible facilitatory effects on object classification compared with meaningless baseline contexts with a matching contrast structure, but incongruent contexts severely degraded performance; (d) any object-context incongruence (orientation or semantic) increased RTs at longer SOAs, indicating dependent processing of object and context; and (e) facilitatory effects of context emerged only when the context shortly preceded the object. The authors conclude that the effects of natural scene context on object classification are primarily inhibitory and discuss possible reasons.
“…Critically, more recent computational instantiations of view interpolation have adopted more flexible representations of image features (Bricolo et al, 1997;Riesenhuber and Poggio, 1998) based on neurophysiological results that provide evidence for view-tuned neurons (Logothetis et al, 1995). Reinforcing the biological plausibility of this approach, Perrett et al (1998) offer specific neurophysiological evidence for 'evidence accumulation' across collections of local features, a mechanism similar to that proposed in some of the recent computational models.…”
Section: Interpolation Across Viewsmentioning
confidence: 85%
“…Indeed, proponents of the imagebased approach have offered a variety of different mechanisms for generalizing from unfamiliar to familiar views, including mental rotation (Tarr and Pinker, 1989), view interpolation (Poggio and Edelman, 1990) and linear combinations of views (Ullman and Basri, 1991). Even more sophisticated (Ullman, 1998) and neurally-plausible (Perrett et al, 1998) generalization mechanisms are presented in this volume.…”
“…Under a viewbased scheme neurons respond most strongly if objects are presented in learned views or configurations. Nevertheless, recognition of objects in varying orientations is thought possible by storing many views of an object Olshausen et al, A C C E P T E D M A N U S C R I P T ACCEPTED MANUSCRIPT 4 1993; Poggio and Edelman, 1990;Tarr and Gauthier, 1998;Tarr, 1995;Ullman, 1998), interpolating across these views (Logothetis et al, 1994;Poggio and Edelman, 1990;Ullman, 1989) or by a distributed neural representation across view-tuned neurons (Perrett et al, 1998).…”
A fundamental issue in visual cognition is whether high-level visual areas code objects in a part-based or a view-based (holistic) format. By examining the viewpoint invariance of object recognition, previous behavioral and neuroimaging studies have yielded ambiguous results, supporting both types of representational formats. A critical factor distinguishing the two formats could be the availability of attentional resources, as a number of studies have found greater viewpoint invariance for attended compared to unattended objects. It has therefore been suggested that attention is necessary to enable part-based representations, whereas holistic representations are automatically activated irrespective of attention. In this functional magnetic resonance imaging study we used a multivariate approach to probe the format of object representations in human lateral occipital complex (LOC) and its dependence on attention. We presented human participants with intact and half-split versions of objects that were either attended or unattended. Cross-classifying between intact and split objects, we found that the objectrelated information coded in activation patterns of intact objects is fully preserved in the patterns of split objects and vice versa. Importantly, the generalization between intact and split objects did not depend on attention. Our findings demonstrate that LOC codes objects in a non-holistic format, both in the presence and absence of attention.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.