The audiovisual stream-bounce effect refers to the resolution of ambiguous motion sequences as streaming or bouncing depending on the presence or absence of a sound. We used a novel experimental design and signal detection theory (SDT) to determine its sensory or decisional origins. To account for issues raised by Witt et al. on the interpretation of SDT results, we devised a pure signal detection (as opposed to signal discrimination) paradigm and measured participants' sensitivity and criterion when detecting a weak tone concurrent with objectively streaming or bouncing visual displays. We observed no change in sensitivity but a significant change in criterion with participants' criterion more liberal with bouncing targets than for streaming targets with. In a second experiment, we tasked participants with detecting a weak tone in noise while viewing an ambiguous motion sequence. They also indicated whether the targets appeared to stream or bounce. Participants' reported equivalent, mostly bouncing responses for hit and false alarm trials, and equivalent, mostly streaming responses for correct rejection and miss trials. Further, differences in participants' sensitivity and criterion measures for detecting tones in subjectively streaming compared to subjectively bouncing targets were inconsistent with sensory factors. These results support a decisional account of the sound-induced switch from mostly streaming to mostly bouncing responses in audiovisual stream-bounce displays.
We tested the sensory versus decisional origins of two established audiovisual crossmodal correspondences (CMCs; lightness/pitch and elevation/pitch), applying a signal discrimination paradigm to low-level stimulus features and controlling for attentional cueing. An audiovisual stimulus randomly varied along two visual dimensions (lightness: black/white; elevation: high/low) and one auditory dimension (pitch: high/low), and participants discriminated either only lightness, only elevation, or both lightness and elevation. The discrimination task and the stimulus duration varied between subjects. To investigate the influence of crossmodal congruency, we considered the effect of each CMC (lightness/pitch and elevation/pitch) on the sensitivity and criterion of each discrimination as a function of stimulus duration. There were three main findings. First, discrimination sensitivity was significantly higher for visual targets paired congruently (compared with incongruently) with tones while criterion was unaffected. Second, the sensitivity increase occurred for all stimulus durations, ruling out attentional cueing effects. Third, the sensitivity increase was feature specific such that only the CMC that related to the feature being discriminated influenced sensitivity (i.e. lightness congruency only influenced lightness discrimination and elevation congruency only influenced elevation discrimination in the single and dual task conditions). We suggest that these congruency effects reflect low-level sensory processes.
The stream-bounce effect refers to a bistable motion stimulus that is interpreted as two targets either ''streaming'' past or ''bouncing'' off one another, and the manipulations that bias responses. Directional bias, according to Bertenthal et al., is an account of the effect proposing that low-level motion integration promotes streaming, and its disruption leads to bouncing, and it is sometimes cited either directly in a bottom-up fashion or indirectly under top-down control despite Sekuler and Sekuler finding evidence inconsistent with it. We tested two key aspects of the hypothesis: (a) comparable changes in speed should produce comparable disruptions and lead to similar effects; and (b) speed changes alone should disrupt integration without the need for additional more complex changes of motion. We found that target motion influences stream-bounce perception, but not as directional bias predicts. Our results support Sekuler and Sekuler and argue against the low-level motion signals driving perceptual outcomes in stream-bounce displays (directly or indirectly) and point to higher level inferential processes involving perceptual history and expectation. Directional bias as a mechanism should be abandoned and either another specific bottom-up process must be proposed and tested or consideration should be given to top-down factors alone driving the effect.
We examine perceptual disambiguation and crossmodal interactions by considering the effect of recent perceptual history on stream-bounce perception. First, we tested the assumption that the audio-visual stream-bounce effect (visual-only trials mostly stream, whereas audio-visual trials mostly bounce) reflects some intrinsic preference for streaming that is broken by sound. Instead, we found that for naïve observers, visual-only stimuli are bistable and bias free. In intermixed trials, sound acts as a polarizing factor (rather than a bounce-inducing factor) and has as much effect on visual-only trials as it does on audio-visual trials. Second, temporal context exerts a comparative influence and exposure to audio-visual stimuli influences responses to visual-only stimuli and vice-versa. Finally, there is a serial dependence in responses and the current stimulus (unisensory or multisensory) is interpreted with a bias to recent interpretations. Recent perceptual history exerts a substantial influence on the perception of stream-bounce stimuli. This influence occurs in the unisensory case and is in line with an extensive literature on visual bistability; it extends to the multisensory case, and there are interactions between the 2 cases. Our combined findings support a role for top-down interpretational influences in the streambounce effect and stream-bounce perception more generally. Public Significance StatementAlthough sensory information is typically ambiguous, our perception is not. This study supports the idea that in dealing with ambiguity, the perceptual system analyses recent perceptual history across the sensory modalities to predict the immediate sensory future and these predictions alter what is perceived. Given the requirement on the brain to provide an informative, clear, and stable representation of a complex and ever-changing outside world, and the computational burden this involves, making heavy use of experience is efficient and adaptive.
Previous research suggests that cognitive factors acting in a top-down manner influence the perceptual interpretation of ambiguous stimuli. To examine the temporal unfolding of these influences as a perceptual decision evolves, we have implemented a modified version of the stream-bounce display. Our novel approach allows us to track responses to stream-bounce stimuli dynamically over the entire course of the motion sequence rather than collecting a subjective report after the fact. Using a trackpad, we had participants control a cursor to track a stream-bounce target actively from start to end and measured tracking speed throughout as the dependent variable. Our paradigm replicated the typical effect of visual-only displays being associated with a streaming bias and audiovisual displays with a bouncing bias. Our main finding is a significant behavioral change preceding a perceptual decision that then predicts that decision. Specifically, for trials in which the sound was presented, tracking speeds were significantly slower starting 500 ms before the point of coincidence and presentation of the sound for bounce compared to stream responses. We suggest that behavioral response may reflect a cognitive expectation of a perceptual outcome that then biases action and the interpretation of sensory input to favor that forthcoming percept in a manner consistent with both the predictivecoding and common-coding theoretical frameworks. Our approach provides a novel behavioral corroboration of recent imaging studies that are suggestive of early brain activity in perception and action. DataOnly data from test blocks (both objective and subjective motion) were recorded for analysis.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.