In the past decade, cultural differences in perceptual judgment and memory have been observed: Westerners attend more to focal objects, whereas East Asians attend more to contextual information. However, the underlying mechanisms for the apparent differences in cognitive processing styles have not been known. In the present study, we examined the possibility that the cultural differences arise from culturally different viewing patterns when confronted with a naturalistic scene. We measured the eye movements of American and Chinese participants while they viewed photographs with a focal object on a complex background. In fact, the Americans fixated more on focal objects than did the Chinese, and the Americans tended to look at the focal object more quickly. In addition, the Chinese made more saccades to the background than did the Americans. Thus, it appears that differences in judgment and memory may have their origins in differences in what is actually attended as people view a scene.attention ͉ culture ͉ memory ͉ eye-tracking ͉ visual cognition
This paper describes the application of the PARADISE evaluation framework to the corpus of 662 human-computer dialogues collected in the June 2000 Darpa Communicator data collection. We describe results based on the standard logfile metrics as well as results based on additional qualitative metrics derived using the DATE dialogue act tagging scheme. We show that performance models derived via using the standard metrics can account for 37% of the variance in user satisfaction, and that the addition of DATE metrics improved the models by an absolute 5%.
The perception of coarticulated speech as it unfolds over time was investigated by monitoring eye movements of participants as they listened to words with oral vowels or with late or early onset of anticipatory vowel nasalization. When listeners heard [CṼNC] and had visual choices of images of CVNC (e.g., send) and CVC (said) words, they fixated more quickly and more often on the CVNC image when onset of nasalization began early in the vowel compared to when the coarticulatory information occurred later. Moreover, when a standard eye movement programming delay is factored in, fixations on the CVNC image began to occur before listeners heard the nasal consonant. Listeners' attention to coarticulatory cues for velum lowering was selective in two respects: (a) listeners assigned greater perceptual weight to coarticulatory information in phonetic contexts in which [Ṽ] but not N is an especially robust property, and (b) individual listeners differed in their perceptual weights. Overall, the time course of perception of velum lowering in American English indicates that the dynamics of perception parallel the dynamics of the gestural information encoded in the acoustic signal. In real-time processing, listeners closely track unfolding coarticulatory information in ways that speed lexical activation.
Three experiments investigated the use of verb argument structure by tracking participants' eye movements across a set of related pictures as they listened to sentences. The assumption was that listeners would naturally look at relevant pictures as they were mentioned or implied. The primary hypothesis was that a verb would implicitly introduce relevant entities (linguistic arguments) that had not yet been mentioned, and thus a picture corresponding to such an entity would draw anticipatory looks. For example, upon hearing ...mother suggested..., participants would look at a potential recipient of the suggestion. The only explicit task was responding to comprehension questions. Experiments 1 and 2 manipulated both the argument structure of the verb and the typicality/co-occurrence frequency of the target argument/adjunct, in order to distinguish between anticipatory looks to arguments specifically and anticipatory looks to pictures that were strongly associated with the verb, but did not have the linguistic status of argument. Experiment 3 manipulated argument status alone. In Experiments 1 and 3, there were more anticipatory looks to potential arguments than to potential adjuncts, beginning about 500 ms after the acoustic onset of the verb. Experiment 2 revealed a main effect of typicality. These findings indicate that both real world knowledge and argument structure guide visual attention within this paradigm, but that argument structure has a privileged status in focusing listener attention on relevant aspects of a visual scene.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.