Abstract.In an experimental environment, we simulated the situation of a user who gives speech input to a system while walking through an airport. The time pressure on the subjects and the requirement to navigate while speaking were manipulated orthogonally. Each of the 32 subjects generated 80 utterances, which were coded semi-automatically with respect to a wide range of features, such as filled pauses. The experiment yielded new results concerning the effects of time pressure and cognitive load on speech. To see whether a system can automatically identify these conditions on the basis of speech input, we had this task performed for each subject by a Bayesian network that had been learned on the basis of the experimental data for the other subjects. The results shed light on the conditions that determine the accuracy of such recognition.
Abstract.One of the central questions addressed in the project READY was that of how a system can automatically recognize situationally determined resource limitations of its user-in particular, time pressure and cognitive load. This chapter summarizes most of the work done in READY on this topic, presenting as well some previously unpublished results. We first consider why on-line recognition or resource limitations can be useful by discussing the ways in which a system might adapt its behavior to perceived resource limitations. We then summarize a number of approaches to the recognition problem that have been taken in READY and other projects, before focusing on one particular approach: the analysis of features of a user's speech. In each of two similarly structured experiments, we created four experimental conditions that varied in terms of whether the user was (a) required to produce spoken utterances quickly or not; and (b) navigating within a simulated airport terminal or standing still. In the second experiment, additional distraction was caused by continuous loudspeaker announcements. The speech produced by the experimental subjects (32 in each experiment) was coded in terms of 7 variables. We report on the extent to which each of these variables was influenced by the subjects' resource limitations. We also trained dynamic Bayesian networks on the resulting data in order to see how well the information in the users' speech could serve as evidence as to which condition the user had been in. The results yield information about the accuracy that can be attained in this way and about the diagnostic value of some specific features of speech.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.