Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
Computational models and metrics, including measures like surprisal and semantic relevance, have been developed to accurately predict and explain language comprehension and processing. However, their efficacy is hindered by their inadequate integration of contextual information. Drawing inspiration from the attention mechanism in transformers and human forgetting mechanism, this study introduces an attention-aware method that thoroughly incorporates contextual information, updating surprisal and semantic relevance into attention-aware metrics respectively. Furthermore, by employing the quantum superposition principle, the study proposes an enhanced approach for integrating and encoding diverse information sources based on the two attention-aware metrics. Both attention-aware and enhanced metrics demonstrate superior effectiveness in comparison to the currently available metrics, leading to improved predictions of eye-movements during naturalistic discourse reading across 13 languages. The proposed approaches are fairly capable of facilitating simulation and evaluation of existing reading models and language processing theories. The metrics computed by the proposed approaches are highly interpretable and exhibit cross-language generalizations in predicting language comprehension. The innovative computational methods proposed in this study hold the great potential to enhance our under-standing of human working memory mechanisms, human reading behavior and cognitive modeling in language processing. Moreover, they have the capacity to revolutionize ongoing research in computational cognition for language processing, offering valuable insights for computational neuroscience, quantum cognition and optimizing the design of AI systems.
Computational models and metrics, including measures like surprisal and semantic relevance, have been developed to accurately predict and explain language comprehension and processing. However, their efficacy is hindered by their inadequate integration of contextual information. Drawing inspiration from the attention mechanism in transformers and human forgetting mechanism, this study introduces an attention-aware method that thoroughly incorporates contextual information, updating surprisal and semantic relevance into attention-aware metrics respectively. Furthermore, by employing the quantum superposition principle, the study proposes an enhanced approach for integrating and encoding diverse information sources based on the two attention-aware metrics. Both attention-aware and enhanced metrics demonstrate superior effectiveness in comparison to the currently available metrics, leading to improved predictions of eye-movements during naturalistic discourse reading across 13 languages. The proposed approaches are fairly capable of facilitating simulation and evaluation of existing reading models and language processing theories. The metrics computed by the proposed approaches are highly interpretable and exhibit cross-language generalizations in predicting language comprehension. The innovative computational methods proposed in this study hold the great potential to enhance our under-standing of human working memory mechanisms, human reading behavior and cognitive modeling in language processing. Moreover, they have the capacity to revolutionize ongoing research in computational cognition for language processing, offering valuable insights for computational neuroscience, quantum cognition and optimizing the design of AI systems.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.