Mobile app stores produce a tremendous amount of data in the form of user reviews, which is a huge source of user requirements and sentiments; such reviews allow app developers to proactively address issues in their apps. However, only a small number of reviews capture common issues and sentiments which creates a need for automatically identifying prominent reviews. Unfortunately, most existing work in text ranking and popularity prediction focuses on social contexts where other signals are available, which renders such works ineffective in the context of app reviews. In this work, we propose a new framework, PPrior, that enables proactive prioritization of app issues through identifying prominent reviews (ones predicted to receive a large number of votes in a given time window). Predicting highlyvoted reviews is challenging given that, unlike social posts, social network features of users are not available. Moreover, there is an issue of class imbalance, since a large number of user reviews receive little to no votes. PPrior employs a pre-trained T5 model and works in three phases. Phase one adapts the pretrained T5 model to the user reviews data in a self-supervised fashion. In phase two, we leverage contrastive training to learn a generic and task-independent representation of user reviews. Phase three uses radius neighbors classifier to make the final predictions. This phase also uses FAISS index for scalability and efficient search. To conduct extensive experiments, we acquired a large dataset of over 2.1 million user reviews from Google Play. Our experimental results demonstrate the effectiveness of the proposed framework when compared against several state-of-theart approaches. Moreover, the accuracy of PPrior in predicting prominent reviews is comparable to that of experienced app developers.
Slot filling is one of the critical tasks in modern conversational systems. The majority of existing literature employs supervised learning methods, which require labeled training data for each new domain. Zero-shot learning and weak supervision approaches, among others, have shown promise as alternatives to manual labeling. Nonetheless, these learning paradigms are significantly inferior to supervised learning approaches in terms of performance. To minimize this performance gap and demonstrate the possibility of open-domain slot filling, we propose a Self-supervised Co-training framework, called SCot, that requires zero in-domain manually labeled training examples and works in three phases. Phase one acquires two sets of complementary pseudo labels automatically.Phase two leverages the power of the pre-trained language model BERT, by adapting it for the slot filling task using these sets of pseudo labels. In phase three, we introduce a self-supervised cotraining mechanism, where both models automatically select highconfidence soft labels to further improve the performance of the other in an iterative fashion. Our thorough evaluations show that SCot outperforms state-of-the-art models by 45.57% and 37.56% on SGD and MultiWoZ datasets, respectively. Moreover, our proposed framework SCot achieves comparable performance when compared to state-of-the-art fully supervised models.
Task-oriented dialog systems empower users to accom-plish their goals by facilitating intuitive and expres-sive natural language interactions. State-of-the-art ap-proaches in task-oriented dialog systems formulate theproblem as a conditional sequence generation task andfine-tune pre-trained causal language models in the su-pervised setting. This requires labeled training datafor each new domain or task, and acquiring such datais prohibitively laborious and expensive, thus makingit a bottleneck for scaling systems to a wide rangeof domains. To overcome this challenge, we intro-duce a novel Zero-Shot generalizable end-to-end Task-oriented Dialog system, ZS-ToD, that leverages domainschemas to allow for robust generalization to unseen do-mains and exploits effective summarization of the dia-log history. We employ GPT-2 as a backbone model andintroduce a two-step training process where the goal ofthe first step is to learn the general structure of the dialogdata and the second step optimizes the response gen-eration as well as intermediate outputs, such as dialogstate and system actions. As opposed to state-of-the-artsystems that are trained to fulfill certain intents in thegiven domains and memorize task-specific conversa-tional patterns, ZS-ToD learns generic task-completionskills by comprehending domain semantics via domainschemas and generalizing to unseen domains seam-lessly. We conduct an extensive experimental evaluationon SGD and SGD-X datasets that span up to 20 uniquedomains and ZS-ToD outperforms state-of-the-art sys-tems on key metrics, with an improvement of +17% onjoint goal accuracy and +5 on inform. Additionally,we present a detailed ablation study to demonstrate theeffectiveness of the proposed components and trainingmechanism.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.