Relation classification is an important semantic processing task in the field of natural language processing (NLP). State-ofthe-art systems still rely on lexical resources such as WordNet or NLP systems like dependency parser and named entity recognizers (NER) to get high-level features. Another challenge is that important information can appear at any position in the sentence. To tackle these problems, we propose Attention-Based Bidirectional Long Short-Term Memory Networks(Att-BLSTM) to capture the most important semantic information in a sentence. The experimental results on the SemEval-2010 relation classification task show that our method outperforms most of the existing methods, with only word vectors.
Implicit discourse relation classification is one of the most difficult tasks in discourse parsing. Previous studies have generally focused on extracting better representations of the relational arguments. In order to solve the task, it is however additionally necessary to capture what events are expected to cause or follow each other. Current discourse relation classifiers fall short in this respect. We here show that this shortcoming can be effectively addressed by using the bidirectional encoder representation from transformers (BERT) proposed by Devlin et al. (2019), which were trained on a nextsentence prediction task, and thus encode a representation of likely next sentences. The BERT-based model outperforms the current state of the art in 11-way classification by 8% points on the standard PDTB dataset. Our experiments also demonstrate that the model can be successfully ported to other domains: on the BioDRB dataset, the model outperforms the state of the art system around 15% points.
Due to the structural problem, the traditional neural network models are prone to problems such as gradient explosion and over-fitting, while the deep GRU neural network model has low update efficiency and poor information processing capability among multiple hidden layers. Based on this, this paper proposes an optimized gated recurrent unit(OGRU) neural network.The OGRU neural network model proposed in this paper improves information processing capability and learning efficiency by optimizing the unit structure and learning mechanism of GRU, and avoids the update gate being interfered by the current forgetting information. The experiment uses Tensorflow framework to establish prediction models for LSTM neural network, GRU neural network and OGRU neural network respectively, and compare the prediction accuracy. The results show that the OGRU model has the highest learning efficiency and better prediction accuracy.
Implicit discourse relation classification is one of the most difficult steps in discourse parsing. The difficulty stems from the fact that the coherence relation must be inferred based on the content of the discourse relational arguments. Therefore, an effective encoding of the relational arguments is of crucial importance. We here propose a new model for implicit discourse relation classification, which consists of a classifier, and a sequence-to-sequence model which is trained to generate a representation of the discourse relational arguments by trying to predict the relational arguments including a suitable implicit connective. Training is possible because such implicit connectives have been annotated as part of the PDTB corpus. Along with a memory network, our model could generate more refined representations for the task. And on the now standard 11-way classification, our method outperforms the previous state of the art systems on the PDTB benchmark on multiple settings including cross validation.
Automatically generating globally coherent stories is a challenging problem. Neural text generation models have been shown to perform well at generating fluent sentences from data, but they usually fail to keep track of the overall coherence of the story after a couple of sentences. Existing work that incorporates a text planning module succeeded in generating recipes and dialogues, but appears quite datademanding. We propose a novel story generation approach that generates globally coherent stories from a fairly small corpus. The model exploits a symbolic text planning module to produce text plans, thus reducing the demand of data; a neural surface realization module then generates fluent text conditioned on the text plan. Human evaluation showed that our model outperforms various baselines by a wide margin and generates stories which are fluent as well as globally coherent.
ABSTRACT:At present, 87% of people's activities are in indoor environment; indoor navigation has become a research issue. As the building structures for people's daily life are more and more complex, many obstacles influence humans' moving. Therefore it is essential to provide an accurate and efficient indoor path planning. Nowadays there are many challenges and problems in indoor navigation. Most existing path planning approaches are based on 2D plans, pay more attention to the geometric configuration of indoor space, often ignore rich semantic information of building components, and mostly consider simple indoor layout without taking into account the furniture. Addressing the above shortcomings, this paper uses BIM (IFC) as the input data and concentrates on indoor navigation considering obstacles in the multi-floor buildings. After geometric and semantic information are extracted, 2D and 3D space subdivision methods are adopted to build the indoor navigation network and to realize a path planning that avoids obstacles. The 3D space subdivision is based on triangular prism. The two approaches are verified by the experiments.
In recent years, network embedding has attracted more and more attention due to its effectiveness and convenience to compress the network structured data. In this paper, we propose a communitybased variational autoencoder (ComVAE) model to learn network embedding, which consists of a community detection module and a deep learning module. In the proposed model, both community information and deep learning techniques are utilized to learn low-dimensional vertex representations. First, community information reveals an implicit relationship between vertices from a global view, which can be a supplement to local information and help to improve the embedding quality. To obtain the community information, community detection algorithms are utilized as a module and the modularization design makes the model more flexible. Second, deep learning techniques can not only integrate and preserve the information from both local and global views efficiently but also strengthen the robustness of vertex representations. To demonstrate the performance of our model, extensive experiments are conducted in four downstream tasks, namely, network reconstruction, node classification, link prediction, and visualization. The experimental results show that our model outperforms the state-of-the-art approaches to real-world datasets. INDEX TERMS Network embedding, community detection, variational autoencoder.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.