Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1150
|View full text |Cite
|
Sign up to set email alerts
|

Towards Scalable and Reliable Capsule Networks for Challenging NLP Applications

Abstract: Obstacles hindering the development of capsule networks for challenging NLP applications include poor scalability to large output spaces and less reliable routing processes. In this paper, we introduce (i) an agreement score to evaluate the performance of routing processes at instance level; (ii) an adaptive optimizer to enhance the reliability of routing; (iii) capsule compression and partial routing to improve the scalability of capsule networks. We validate our approach on two NLP tasks, namely: multi-label… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
49
0
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 118 publications
(51 citation statements)
references
References 40 publications
0
49
0
1
Order By: Relevance
“…For example, Tang et al [58] used a convolutional neural network (CNN) to obtain word embeddings for words frequently used in tweets and dos Santos and Gatti [17] employed a deep CNN for sentiment detection in short texts. More recent approaches have been focusing on the development of sentiment-specific word embeddings [44], which are able to encode more affective clues than regular word vectors, and on the use of context-aware subsymbolic approaches such as attention modeling [32,33] and capsule networks [13,66].…”
Section: Related Workmentioning
confidence: 99%
“…For example, Tang et al [58] used a convolutional neural network (CNN) to obtain word embeddings for words frequently used in tweets and dos Santos and Gatti [17] employed a deep CNN for sentiment detection in short texts. More recent approaches have been focusing on the development of sentiment-specific word embeddings [44], which are able to encode more affective clues than regular word vectors, and on the use of context-aware subsymbolic approaches such as attention modeling [32,33] and capsule networks [13,66].…”
Section: Related Workmentioning
confidence: 99%
“…Recently, deep learning-based aspect extraction [52], attention-based LSTM [53] and capsule networks [54,55] have also been widely used for sentiment analysis and challenging NLP applications that yield state-of-the-art prediction results. In this work, our focus is on the design of a unified sentiment analysis framework for multi-domain textual sentiment classification.…”
Section: Related Workmentioning
confidence: 99%
“…The output of capsule networks can serve as conventional probability of hypothesis while pose vector can bring specific interpretable information as auxiliary evidence to final decision. Also, since the extrapolation during affine transformation and generative routing method, capsule network can be learned with many fewer samples and prevail deep models [9], [10].…”
Section: Capsule Networkmentioning
confidence: 99%
“…4. Here we compared three methods: XML-CNN [14], NLPCap [9] and our proposed NLP-Cap with perturbed samples (NLP-Cap-Ge). The performances at two standard rank-based tasks, Precision@k and Normalized Discounted Cumulative Gain (NDCG@k), are considered.…”
Section: A Text Classification With Fewer Samplesmentioning
confidence: 99%
See 1 more Smart Citation