2014
DOI: 10.1007/978-3-662-44851-9_28
|View full text |Cite
|
Sign up to set email alerts
|

Large-Scale Multi-label Text Classification — Revisiting Neural Networks

Abstract: Abstract. Neural networks have recently been proposed for multi-label classification because they are able to capture and model label dependencies in the output layer. In this work, we investigate limitations of BP-MLL, a neural network (NN) architecture that aims at minimizing pairwise ranking error. Instead, we propose to use a comparably simple NN approach with recently proposed learning techniques for large-scale multi-label text classification tasks. In particular, we show that BP-MLL's ranking loss minim… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
224
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 294 publications
(248 citation statements)
references
References 23 publications
2
224
0
Order By: Relevance
“…The best performing network is the 2L-CNN with randomly initialized embeddings. The resulting F-measure is comparable to the value of 87.89% presented in (Nam et al, 2014).…”
Section: Results On the English Reuters Datasetsupporting
confidence: 78%
“…The best performing network is the 2L-CNN with randomly initialized embeddings. The resulting F-measure is comparable to the value of 87.89% presented in (Nam et al, 2014).…”
Section: Results On the English Reuters Datasetsupporting
confidence: 78%
“…That is why we focused directly on deep-learning methods, as they are capable of learning and predicting a full label distribution (Nam et al, 2014).…”
Section: Predicting Full Multi-label Distributionmentioning
confidence: 99%
“…There are also several relevant works that propose the inclusion of multi-label co-occurrence into loss functions such as pairwise ranking loss (Zhang and Zhou, 2006) and more recent work by Nam et al (2014), who report that binary crossentropy can outperform the pairwise ranking loss by leveraging rectified linear units (ReLUs) for nonlinearity.…”
Section: Related Workmentioning
confidence: 99%