Proceedings of the Symposium on Applied Computing 2017
DOI: 10.1145/3019612.3019664
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical multi-label classification with chained neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
207
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 126 publications
(207 citation statements)
references
References 16 publications
0
207
0
Order By: Relevance
“…Secondly, we found that the original performance of HMCN (Wehrmann et al, 2018) is sometimes much lower than expected. After tuning their model, we observed that if we first conduct a weighted sum of the local and global outputs and then apply the sigmoid function, the performance of HMCN becomes much better (see Table 7) than doing them in the opposite order as in Wehrmann et al (2018). In addition, we found that HMCN + HAN (Yang et al, 2016) would result in extremely low performance.…”
Section: B Performance Analysis Of Baselinesmentioning
confidence: 69%
See 3 more Smart Citations
“…Secondly, we found that the original performance of HMCN (Wehrmann et al, 2018) is sometimes much lower than expected. After tuning their model, we observed that if we first conduct a weighted sum of the local and global outputs and then apply the sigmoid function, the performance of HMCN becomes much better (see Table 7) than doing them in the opposite order as in Wehrmann et al (2018). In addition, we found that HMCN + HAN (Yang et al, 2016) would result in extremely low performance.…”
Section: B Performance Analysis Of Baselinesmentioning
confidence: 69%
“…There are not many neural methods that specifically target HTC. We mainly compare with two latest neural models: HR-DGCNN (Peng et al, 2018), which extends hierarchical regularization (Gopal and Yang, 2013) to Graph-CNN and compares favorably to flat models like RCNN (Lai et al, 2015) and XML-CNN (Liu et al, 2017), and HMCN (Wehrmann et al, 2018), which outperforms state-of-the-art HTC methods such as HMC-LMLP (Cerri et al, 2016). We also compare with the base models that we use for feature encoding.…”
Section: Compared Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Concretely, we select a best model through coarse-grained experiments on each of the two benchmarks and fix it, and then fine-tune the features and hyperparameters, such as model structures, input representations, activation functions, optimizers, learning rate, etc. The best performance models 6 are as Models RCV1 Yelp Micro-F1 Micro-F1 HR-DGCNN (Peng et al, 2018) 0.7610 -HMCN (Wehrmann et al, 2018) 0.8080 0.6640 Our best models 0.8099 0.6704 follows: (1) RCNN with two-layers Bi-GRU and one-layer CNN for RCV1 dataset (input = word, optimizer = Adam, learning rate = 0.008);…”
Section: Results Of Using Rich Models and Featuresmentioning
confidence: 99%