2016
DOI: 10.48550/arxiv.1604.04573
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CNN-RNN: A Unified Framework for Multi-label Image Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
37
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(38 citation statements)
references
References 16 publications
1
37
0
Order By: Relevance
“…To achieve this, a series of works introduced graphic models, such as Conditional Random Field [8], Dependency Network [10], or co-occurrence matrix [29] to capture pairwise label correlations. Recently, Wang et al [24] formulated a CNN-RNN framework that utilized the semantic redundancy and the co-occurrence dependency implicitly to facilitate effective multi-label classification. Some works [33,2] further took advantage of proposal generation/visual attention mechanism to search local discriminative regions and LSTM [13] to explicitly model label dependencies.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…To achieve this, a series of works introduced graphic models, such as Conditional Random Field [8], Dependency Network [10], or co-occurrence matrix [29] to capture pairwise label correlations. Recently, Wang et al [24] formulated a CNN-RNN framework that utilized the semantic redundancy and the co-occurrence dependency implicitly to facilitate effective multi-label classification. Some works [33,2] further took advantage of proposal generation/visual attention mechanism to search local discriminative regions and LSTM [13] to explicitly model label dependencies.…”
Section: Related Workmentioning
confidence: 99%
“…Since the ground truth annotations of test set are unavailable, our method and all existing competitors are trained on the training set and evaluated on the validation set. For the OP, OR, OF1 and CP, CR, CF1 metrics with top-3 constraint, we follow existing methods [24] to exclude the labels with probabilities lower than a threshold (0.5 in our experiments).…”
Section: Comparison On Microsoft Cocomentioning
confidence: 99%
See 1 more Smart Citation
“…However, the model must allow appropriate label interactions for beneficial results. Multi-task learning is frequently applied to tag images with multiple labels (Wang et al, 2016;Wei et al, 2015). Multi-task sequence learning (Sutton et al, 2007;Collobert et al, 2011) is the task to jointly tag sequence values with multiple label categories.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Kiros et al used deep representation of images for automatic image annotation [52], [72]. While convolutional neural networks like [53] and [91] were designed to identify a single object label for images, deep and recurrent neural networks have been employed for multi-label ranking and classification problem [33], [111]. Training a deep neural network requires availability of large training dataset, and is computationally extremely expensive.…”
Section: B Deep Convolutional Neural Networkmentioning
confidence: 99%