2023
DOI: 10.1145/3519030
|View full text |Cite
|
Sign up to set email alerts
|

Double Attention Based on Graph Attention Network for Image Multi-Label Classification

Abstract: The task of image multi-label classification is to accurately recognize multiple objects in an input image. Most of the recent works need to leverage the label co-occurrence matrix counted from training data to construct the graph structure, which are inflexible and may degrade model generalizability. In addition, these methods fail to capture the semantic correlation between the channel feature maps to further improve the model performance. To address these issues, we propose a D ouble… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(7 citation statements)
references
References 54 publications
0
7
0
Order By: Relevance
“…Compared to the remaining methods, our method still maintains the optimal performance on 10 categories. Notably, our method performs well in the categories of boats, bottles, sofas, and TVs, achieving a significant improvement of nearly 1%$\%$ over COP [14], DER [15], CPCL [16], DA‐GAT [26], CANet [27], and IA‐GCN [28]. For the few categories where the prediction accuracy is not optimal, the difference between our method and the optimal value remains within a very small range.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Compared to the remaining methods, our method still maintains the optimal performance on 10 categories. Notably, our method performs well in the categories of boats, bottles, sofas, and TVs, achieving a significant improvement of nearly 1%$\%$ over COP [14], DER [15], CPCL [16], DA‐GAT [26], CANet [27], and IA‐GCN [28]. For the few categories where the prediction accuracy is not optimal, the difference between our method and the optimal value remains within a very small range.…”
Section: Resultsmentioning
confidence: 99%
“…With extensive research [26][27][28][29], all of these methods try to strengthen the learning of regional features related to category. However, they neglect the possible impact of category scale diversity and always perform category label identification and supervision at the feature layer of the same scale.…”
Section: Spatial Location Information Utilizationmentioning
confidence: 99%
“…The GAT model has higher computational efficiency and has been widely applied in various fields, such as social network analysis, image processing, and natural language processing. 13 Recently, the GAT is also used in code defect detection, software security analysis, and code similarity detection. 27 Therefore, we apply the GAT model to explore effective features and predict the location of faults.…”
Section: Gatmentioning
confidence: 99%
“…GAT is based on the transformer model and introduces a masked self-attention mechanism that assigns different weights to the representation of each node according to the different features of its neighboring nodes. 12,13 To address the traditional SBFL and DLFL techniques' limitations, we propose a fault localization approach based on the WEGAT. It abstracts the coverage information of test cases and program elements into an execution graph and analyzes the information of the graph using a GAT.…”
Section: Introductionmentioning
confidence: 99%
“…Channel attention has shown significant advantages in many image processing tasks. For example, in image classification tasks, channel attention can help the network to better distinguish feature differences between different categories [ 13 , 14 , 15 , 16 , 17 ]. In a target detection task, channel attention improves the network’s ability to accurately locate and recognize targets [ 18 , 19 , 20 , 21 , 22 ].…”
Section: Introductionmentioning
confidence: 99%