2023
DOI: 10.1016/j.patcog.2023.109547
|View full text |Cite
|
Sign up to set email alerts
|

AA-trans: Core attention aggregating transformer with information entropy selector for fine-grained visual classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(4 citation statements)
references
References 11 publications
0
4
0
Order By: Relevance
“…We first compile a list of common model structures to find the best pre-trained models for plant disease diagnosis. As shown in Table 4 , we gather a variety of structural models such as AlexNet [ 8 ], VGGNet [ 9 ] GoogleNet [ 10 ], ResNet [ 21 ], DenseNet [ 42 ], EffecientNet [ 43 ], and others that can be used directly for plant disease classification or as backbone networks for other more complex models [ 44 , 45 ] to implement different tasks such as plant disease detection and segmentation. Furthermore, to facilitate the use of plant disease pre-trained models in plant disease detection and segmentation tasks, we collect the backbone network structures of the YOLO (You Only Look Once) series, MaskRCNN, FCN (Fully Convolutional Network), and DeepLab as part of the list of the pre-trained models.…”
Section: Methodsmentioning
confidence: 99%
“…We first compile a list of common model structures to find the best pre-trained models for plant disease diagnosis. As shown in Table 4 , we gather a variety of structural models such as AlexNet [ 8 ], VGGNet [ 9 ] GoogleNet [ 10 ], ResNet [ 21 ], DenseNet [ 42 ], EffecientNet [ 43 ], and others that can be used directly for plant disease classification or as backbone networks for other more complex models [ 44 , 45 ] to implement different tasks such as plant disease detection and segmentation. Furthermore, to facilitate the use of plant disease pre-trained models in plant disease detection and segmentation tasks, we collect the backbone network structures of the YOLO (You Only Look Once) series, MaskRCNN, FCN (Fully Convolutional Network), and DeepLab as part of the list of the pre-trained models.…”
Section: Methodsmentioning
confidence: 99%
“…To differentiate objectives from inferior categorization is the goal of smooth visual categorization [22]. It is thought to be a very challenging assignment since smooth images naturally exhibit significant inter-class variations and tiny intra-class variability.…”
Section: Related Workmentioning
confidence: 99%
“…( 22) displays the percentage of positive situations that were expected to be negative but ended up being positive. (22) The contrast provided Table II highlights the FPR and FNR performance of the Hybrid DL-Attention Mechanism. The results highlight the potential of deep learning's attention mechanisms by showing how they may considerably improve the model's performance in tasks that call for striking a careful balance between reducing false alarms and missed detections.…”
Section: Training and Testingmentioning
confidence: 99%
“…Due to these practical issues, researchers have spent a lot of effort on recognition models based on a single image, but currently they can only achieve an accuracy of about 75% at best (Wang et al, 2023), with little improvement. It appears that the accuracy has reached a ceiling.…”
Section: Introductionmentioning
confidence: 99%