2023
DOI: 10.3390/app13063614
|View full text |Cite
|
Sign up to set email alerts
|

Human Pose Estimation Based on Lightweight Multi-Scale Coordinate Attention

Abstract: Heatmap-based traditional approaches for estimating human pose usually suffer from drawbacks such as high network complexity or suboptimal accuracy. Focusing on the issue of multi-person pose estimation without heatmaps, this paper proposes an end-to-end, lightweight human pose estimation network using a multi-scale coordinate attention mechanism based on the Yolo-Pose network to improve the overall network performance while ensuring the network is lightweight. Specifically, the lightweight network GhostNet wa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 43 publications
0
2
0
Order By: Relevance
“…To derive a set of feature vectors, the GhostNet model is used. The GhostNet model removes features with some parameters and efficiently receives unwanted data from the network [ 21 ]. The GhostNet element turns the typical convolutional function into 2‐step operations.…”
Section: The Proposed Modelmentioning
confidence: 99%
“…To derive a set of feature vectors, the GhostNet model is used. The GhostNet model removes features with some parameters and efficiently receives unwanted data from the network [ 21 ]. The GhostNet element turns the typical convolutional function into 2‐step operations.…”
Section: The Proposed Modelmentioning
confidence: 99%
“…The attention mechanism has been increasingly applied in the field of image generation to enhance the extraction of specific features. For example, channel attention and spatial attention make the model focus on informative features [18]. The transformer developed recently in vision can effectively capture global attention due to self-attention [19].…”
Section: Introductionmentioning
confidence: 99%
“…However, the operating window size in maximum and average pooling affects the ability of SA in preserving important features. To further enhance the convolution feature capabilities in SA and CA, a multi scale feature fusion attention was proposed with coordinate attention (CA) mechanism [8] on a light weight bidirectional feature pyramid network.…”
Section: Introductionmentioning
confidence: 99%