2023
DOI: 10.1016/j.neucom.2023.126284
|View full text |Cite
|
Sign up to set email alerts
|

3D human pose and shape estimation via de-occlusion multi-task learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
10

Relationship

0
10

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 15 publications
0
2
0
Order By: Relevance
“…(d) Multi-modal question answering: Future question-answering systems need to support multiple forms of input, such as text, images, and voice [69], to meet the diverse needs of users. For example, Ning et al [70] proposed a novel method called Differentiable Image-Language Fusion (DILF) for multi-view image and language fusion.…”
Section: Trends and Conclusionmentioning
confidence: 99%
“…(d) Multi-modal question answering: Future question-answering systems need to support multiple forms of input, such as text, images, and voice [69], to meet the diverse needs of users. For example, Ning et al [70] proposed a novel method called Differentiable Image-Language Fusion (DILF) for multi-view image and language fusion.…”
Section: Trends and Conclusionmentioning
confidence: 99%
“…The average accuracy of predicting key points is the highest when introducing a regression module based on the anchor pose and a module fused with 3D pose data input, with only a few keypoints showing little decrease. Overall, the FPNet has the highest prediction accuracy, but there is still room for improvement [45]. 7, exhibit significant attitude changes and scale variances.…”
Section: Performance Between Multiple Datasetsmentioning
confidence: 99%
“…However, despite the remarkable achievements made in the field of robot music performance, there still exist several limitations and unresolved issues in existing research (Ran et al, 2023). Firstly, current robot music performances face challenges in the realm of multi-modal fusion.…”
Section: Related Workmentioning
confidence: 99%