2021
DOI: 10.3389/fncel.2021.621252
|View full text |Cite
|
Sign up to set email alerts
|

OptiFlex: Multi-Frame Animal Pose Estimation Combining Deep Learning With Optical Flow

Abstract: Animal pose estimation tools based on deep learning have greatly improved animal behaviour quantification. These tools perform pose estimation on individual video frames, but do not account for variability of animal body shape in their prediction and evaluation. Here, we introduce a novel multi-frame animal pose estimation framework, referred to as OptiFlex. This framework integrates a flexible base model (i.e., FlexibleBaseline), which accounts for variability in animal body shape, with an OpticalFlow model t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(7 citation statements)
references
References 37 publications
0
7
0
Order By: Relevance
“…The 3D poses can be inferred from these 2D features by means of classical calibrated camera setups 59 ; however the 2D detection in one camera image does not benefit from the information from other cameras and the triangulation may suffer from resulting mislabeling of 2D features as well as missing detections due to occluded features. A recent approach 38,60 overcomes many of these issues by mapping from recorded images directly to 3D feature locations, again using deep learning, and is capable of classifying animal behaviors across many species 38 . An alternative approach is to use measurements that directly yield 3D information, for example RGBD 61 .…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The 3D poses can be inferred from these 2D features by means of classical calibrated camera setups 59 ; however the 2D detection in one camera image does not benefit from the information from other cameras and the triangulation may suffer from resulting mislabeling of 2D features as well as missing detections due to occluded features. A recent approach 38,60 overcomes many of these issues by mapping from recorded images directly to 3D feature locations, again using deep learning, and is capable of classifying animal behaviors across many species 38 . An alternative approach is to use measurements that directly yield 3D information, for example RGBD 61 .…”
Section: Discussionmentioning
confidence: 99%
“…An alternative approach is to use measurements that directly yield 3D information, for example RGBD 61 . In parallel, there has been substantial developments in pose estimation of humans, including the possibility to track multiple individuals in real time 60,[62][63][64] , some of which include explicit models of kinematics 65,66 . In general, these approaches triangulate joint positions of readily detectable key points in the images, which has the advantage of not requiring application of surface markers.…”
Section: Discussionmentioning
confidence: 99%
“…To quantify the performance of markerless pose-estimation, automatically determined marker positions are needed to be compared against labels determined in another way, usually from manual labelling. To compare between evaluation metrics, the literature has established a trustable criterion, called the average Percentage of Correct Keypoints (aPCK) [39, 47, 48]. The aPCK approach needs a human assessor to manually annotate a validation set (on the top of training/test set).…”
Section: Methods and Computational Approachmentioning
confidence: 99%
“…Hahn-Klimroth et al [24] presented a multistep CNN system to detect three typical African ungulate stances in zoo enclosures, including model averaging and postprocessing rules to make the system robust to outliers. Liu et al [25] used a ResNet backbone, three transposed convolution layers, and a final output layer to estimate the pose. The model is evaluated using data sets from four different animal species (mouse, fruit fly, zebrafish, and monkey).…”
Section: Related Workmentioning
confidence: 99%