2020
DOI: 10.1007/s11571-020-09615-4
|View full text |Cite
|
Sign up to set email alerts
|

End-to-end face parsing via interlinked convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 28 publications
0
13
0
Order By: Relevance
“…The localized regions are then sent to TaNet's segmentation pathways for pixel-wise prediction as shown in Figure 1. After segmenting the ROIs using TaNet's pathways, these regions can be remapped [23] to their original positions using a reverse grid transformer (G −1 ).…”
Section: ) Bilinear Sampler (S)mentioning
confidence: 99%
See 1 more Smart Citation
“…The localized regions are then sent to TaNet's segmentation pathways for pixel-wise prediction as shown in Figure 1. After segmenting the ROIs using TaNet's pathways, these regions can be remapped [23] to their original positions using a reverse grid transformer (G −1 ).…”
Section: ) Bilinear Sampler (S)mentioning
confidence: 99%
“…The ground truth transformation matrix is calculated for each region (θ gt r , r ∈ {1, 2, ..N }) as described in [23]. Specifically, we calculated the central coordinates (x, y) for each coarsely segmented ROI r (r ∈ {1, ..N }) and estimated θ gt as:…”
Section: Tanet Trainingmentioning
confidence: 99%
“…Localization Network Prior to the use of localization network, we use FCN-8 [16,26] model for coarse segmentation. Providing a coarse segmentation of different ROIs allows the localization network to 1) generate the transformation parameters (θ) for these regions and 2) learn the context relationship among them.…”
Section: Rois Localizationmentioning
confidence: 99%
“…where θ ∈ R N ×2×3 ; here N = 5 as there are five cardiac regions as shown in Figure 1. As for the localization network (L), we used a simplified version of VGG16 [26] that has 8 convolutional layers and a final regression layer to generate N × 2 × 3 spatial transformation matrix (θ). L outputs the spatial transformation matrix (θ) as shown in Figure 1.…”
Section: Rois Localizationmentioning
confidence: 99%
“…We removed all skip-connections from S-FRCNN and obtained a model called S-FRCNN (no-skip). A CNN model with similar architecture has been used in face parsing [42,39], but different blocks do not share † https://cslikai.cn/project/AFRCNN ‡ https://github.com/etzinis/sudo_rm_rf/blob/master/sudo_rm_rf/dnn/models/ improved_sudormrf.py 1, it is seen that S-FRCNN (no-skip) achieved worse results than the original S-FRCNN.…”
Section: Comparison Of Micro-level Updating Schemesmentioning
confidence: 99%