2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2020
DOI: 10.1109/cvprw50498.2020.00361
|View full text |Cite
|
Sign up to set email alerts
|

AdaMT-Net: An Adaptive Weight Learning Based Multi-Task Learning Model For Scene Understanding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 13 publications
0
8
0
Order By: Relevance
“…As shown in the table, our model outperforms all previous works except for AdaMT-Net [5]. Compared to AdaMT-Net, our model improves mIoU and relative depth error by a fair margin.…”
Section: E Resultsmentioning
confidence: 66%
See 2 more Smart Citations
“…As shown in the table, our model outperforms all previous works except for AdaMT-Net [5]. Compared to AdaMT-Net, our model improves mIoU and relative depth error by a fair margin.…”
Section: E Resultsmentioning
confidence: 66%
“…We think that this is due to the difference of using attention modules or not. Previous works such as [4], [5], [8], [29] have used attention modules in their networks, enabling the model to "look" at the entire image in the training phase. On the other hand, since our model only uses convolutional layers, the model can only learn from pixels nearby.…”
Section: E Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…It is also possible to make a distinction according to the level of parameter sharing in a multi-task framework. Particularly, Jha et al [120] distinguishes between soft and hard-sharing. In the first case, models have a separate network for each task under consideration, resulting in a disjoint set of parameters.…”
Section: Depth As Predictionmentioning
confidence: 99%
“…Furthermore, learning multiple tasks at once can improve generalization ability and lead to better results compared to single-task performance. A number of works exists, which tackle the different tasks for scene understanding in a multi-task setting [18,28,29,48,55,63,68,74,79,80]. Goel et al [18] propose QuadroNet, a real-time capable model to predict 2D bounding boxes, panoptic segmentation, and depth from single images.…”
Section: Multi-task Learningmentioning
confidence: 99%