2021
DOI: 10.1007/978-3-030-90525-5_17
|View full text |Cite
|
Sign up to set email alerts
|

Gaze Assisted Visual Grounding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 23 publications
0
1
0
Order By: Relevance
“…Consequently, the combination of novel categories of objects and complex referring expressions results in decreased performance on RefMD. Improvement in disambiguation and REC performance can be achieved by comparing and exploring different and more sophisticated disambiguation approaches, such as attribute-guided disambiguation [48], to improve the accuracy of grounding as well as by incorporating gesture [49] and gaze [50] information. While the adapted model comprehends the natural language object descriptions with 82% accuracy in the user study, the domain gap between the synthetic and real-world application can further be reduced by incorporating more variation and randomization [29] in the RefMD dataset.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…Consequently, the combination of novel categories of objects and complex referring expressions results in decreased performance on RefMD. Improvement in disambiguation and REC performance can be achieved by comparing and exploring different and more sophisticated disambiguation approaches, such as attribute-guided disambiguation [48], to improve the accuracy of grounding as well as by incorporating gesture [49] and gaze [50] information. While the adapted model comprehends the natural language object descriptions with 82% accuracy in the user study, the domain gap between the synthetic and real-world application can further be reduced by incorporating more variation and randomization [29] in the RefMD dataset.…”
Section: Discussion and Future Workmentioning
confidence: 99%