2021 10th IEEE International Conference on Communication Systems and Network Technologies (CSNT) 2021
DOI: 10.1109/csnt51715.2021.9509657
|View full text |Cite
|
Sign up to set email alerts
|

Neural Dense Captioning with Visual Attention

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 7 publications
0
1
0
Order By: Relevance
“…Meanwhile, research [52] employs the Inception-v3 encoder and gated recurrent unit (GRU) decoder to generate captions in Indonesian. On the other hand, several studies summarized in [53], [54] add an attention layer to the decoder network. This attention mechanism is effective in increasing accuracy because it allows the model to focus on essential parts when processing input into output sequences [55].…”
Section: Introductionmentioning
confidence: 99%
“…Meanwhile, research [52] employs the Inception-v3 encoder and gated recurrent unit (GRU) decoder to generate captions in Indonesian. On the other hand, several studies summarized in [53], [54] add an attention layer to the decoder network. This attention mechanism is effective in increasing accuracy because it allows the model to focus on essential parts when processing input into output sequences [55].…”
Section: Introductionmentioning
confidence: 99%