Thirteenth International Conference on Digital Image Processing (ICDIP 2021) 2021
DOI: 10.1117/12.2600465
|View full text |Cite
|
Sign up to set email alerts
|

Bidirectional LSTM approach to image captioning with scene features

Abstract: I hereby certify that the information contained in this (my submission) is information pertaining to research I conducted for this project. All information other than my own contribution will be fully referenced and listed in the relevant bibliography section at the rear of the project. ALL internet material must be referenced in the bibliography section. Students are required to use the Referencing Standard specified in the report template. To use other author's written or electronic work is illegal (plagiari… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 23 publications
0
3
0
Order By: Relevance
“…It produces around 73.80% accuracy for cosine similarity algorithm. In another experiment we have done comparative analysis with CNN-LSTM (4) and CNN-Bi-LSTM (5) as existing system for caption generation. The below According to this Figure 9 we conclude our system predicts superior results in terms of all performance parameters.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…It produces around 73.80% accuracy for cosine similarity algorithm. In another experiment we have done comparative analysis with CNN-LSTM (4) and CNN-Bi-LSTM (5) as existing system for caption generation. The below According to this Figure 9 we conclude our system predicts superior results in terms of all performance parameters.…”
Section: Resultsmentioning
confidence: 99%
“…The below According to this Figure 9 we conclude our system predicts superior results in terms of all performance parameters. The proposed algorithm is compared with the deep CNN (4) and CNN-Ni-LSTM (5) algorithm. We also evaluate some machine learning algorithm such as such as Navie Bayes, Random Forest & Support Vector Machine Here we conclude the performance of the system is better than compared to the existing system.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Pretrained CNN models are used for the encoder while long short-term memory (LSTM) or gated recurrent unit (GRU) neural networks are commonly used for the language generation. However, the encoder-decoder method is limited in its ability to preserve all source information in the fxedlength vector, and the unidirectional LSTM decoder only preserves past information which leads to poor outcomes for long sequential data [9].…”
Section: Introductionmentioning
confidence: 99%