2022
DOI: 10.1016/j.knosys.2022.109763
|View full text |Cite
|
Sign up to set email alerts
|

AMAM: An Attention-based Multimodal Alignment Model for Medical Visual Question Answering

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(1 citation statement)
references
References 13 publications
0
1
0
Order By: Relevance
“…Scholars from various countries have proposed many VQA models using deep neural networks and achieved good prediction results. At present, these models can be divided into four categories, namely, models of attention-based [6], Bilinear Superdiagonal Fusion [7], Multimodal Relational Reasoning [8] and external knowledge-based [9]. Ren et al used VGG19 algorithm to extract features of images, and then input natural language questions into LSTM model as vectors.…”
Section: Related Workmentioning
confidence: 99%
“…Scholars from various countries have proposed many VQA models using deep neural networks and achieved good prediction results. At present, these models can be divided into four categories, namely, models of attention-based [6], Bilinear Superdiagonal Fusion [7], Multimodal Relational Reasoning [8] and external knowledge-based [9]. Ren et al used VGG19 algorithm to extract features of images, and then input natural language questions into LSTM model as vectors.…”
Section: Related Workmentioning
confidence: 99%