Cargando…
Multi-View Visual Question Answering with Active Viewpoint Selection
This paper proposes a framework that allows the observation of a scene iteratively to answer a given question about the scene. Conventional visual question answering (VQA) methods are designed to answer given questions based on single-view images. However, in real-world applications, such as human–r...
Autores principales: | Qiu, Yue, Satoh, Yutaka, Suzuki, Ryota, Iwata, Kenji, Kataoka, Hirokatsu |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7219048/ https://www.ncbi.nlm.nih.gov/pubmed/32316433 http://dx.doi.org/10.3390/s20082281 |
Ejemplares similares
-
Indoor Scene Change Captioning Based on Multimodality Data
por: Qiu, Yue, et al.
Publicado: (2020) -
Multi-modal adaptive gated mechanism for visual question answering
por: Xu, Yangshuyi, et al.
Publicado: (2023) -
Multi-Modal Explicit Sparse Attention Networks for Visual Question Answering
por: Guo, Zihan, et al.
Publicado: (2020) -
Parallel multi-head attention and term-weighted question embedding for medical visual question answering
por: Manmadhan, Sruthy, et al.
Publicado: (2023) -
The multi-modal fusion in visual question answering: a review of attention mechanisms
por: Lu, Siyu, et al.
Publicado: (2023)