Cargando…
Image to English translation and comprehension: INT2-VQA method based on inter-modality and intra-modality collaborations
Existing visual question answering methods typically concentrate only on visual targets in images, ignoring the key textual content in the images, thereby limiting the depth and accuracy of image content comprehension. Inspired by this, we pay attention to the task of text-based visual question answ...
Autor principal: | Sheng, Xianli |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10468077/ https://www.ncbi.nlm.nih.gov/pubmed/37647277 http://dx.doi.org/10.1371/journal.pone.0290315 |
Ejemplares similares
-
Joint embedding VQA model based on dynamic word vector
por: Ma, Zhiyang, et al.
Publicado: (2021) -
Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
por: Li, Jiangfeng, et al.
Publicado: (2022) -
MRI Cross-Modality Image-to-Image Translation
por: Yang, Qianye, et al.
Publicado: (2020) -
On-chip inter-modal Brillouin scattering
por: Kittlaus, Eric A., et al.
Publicado: (2017) -
Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
por: Fu, Changzeng, et al.
Publicado: (2020)