Cargando…
UAT: Universal Attention Transformer for Video Captioning
Video captioning via encoder–decoder structures is a successful sentence generation method. In addition, using various feature extraction networks for extracting multiple features to obtain multiple kinds of visual features in the encoding process is a standard method for improving model performance...
Autores principales: | Im, Heeju, Choi, Yong-Suk |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269373/ https://www.ncbi.nlm.nih.gov/pubmed/35808316 http://dx.doi.org/10.3390/s22134817 |
Ejemplares similares
-
Hydrocarbon Sorption in Flexible MOFs—Part I: Thermodynamic Analysis with the Dubinin-Based Universal Adsorption Theory (D-UAT)
por: Preißler-Kurzhöfer, Hannes, et al.
Publicado: (2022) -
UAT defined: a guide to practical user acceptance testing as a silver bullet
por: Cimperman, Rob
Publicado: (2006) -
Video captioning with stacked attention and semantic hard pull
por: Rahman, Md. Mushfiqur, et al.
Publicado: (2021) -
Video captioning based on vision transformer and reinforcement learning
por: Zhao, Hong, et al.
Publicado: (2022) -
Lightweight dense video captioning with cross-modal attention and knowledge-enhanced unbiased scene graph
por: Han, Shixing, et al.
Publicado: (2023)