Cargando…
PLG-ViT: Vision Transformer with Parallel Local and Global Self-Attention
Recently, transformer architectures have shown superior performance compared to their CNN counterparts in many computer vision tasks. The self-attention mechanism enables transformer networks to connect visual dependencies over short as well as long distances, thus generating a large, sometimes even...
Autores principales: | Ebert, Nikolas, Stricker, Didier, Wasenmüller, Oliver |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098752/ https://www.ncbi.nlm.nih.gov/pubmed/37050507 http://dx.doi.org/10.3390/s23073447 |
Ejemplares similares
-
Gait-ViT: Gait Recognition with Vision Transformer
por: Mogan, Jashila Nair, et al.
Publicado: (2022) -
PF-ViT: Parallel and Fast Vision Transformer for Offline Handwritten Chinese Character Recognition
por: Dan, Yongping, et al.
Publicado: (2022) -
HGR-ViT: Hand Gesture Recognition with Vision Transformer
por: Tan, Chun Keat, et al.
Publicado: (2023) -
The classification of the bladder cancer based on Vision Transformers (ViT)
por: Khedr, Ola S., et al.
Publicado: (2023) -
RT-ViT: Real-Time Monocular Depth Estimation Using Lightweight Vision Transformers
por: Ibrahem, Hatem, et al.
Publicado: (2022)