Cargando…

On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization

Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The sta...

Descripción completa

Detalles Bibliográficos
Autores principales: Sahay, Rajat, Thomas, Georgi, Jahan, Chowdhury Sadman, Manjrekar, Mihir, Popp, Dan, Savakis, Andreas
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10611075/
https://www.ncbi.nlm.nih.gov/pubmed/37896503
http://dx.doi.org/10.3390/s23208409
_version_ 1785128406319366144
author Sahay, Rajat
Thomas, Georgi
Jahan, Chowdhury Sadman
Manjrekar, Mihir
Popp, Dan
Savakis, Andreas
author_facet Sahay, Rajat
Thomas, Georgi
Jahan, Chowdhury Sadman
Manjrekar, Mihir
Popp, Dan
Savakis, Andreas
author_sort Sahay, Rajat
collection PubMed
description Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The standard feature extraction method in domain adaptation has been convolutional neural networks (CNNs). Recently, attention-based transformer models have emerged as effective alternatives for computer vision tasks. In this paper, we benchmark three attention-based architectures, specifically vision transformer (ViT), shifted window transformer (SWIN), and dual attention vision transformer (DAViT), against convolutional architectures ResNet, HRNet and attention-based ConvNext, to assess the performance of different backbones for domain generalization and adaptation. We incorporate these backbone architectures as feature extractors in the source hypothesis transfer (SHOT) framework for UDA. SHOT leverages the knowledge learned in the source domain to align the image features of unlabeled target data in the absence of source domain data, using self-supervised deep feature clustering and self-training. We analyze the generalization and adaptation performance of these models on standard UDA datasets and aerial UDA datasets. In addition, we modernize the training procedure commonly seen in UDA tasks by adding image augmentation techniques to help models generate richer features. Our results show that ConvNext and SWIN offer the best performance, indicating that the attention mechanism is very beneficial for domain generalization and adaptation with both transformer and convolutional architectures. Our ablation study shows that our modernized training recipe, within the SHOT framework, significantly boosts performance on aerial datasets.
format Online
Article
Text
id pubmed-10611075
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-106110752023-10-28 On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization Sahay, Rajat Thomas, Georgi Jahan, Chowdhury Sadman Manjrekar, Mihir Popp, Dan Savakis, Andreas Sensors (Basel) Article Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The standard feature extraction method in domain adaptation has been convolutional neural networks (CNNs). Recently, attention-based transformer models have emerged as effective alternatives for computer vision tasks. In this paper, we benchmark three attention-based architectures, specifically vision transformer (ViT), shifted window transformer (SWIN), and dual attention vision transformer (DAViT), against convolutional architectures ResNet, HRNet and attention-based ConvNext, to assess the performance of different backbones for domain generalization and adaptation. We incorporate these backbone architectures as feature extractors in the source hypothesis transfer (SHOT) framework for UDA. SHOT leverages the knowledge learned in the source domain to align the image features of unlabeled target data in the absence of source domain data, using self-supervised deep feature clustering and self-training. We analyze the generalization and adaptation performance of these models on standard UDA datasets and aerial UDA datasets. In addition, we modernize the training procedure commonly seen in UDA tasks by adding image augmentation techniques to help models generate richer features. Our results show that ConvNext and SWIN offer the best performance, indicating that the attention mechanism is very beneficial for domain generalization and adaptation with both transformer and convolutional architectures. Our ablation study shows that our modernized training recipe, within the SHOT framework, significantly boosts performance on aerial datasets. MDPI 2023-10-12 /pmc/articles/PMC10611075/ /pubmed/37896503 http://dx.doi.org/10.3390/s23208409 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sahay, Rajat
Thomas, Georgi
Jahan, Chowdhury Sadman
Manjrekar, Mihir
Popp, Dan
Savakis, Andreas
On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title_full On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title_fullStr On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title_full_unstemmed On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title_short On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization
title_sort on the importance of attention and augmentations for hypothesis transfer in domain adaptation and generalization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10611075/
https://www.ncbi.nlm.nih.gov/pubmed/37896503
http://dx.doi.org/10.3390/s23208409
work_keys_str_mv AT sahayrajat ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization
AT thomasgeorgi ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization
AT jahanchowdhurysadman ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization
AT manjrekarmihir ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization
AT poppdan ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization
AT savakisandreas ontheimportanceofattentionandaugmentationsforhypothesistransferindomainadaptationandgeneralization