Back to Search Start Over

Are transformer-based models more robust than CNN-based models?

Authors :
Liu, Zhendong
Qian, Shuwei
Xia, Changhong
Wang, Chongjun
Source :
Neural Networks. Apr2024, Vol. 172, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

As the deployment of artificial intelligence (AI) models in real-world settings grows, their open-environment robustness becomes increasingly critical. This study aims to dissect the robustness of deep learning models, particularly comparing transformer-based models against CNN-based models. We focus on unraveling the sources of robustness from two key perspectives: structural and process robustness. Our findings suggest that transformer-based models generally outperform convolution-based models in robustness across multiple metrics. However, we contend that these metrics may not wholly represent true model robustness, such as the mean of corruption error. To better understand the underpinnings of this robustness advantage, we analyze models through the lens of Fourier transform and game interaction. From our insights, we propose a calibrated evaluation metric for robustness against real-world data, and a blur-based method to enhance robustness performance. Our approach achieves state-of-the-art results, with mCE scores of 2.1% on CIFAR-10-C, 12.4% on CIFAR-100-C, and 24.9% on TinyImageNet-C. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
08936080
Volume :
172
Database :
Academic Search Index
Journal :
Neural Networks
Publication Type :
Academic Journal
Accession number :
175643396
Full Text :
https://doi.org/10.1016/j.neunet.2023.12.045