Back to Search Start Over

Deep Transfer Learning Method Using Self-Pixel and Global Channel Attentive Regularization

Authors :
Changhee Kang
Sang-ug Kang
Source :
Sensors, Vol 24, Iss 11, p 3522 (2024)
Publication Year :
2024
Publisher :
MDPI AG, 2024.

Abstract

The purpose of this paper is to propose a novel transfer learning regularization method based on knowledge distillation. Recently, transfer learning methods have been used in various fields. However, problems such as knowledge loss still occur during the process of transfer learning to a new target dataset. To solve these problems, there are various regularization methods based on knowledge distillation techniques. In this paper, we propose a transfer learning regularization method based on feature map alignment used in the field of knowledge distillation. The proposed method is composed of two attention-based submodules: self-pixel attention (SPA) and global channel attention (GCA). The self-pixel attention submodule utilizes both the feature maps of the source and target models, so that it provides an opportunity to jointly consider the features of the target and the knowledge of the source. The global channel attention submodule determines the importance of channels through all layers, unlike the existing methods that calculate these only within a single layer. Accordingly, transfer learning regularization is performed by considering both the interior of each single layer and the depth of the entire layer. Consequently, the proposed method using both of these submodules showed overall improved classification accuracy than the existing methods in classification experiments on commonly used datasets.

Details

Language :
English
ISSN :
24113522 and 14248220
Volume :
24
Issue :
11
Database :
Directory of Open Access Journals
Journal :
Sensors
Publication Type :
Academic Journal
Accession number :
edsdoj.1540adb981d9456586038b54dc4c41d0
Document Type :
article
Full Text :
https://doi.org/10.3390/s24113522