Back to Search
Start Over
WegFormer: Transformers for weakly supervised semantic segmentation.
- Source :
-
Expert Systems . Mar2024, Vol. 41 Issue 3, p1-17. 17p. - Publication Year :
- 2024
-
Abstract
- Although convolutional neural networks (CNNs) have achieved remarkable progress in weakly supervised semantic segmentation (WSSS), there are still deficiencies of object incompleteness due to the lack of receptive field and insufficient utilization of global context information of CNN. Based on the above observations, we propose a simple and effective method, WegFormer. Specifically, WegFormer captures the global context information with the Vision Transformer (ViT) as the classification network and is equipped with Deep Taylor Decomposition (DTD) principle and Soft Erase (SE) module to generate more integral pseudo labels and smooth further. However, we observe that although the generated pseudo‐labels are more complete, they intrude into the background region, that is, background incompleteness problem. The Efficient Potential Object Mining (EPOM) module we propose solves this problem well. Extensive experiments on the challenging PASCAL VOC 2012 and MS COCO 2014 demonstrate the effectiveness of WegFormer, where superior results of 71.4% and 38.3% are obtained on the PASCAL VOC 2012 and MS COCO 2014 validation sets, respectively. [ABSTRACT FROM AUTHOR]
- Subjects :
- *TRANSFORMER models
*CONVOLUTIONAL neural networks
*COMPUTER vision
*DEEP learning
Subjects
Details
- Language :
- English
- ISSN :
- 02664720
- Volume :
- 41
- Issue :
- 3
- Database :
- Academic Search Index
- Journal :
- Expert Systems
- Publication Type :
- Academic Journal
- Accession number :
- 175283172
- Full Text :
- https://doi.org/10.1111/exsy.13495