Back to Search Start Over

Training Noise Token Pruning

Authors :
Rao, Mingxing
Jiang, Bohan
Moyer, Daniel
Publication Year :
2024

Abstract

In the present work we present Training Noise Token (TNT) Pruning for vision transformers. Our method relaxes the discrete token dropping condition to continuous additive noise, providing smooth optimization in training, while retaining discrete dropping computational gains in deployment settings. We provide theoretical connections to Rate-Distortion literature, and empirical evaluations on the ImageNet dataset using ViT and DeiT architectures demonstrating TNT's advantages over previous pruning methods.<br />Comment: 25 pages, 8 figures

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2411.18092
Document Type :
Working Paper