Back to Search Start Over

Accurate and Structured Pruning for Efficient Automatic Speech Recognition

Authors :
Jiang, Huiqiang
Zhang, Li Lyna
Li, Yuang
Wu, Yu
Cao, Shijie
Cao, Ting
Yang, Yuqing
Li, Jinyu
Yang, Mao
Qiu, Lili
Publication Year :
2023

Abstract

Automatic Speech Recognition (ASR) has seen remarkable advancements with deep neural networks, such as Transformer and Conformer. However, these models typically have large model sizes and high inference costs, posing a challenge to deploy on resource-limited devices. In this paper, we propose a novel compression strategy that leverages structured pruning and knowledge distillation to reduce the model size and inference cost of the Conformer model while preserving high recognition performance. Our approach utilizes a set of binary masks to indicate whether to retain or prune each Conformer module, and employs L0 regularization to learn the optimal mask values. To further enhance pruning performance, we use a layerwise distillation strategy to transfer knowledge from unpruned to pruned models. Our method outperforms all pruning baselines on the widely used LibriSpeech benchmark, achieving a 50% reduction in model size and a 28% reduction in inference cost with minimal performance loss.<br />Comment: Accepted at INTERSPEECH 2023

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2305.19549
Document Type :
Working Paper