Back to Search
Start Over
Compressing Deep Model With Pruning and Tucker Decomposition for Smart Embedded Systems
- Source :
- IEEE Internet of Things Journal. 9:14490-14500
- Publication Year :
- 2022
- Publisher :
- Institute of Electrical and Electronics Engineers (IEEE), 2022.
-
Abstract
- Deep learning has been proved to be one of the most effective method in feature encoding for different intelligent applications such as video based human action recognition. However, its non-convex optimization mechanism leads large memory consumption, which hinders its deployment on the smart embedded systems with limited computation resources. To overcome this challenge, we propose a novel deep model compression technique for smart embedded systems, which realizes both the memory size reduction and inference complexity decrease within a small drop of accuracy. First, we propose an improved naive Bayes inference based channel parameter pruning to obtain a sparse model with higher accuracy. Then, to improve the inference efficiency, the improved Tucker decomposition method is proposed, where an improved genetic algorithm is used to optimize the Tucker ranks. Finally, to elevate effectiveness of our proposed method, extensive experiments are conducted. The experimental results show that our method can achieve the state-of-the-art performance compared with existing methods in terms of accuracy, parameter compression and floating point operations reduction.
- Subjects :
- Computer Networks and Communications
Computer science
business.industry
Deep learning
Inference
Computer Science Applications
Reduction (complexity)
Naive Bayes classifier
Hardware and Architecture
Encoding (memory)
Embedded system
Signal Processing
Genetic algorithm
Pruning (decision trees)
Artificial intelligence
business
Information Systems
Tucker decomposition
Subjects
Details
- ISSN :
- 23722541
- Volume :
- 9
- Database :
- OpenAIRE
- Journal :
- IEEE Internet of Things Journal
- Accession number :
- edsair.doi...........665e2bb67fbe1aa2e7900a4d5fd4e89c