Back to Search
Start Over
An End-to-End Workflow to Efficiently Compress and Deploy DNN Classifiers on SoC/FPGA.
- Source :
- IEEE Embedded Systems Letters; Sep2024, Vol. 16 Issue 3, p255-258, 4p
- Publication Year :
- 2024
-
Abstract
- Machine learning (ML) models have demonstrated discriminative and representative learning capabilities over a wide range of applications, even at the cost of high-computational complexity. Due to their parallel processing capabilities, reconfigurability, and low-power consumption, systems on chip based on a field programmable gate array (SoC/FPGA) have been used to face this challenge. Nevertheless, SoC/FPGA devices are resource-constrained, which implies the need for optimal use of technology for the computation and storage operations involved in ML-based inference. Consequently, mapping a deep neural network (DNN) architecture to a SoC/FPGA requires compression strategies to obtain a hardware design with a good compromise between effectiveness, memory footprint, and inference time. This letter presents an efficient end-to-end workflow for deploying DNNs on an SoC/FPGA by integrating hyperparameter tuning through Bayesian optimization (BO) with an ensemble of compression techniques. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 19430663
- Volume :
- 16
- Issue :
- 3
- Database :
- Complementary Index
- Journal :
- IEEE Embedded Systems Letters
- Publication Type :
- Academic Journal
- Accession number :
- 179296480
- Full Text :
- https://doi.org/10.1109/LES.2023.3343030