Back to Search
Start Over
Histogram-Equalized Quantization for logic-gated Residual Neural Networks
- Source :
- 2022 IEEE International Symposium on Circuits and Systems (ISCAS), Austin, TX, USA, 2022, pp. 1289-1293
- Publication Year :
- 2025
-
Abstract
- Adjusting the quantization according to the data or to the model loss seems mandatory to enable a high accuracy in the context of quantized neural networks. This work presents Histogram-Equalized Quantization (HEQ), an adaptive framework for linear symmetric quantization. HEQ automatically adapts the quantization thresholds using a unique step size optimization. We empirically show that HEQ achieves state-of-the-art performances on CIFAR-10. Experiments on the STL-10 dataset even show that HEQ enables a proper training of our proposed logic-gated (OR, MUX) residual networks with a higher accuracy at a lower hardware complexity than previous work.<br />Comment: Published at IEEE ISCAS 2022
- Subjects :
- Computer Science - Machine Learning
Computer Science - Hardware Architecture
Subjects
Details
- Database :
- arXiv
- Journal :
- 2022 IEEE International Symposium on Circuits and Systems (ISCAS), Austin, TX, USA, 2022, pp. 1289-1293
- Publication Type :
- Report
- Accession number :
- edsarx.2501.04517
- Document Type :
- Working Paper
- Full Text :
- https://doi.org/10.1109/ISCAS48785.2022.9937290