4 results on '"weight quantization"'
Search Results
2. Weight-Quantized SqueezeNet for Resource-Constrained Robot Vacuums for Indoor Obstacle Classification
- Author
-
Qian Huang
- Subjects
robot vacuums ,memory efficient ,deep learning ,weight quantization ,SqueezeNet ,General Earth and Planetary Sciences ,GeneralLiterature_MISCELLANEOUS ,General Environmental Science - Abstract
With the rapid development of artificial intelligence (AI) theory, particularly deep learning neural networks, robot vacuums equipped with AI power can automatically clean indoor floors by using intelligent programming and vacuuming services. To date, several deep AI models have been proposed to distinguish indoor objects between cleanable litter and noncleanable hazardous obstacles. Unfortunately, these existing deep AI models focus entirely on the accuracy enhancement of object classification, and little effort has been made to minimize the memory size and implementation cost of AI models. As a result, these existing deep AI models require far more memory space than a typical robot vacuum can provide. To address this shortcoming, this paper aims to study and find an efficient deep AI model that can achieve a good balance between classification accuracy and memory usage (i.e., implementation cost). In this work, we propose a weight-quantized SqueezeNet model for robot vacuums. This model can classify indoor cleanable litters from noncleanable hazardous obstacles based on the image or video captures from robot vacuums. Furthermore, we collect videos or pictures captured by built-in cameras of robot vacuums and use them to construct a diverse dataset. The dataset contains 20,000 images with a ground-view perspective of dining rooms, kitchens and living rooms for various houses under different lighting conditions. Experimental results show that the proposed deep AI model can achieve comparable object classification accuracy of around 93% while reducing memory usage by at least 22.5 times. More importantly, the memory footprint required by our AI model is only 0.8 MB, indicating that this model can run smoothly on resource-constrained robot vacuums, where low-end processors or microcontrollers are dedicated to running AI algorithms.
- Published
- 2022
- Full Text
- View/download PDF
3. Optimized programming algorithms for multilevel RRAM in hardware neural networks
- Author
-
Christian Wenger, Cristian Zambelli, Daniele Ielmini, Oscar G. Ossorio, Piero Olivo, Mamathamba Kalishettyhalli Mahadevaiah, Francesco Anzalone, Eduardo Perez, and Valerio Milo
- Subjects
Scheme (programming language) ,Computer science ,Reliability (computer networking) ,hardware neural networks ,in-memory computing ,multilevel programming ,resistance variability ,Resistive-switching random access memory (RRAM) ,weight quantization ,02 engineering and technology ,01 natural sciences ,NO ,Software ,In-Memory Processing ,0103 physical sciences ,PE7_5 ,computer.programming_language ,010302 applied physics ,Artificial neural network ,business.industry ,021001 nanoscience & nanotechnology ,Resistive random-access memory ,Key (cryptography) ,0210 nano-technology ,business ,Algorithm ,computer ,MNIST database - Abstract
A key requirement for RRAM in neural network accelerators with a large number of synaptic parameters is the multilevel programming. This is hindered by resistance imprecision due to cycle-to-cycle and device-to-device variations. Here, we compare two multilevel programming algorithms to minimize resistance variations in a 4-kbit array of HfO 2 RRAM. We show that gate-based algorithms have the highest reliability. The optimized scheme is used to implement a neural network with 9-level weights, achieving 91.5% (vs. software 93.27%) in MNIST recognition.
- Published
- 2021
- Full Text
- View/download PDF
4. Learning Sparse Convolutional Neural Network via Quantization With Low Rank Regularization
- Author
-
Liu Yan, Zongcheng Ben, Xin Long, Maojun Zhang, Xiangrong Zeng, and Zhou Dianle
- Subjects
General Computer Science ,Computer science ,Computation ,Convolutional neural network (CNN) ,Inference ,02 engineering and technology ,Convolutional neural network ,Regularization (mathematics) ,Quantization (physics) ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,Tensor ,visualization ,weight quantization ,Artificial neural network ,Quantization (signal processing) ,sparsity ,020208 electrical & electronic engineering ,General Engineering ,channel pruning ,spectral regularization ,020201 artificial intelligence & image processing ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,lcsh:TK1-9971 ,Algorithm ,Communication channel - Abstract
With the refinement of tasks in artificial intelligence, bringing in exponential level increments in computation cost and storage. Therefore, the augment of computation resource for complicated neural networks severely hinders their applications on limited-power devices in recent years. As a result, there is an impending necessity to compress and accelerate the deep networks by special ways. Considering the different peculiarities of weight quantization and sparse regularization, in this paper, we propose a low rank sparse quantization (LRSQ) method to quantize network weights and regularize the corresponding structures at the same time. Our LRSQ can: 1) obtain low-bit quantized networks to reduce memory and computation cost and 2) learn a compact structure from complex convolutional networks for subsequent channel pruning which has significant reduction on FLOPs. In experimental sections, we evaluate the proposed method on several popular models such as VGG-7/16/19 and ResNet-18/34/50, and results show that this method can dramatically reduce parameters and channels of the network with slight inference accuracy loss. Furthermore, we also visualize and analyze the four-dimensional weight tensors, which shows the low rank and group-sparsity structure of it. Finally, we try pruning unimportant channels which are zero-channels in our quantized model, and finding even a little better precision than the standard full-precision network.
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.