Back to Search
Start Over
Effects of hidden layer sizing on CNN fine-tuning
- Source :
- Future Generation Computer Systems. 118:48-55
- Publication Year :
- 2021
- Publisher :
- Elsevier BV, 2021.
-
Abstract
- Some applications have the property of being resilient, meaning that they are robust to noise (e.g. due to error) in the data. This characteristic is very useful in situations where an approximate computation allows to perform the task in less time or to deploy the algorithm on embedded hardware. Deep learning is one of the fields that can benefit from approximate computing to reduce the high number of involved parameters thanks to its impressive generalization ability. A common approach is to prune some neurons and perform an iterative re-training with the aim of both reducing the required memory and to speed-up the inference stage. In this work we propose to face CNN size reduction from a different perspective: instead of reducing the network weights or look for an approximated network very close to the Pareto frontier, we investigate whether it is possible to remove some neurons only from the fully connected layers before the network training without substantially affecting the network performance. As a case study, we will focus on “fine-tuning”, a branch of transfer learning that has shown its effectiveness especially in domains lacking effective expert-designed features. To further compact the network, we apply weight quantization to the convolutional kernels. Results show that it is possible to tailor some layers to reduce the network size, both in terms of the number of parameters to learn and required memory, without statistically affecting the performance and without the need for any additional training. Finally, we investigate to what extent the sizing operation affects the network robustness against adversarial perturbations, a set of approaches aimed at misleading deep neural networks.
- Subjects :
- Fine-tuning
Computer Networks and Communications
business.industry
Computer science
Computation
Deep learning
Pareto principle
020206 networking & telecommunications
02 engineering and technology
Sizing
Hardware and Architecture
Robustness (computer science)
0202 electrical engineering, electronic engineering, information engineering
020201 artificial intelligence & image processing
Network performance
Artificial intelligence
business
Transfer of learning
Algorithm
Software
Subjects
Details
- ISSN :
- 0167739X
- Volume :
- 118
- Database :
- OpenAIRE
- Journal :
- Future Generation Computer Systems
- Accession number :
- edsair.doi.dedup.....72b788cb991a1809eb89290bebb3b392