1. Optimization and Deployment of Memory-Intensive Operations in Deep Learning Model on Edge
- Author
-
Peng XU, Jianxin ZHAO, Chi Harold LIU
- Subjects
memory optimization ,deep compiler ,computation optimization ,model deployment ,edge computing ,Computer software ,QA76.75-76.765 ,Technology (General) ,T1-995 - Abstract
As a large amount of data is increasingly generated from edge devices,such as smart homes,mobile phones,and wearable devices,it becomes crucial for many applications to deploy machine learning modes across edge devices.The execution speed of the deployed model is a key element to ensure service quality.Considering a highly heterogeneous edge deployment scenario,deep learning compiling is a novel approach that aims to solve this problem.It defines models using certain DSLs and generates efficient code implementations on different hardware devices.However,there are still two aspects that are not yet thoroughly investigated yet.The first is the optimization of memory-intensive operations,and the second problem is the heterogeneity of the deployment target.To that end,in this work,we propose a system solution that optimizes memory-intensive operation,optimizes the subgraph distribution,and enables the compiling and deployment of DNN models on multiple targets.The evaluation results show the performance of our proposed system.
- Published
- 2023
- Full Text
- View/download PDF