Joint Optimization With DNN Partitioning and Resource Allocation in Mobile Edge Computing
暂无分享,去创建一个
With the rapid development of computing power and artificial intelligence, IoT devices equipped with ubiquitous sensors are gradually installed with intelligence. People can enjoy many conveniences with intelligent devices, such as face recognition, video understanding, and motion estimation. Currently, deep neural networks are the mainstream technology in intelligent mobile applications. Inspired by DNN model partition schemes, the paradigm of edge computing could be utilized collaboratively to improve the effectiveness of intelligent task execution in IoT devices. However, due to the dynamics of the wireless network environment and the increasing number of IoT devices, a DNN partition policy without adequate consideration would pose a significant challenge to the efficiency of task inference. Moreover, the shortage and high rental cost of edge computing resources make the optimization of DNN-based task execution more difficult. To cope with those situations, we propose a joint method by a self-adaptive DNN partition with cost-effective resource allocation to facilitate collaborative computation between IoT devices and edge servers. Our proposed online algorithm can be proved to ensure the overall rental cost within an upper bound above the optimal solution while guaranteeing the latency for DNN-based task inference. To evaluate the performance of our strategy, we conduct extensive trace-driven illustrative studies and show that the proposed method can achieve sub-optimal results and outperforms other alternative methods.