计算机与现代化 ›› 2022, Vol. 0 ›› Issue (11): 75-80.

• 人工智能 • 上一篇    下一篇

基于深度Q学习的电力物联网任务卸载研究

  

  1. (1.国网电力科学研究院有限公司,江苏南京211000;2.南京邮电大学通信与信息工程学院,江苏南京210003)
  • 出版日期:2022-11-30 发布日期:2022-11-30
  • 作者简介:丁忠林(1982—),男,江苏南通人,工程师,研究方向:电力无线通信技术,E-mail: dingzhonglin@sgepri.sgcc.com.cn; 李洋(1975—),男,江苏灌云人,高级工程师,学士,研究方向:电力无线通信技术,E-mail: liyang2@sgepri.sgcc.com.cn; 曹委(1993—),男,江苏南京人,工程师,研究方向:电力无线通信技术,E-mail: caowei1@sgepri.sgcc.com.cn; 通信作者:谈宇浩(1998—),男,湖北孝感人,硕士研究生,研究方向:分布式学习,E-mail: 1136851629@qq.com; 徐波(1995—),男,江苏南京人,博士研究生,研究方向:无线通信,E-mail: 10108010321@njupt.edu.cn。
  • 基金资助:
    国家电网有限公司总部管理科技项目资助(SGZJXT00JSJS2000455)

Deep Q-learning Based Task Offloading in Power IoT

  1. (1. State Grid Electric Power Research Institute, Nari Group Corporation, Nanjing 211000, China;2. College of Telecommunications and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
  • Online:2022-11-30 Published:2022-11-30

摘要: 随着现代化城市与工业生产中电力需求的不断提高,电力物联网(Power Internet of Things, PIoT)作为一种能够显著提高电力系统效率的解决方案受到了广泛关注。为有效解决接入问题,现有的电力设备往往已配备内置轻量级人工智能的5G模组。然而,受制于模组有限的计算能力和通信能力,设备产生的海量数据难以实时处理和分析。基于该问题,本文主要研究电力物联网系统中的任务卸载问题,通过联合优化卸载决策和边缘服务器的计算资源分配,从而降低时延与能耗的加权和。此外本文提出一种基于深度强化学习的任务卸载算法,首先任务在边缘服务器的处理过程建模为队列,其次基于凸优化理论对本地计算资源分配进行优化,最后采用深度Q学习算法优化任务卸载决策。实验结果表明,本文提出的方法能够有效降低系统时延与能耗的加权和。

关键词: 电力物联网, 边缘卸载, 资源分配, 深度强化学习, 5G模组

Abstract: With the increasing demand for electricity in modern cities and industrial production, power internet of things (PIoT) has attracted extensive attention. PIoT is considered as a solution which can significantly improve the efficiency of power systems. In order to establish effective access, power equipment now is often equipped with 5G modules with lightweight built-in AI. However, limited to computing and communication capabilities of the modules, great challengs are brought by real-time processing and analysis of massive data generated by the equipment. In this paper, we mainly focus on task offloading in the PIoT system. By jointly optimizing the task scheduling and the computing resource allocation of edge servers, the weighted sum of latency and energy consumption turn out to be reduced. We propose a task offloading algorithm based on deep reinforcement learning. Firstly, the task execution on each edge server is modeled as a queuing system. Then, the local computing resource allocation is optimized based on convex optimization theory. Finally, a deep Q-learning algorithm is proposed to optimize the task offloading decisions. Simulation results show that, the proposed algorithm can reduce the latency and energy consumption significantly.

Key words: power internet of things, edge offloading, resource allocation, deep reinforcement learning, 5G modules