首页 | 本学科首页   官方微博 | 高级检索  
     

基于深度强化学习的服务功能链映射算法
引用本文:金明,李琳琳,张文瑾,刘文.基于深度强化学习的服务功能链映射算法[J].计算机应用研究,2020,37(11):3456-3460,3466.
作者姓名:金明  李琳琳  张文瑾  刘文
作者单位:火箭军工程大学 作战保障学院,西安710000;解放军96941 部队,北京 100089;解放军96872 部队,陕西 宝鸡721000
摘    要:针对服务功能链映射对网络时延和部署失败率的影响,提出了一种基于深度强化学习的服务功能链映射算法DQN-SFC。首先构建了一个多层次NFV管理编排架构,以满足算法对资源感知和设备配置的需求;然后基于马尔可夫决策过程建模,对SFC映射问题进行形式化描述;最后构建了一个深度强化学习网络,将网络平均时延和部署失败产生的运维开销作为奖惩反馈,经过训练后可根据网络状态决定虚拟网络功能的部署位置。通过仿真实验,对该算法的正确性和性能优势进行了验证。实验表明:与传统算法相比,该算法能有效降低网络平均时延和部署失败率,同时算法运行时间具有一定优势。

关 键 词:网络功能虚拟化  服务功能链  深度强化学习  网络时延  网络运维开销
收稿时间:2019/8/13 0:00:00
修稿时间:2020/9/26 0:00:00

SFC mapping algorithm based on deep reinforcement learning
Affiliation:Rocket Force University of Engineer,,,
Abstract:This paper proposes an algorithm for SFC mapping based on deep reinforcement learning which was called DQN-SFC, aiming at reducing the influence of SFC mapping on the average time delay and deployment failure ratio in the network. Firstly, it constructed a multi-layer NFV management and scheduling architecture to meet the requirements of resource awareness and equipment configuration of the algorithm. Secondly, based on Markov decision process, it formally described the SFC mapping problem. Finally, it constructed a deep reinforcement learning network, which used the average network delay and the operation expense of the deployment as reward and punishment feedback. After training, the target position of the virtual network function where to be deployed can be determined according to the network status. The simulation experiment verifies correctness and performance of this algorithm. Experiment shows that this algorithm can effectively reduce the average network delay and deployment failure ratio, and has certain advantages in algorithm running time.
Keywords:NFV  SFC  deep reinforcement learning  network delay  OPEX
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号