{"title":"基于繁忙程度的深度强化学习方法,用于弹性光网络的路由、调制和频谱分配","authors":"Chengsheng Liang, Yuqi Tu, Yue-Cai Huang","doi":"10.1117/12.3007879","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning (DRL) has been introduced in routing, modulation and spectrum assignment (RMSA) of the elastic optical networks. Since the DRL agent’s learning is based on the state it observes and the reward it receives, key information should be embedded in the state and the reward. In previous studies, the observed and feedback information is limited. In this paper, we propose a busyness level-based DRL method for the RMSA of the elastic optical networks. Since the busyness of the links or transmission paths highly affects the performance, we believe the busyness information should be perceived by the agent to learn a good RMSA policy. Specifically, we define two indicators to quantify busyness level, and then combine these two indicators into the design of reward and state. Simulation results show that our approach works better than the case that busyness is not","PeriodicalId":502341,"journal":{"name":"Applied Optics and Photonics China","volume":"11 3","pages":"1296624 - 1296624-7"},"PeriodicalIF":0.0000,"publicationDate":"2023-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Busyness level-based deep reinforcement learning method for routing, modulation, and spectrum assignment of elastic optical networks\",\"authors\":\"Chengsheng Liang, Yuqi Tu, Yue-Cai Huang\",\"doi\":\"10.1117/12.3007879\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep reinforcement learning (DRL) has been introduced in routing, modulation and spectrum assignment (RMSA) of the elastic optical networks. Since the DRL agent’s learning is based on the state it observes and the reward it receives, key information should be embedded in the state and the reward. In previous studies, the observed and feedback information is limited. In this paper, we propose a busyness level-based DRL method for the RMSA of the elastic optical networks. Since the busyness of the links or transmission paths highly affects the performance, we believe the busyness information should be perceived by the agent to learn a good RMSA policy. Specifically, we define two indicators to quantify busyness level, and then combine these two indicators into the design of reward and state. Simulation results show that our approach works better than the case that busyness is not\",\"PeriodicalId\":502341,\"journal\":{\"name\":\"Applied Optics and Photonics China\",\"volume\":\"11 3\",\"pages\":\"1296624 - 1296624-7\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-12-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Optics and Photonics China\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.3007879\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Optics and Photonics China","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3007879","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Busyness level-based deep reinforcement learning method for routing, modulation, and spectrum assignment of elastic optical networks
Deep reinforcement learning (DRL) has been introduced in routing, modulation and spectrum assignment (RMSA) of the elastic optical networks. Since the DRL agent’s learning is based on the state it observes and the reward it receives, key information should be embedded in the state and the reward. In previous studies, the observed and feedback information is limited. In this paper, we propose a busyness level-based DRL method for the RMSA of the elastic optical networks. Since the busyness of the links or transmission paths highly affects the performance, we believe the busyness information should be perceived by the agent to learn a good RMSA policy. Specifically, we define two indicators to quantify busyness level, and then combine these two indicators into the design of reward and state. Simulation results show that our approach works better than the case that busyness is not