Elastic optical networks (EONs) face with more intense resource competition and difficulties in ensuring service quality for time-varying services. This study uses deep Q-network to optimize routing, modulation and spectrum allocation (DQN-RMSA) for time-varying services. The DQN is introduced to combine the deep reinforcement learning (DRL) framework and Q-network to manage RMSA by sliding windows for time-varying services. Specifically, a priority-aware state representation is adopted to integrate time-varying service characteristics and network resource state capturing by the DRL agent. An enhanced prioritized experience replay mechanism with a Sumtree structure is used to accelerate convergence by prioritizing critical learning samples. Moreover, a delayed action is designed by using Markov decision process (MDP) to monitor state transition for improving the performance of RMSA, and the delay action is reflected in the reward function to balance delay and resource utilization in terms of decreasing blocking probability. Simulation results demonstrate that DQN-RMSA can achieve superior performance over traditional heuristic and other DRL methods, with notable improvements in service blocking probability, delay tolerance, and spectrum utilization. The DQN-RMSA can adapt to fluctuations in time-varying service, highlighting its potential for achieving self-optimization and robust resource allocation in dynamic environments.
扫码关注我们
求助内容:
应助结果提醒方式:
