Lingling Chen, Ziwei Wang, Xiaohui Zhao, Xuan Shen, Wei He
{"title":"A dynamic spectrum access algorithm based on deep reinforcement learning with novel multi-vehicle reward functions in cognitive vehicular networks","authors":"Lingling Chen, Ziwei Wang, Xiaohui Zhao, Xuan Shen, Wei He","doi":"10.1007/s11235-024-01188-5","DOIUrl":null,"url":null,"abstract":"<p>As a revolution in the field of transportation, the demand for communication of vehicles is increasing. Therefore, how to improve the success rate of vehicle spectrum access has become a major problem to be solved. The case of a single vehicle accessing a channel was only considered in the previous research on dynamic spectrum access in cognitive vehicular networks (CVNs), and the spectrum resources could not be fully utilized. In order to fully utilize spectrum resources, a model for spectrum sharing among multiple secondary vehicles (SVs) and a primary vehicle (PV) is proposed. This model includes scenarios where multiple SVs share spectrum to maximize the average quality of service (QoS) for vehicles. And the condition is considered that the total interference generated by vehicles accessing the same channel is less than the interference threshold. In this paper, a deep Q-network method with a modified reward function (IDQN) algorithm is proposed to maximize the average QoS of PVs and SVs and improve spectrum utilization. The algorithm is designed with different reward functions according to the QoS of PVs and SVs under different situations. Finally, the proposed algorithm is compared with the deep Q-network (DQN) and Q-learning algorithms under the Python simulation platform. The average access success rate of SVs in the IDQN algorithm proposed can reach 98<span>\\(\\%\\)</span>, which is improved by 18<span>\\(\\%\\)</span> compared with the Q-learning algorithm. And the convergence speed is 62.5<span>\\(\\%\\)</span> faster than the DQN algorithm. At the same time, the average QoS of PVs and the average QoS of SVs in the IDQN algorithm can reach 2.4, which is improved by 50<span>\\(\\%\\)</span> and 33<span>\\(\\%\\)</span> compared with the DQN algorithm, and improved by 60<span>\\(\\%\\)</span> and 140<span>\\(\\%\\)</span> compared with the Q-learning algorithm.</p>","PeriodicalId":51194,"journal":{"name":"Telecommunication Systems","volume":"357 1","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2024-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Telecommunication Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s11235-024-01188-5","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
As a revolution in the field of transportation, the demand for communication of vehicles is increasing. Therefore, how to improve the success rate of vehicle spectrum access has become a major problem to be solved. The case of a single vehicle accessing a channel was only considered in the previous research on dynamic spectrum access in cognitive vehicular networks (CVNs), and the spectrum resources could not be fully utilized. In order to fully utilize spectrum resources, a model for spectrum sharing among multiple secondary vehicles (SVs) and a primary vehicle (PV) is proposed. This model includes scenarios where multiple SVs share spectrum to maximize the average quality of service (QoS) for vehicles. And the condition is considered that the total interference generated by vehicles accessing the same channel is less than the interference threshold. In this paper, a deep Q-network method with a modified reward function (IDQN) algorithm is proposed to maximize the average QoS of PVs and SVs and improve spectrum utilization. The algorithm is designed with different reward functions according to the QoS of PVs and SVs under different situations. Finally, the proposed algorithm is compared with the deep Q-network (DQN) and Q-learning algorithms under the Python simulation platform. The average access success rate of SVs in the IDQN algorithm proposed can reach 98\(\%\), which is improved by 18\(\%\) compared with the Q-learning algorithm. And the convergence speed is 62.5\(\%\) faster than the DQN algorithm. At the same time, the average QoS of PVs and the average QoS of SVs in the IDQN algorithm can reach 2.4, which is improved by 50\(\%\) and 33\(\%\) compared with the DQN algorithm, and improved by 60\(\%\) and 140\(\%\) compared with the Q-learning algorithm.
期刊介绍:
Telecommunication Systems is a journal covering all aspects of modeling, analysis, design and management of telecommunication systems. The journal publishes high quality articles dealing with the use of analytic and quantitative tools for the modeling, analysis, design and management of telecommunication systems covering:
Performance Evaluation of Wide Area and Local Networks;
Network Interconnection;
Wire, wireless, Adhoc, mobile networks;
Impact of New Services (economic and organizational impact);
Fiberoptics and photonic switching;
DSL, ADSL, cable TV and their impact;
Design and Analysis Issues in Metropolitan Area Networks;
Networking Protocols;
Dynamics and Capacity Expansion of Telecommunication Systems;
Multimedia Based Systems, Their Design Configuration and Impact;
Configuration of Distributed Systems;
Pricing for Networking and Telecommunication Services;
Performance Analysis of Local Area Networks;
Distributed Group Decision Support Systems;
Configuring Telecommunication Systems with Reliability and Availability;
Cost Benefit Analysis and Economic Impact of Telecommunication Systems;
Standardization and Regulatory Issues;
Security, Privacy and Encryption in Telecommunication Systems;
Cellular, Mobile and Satellite Based Systems.