{"title":"Performance comparison of explainable DQN and DDPG models for cooperative lane change decision-making in multi-intelligent industrial IoT vehicles","authors":"Hao-bai ZHAN","doi":"10.1016/j.iot.2025.101552","DOIUrl":null,"url":null,"abstract":"<div><div>With the rapid advancement of intelligent connected vehicles (ICVs) technology, efficient and safe vehicular lane-changing decisions have become a focal point of interest for intelligent transportation systems (ITS). This paper investigates the application of explainable artificial intelligence (XAI) techniques to deep reinforcement learning algorithms, specifically deep Q-networks (DQN) and deep deterministic policy gradient (DDPG), for lane-changing decisions in industrial internet of things (IIoT) vehicles. By integrating innovative reward functions, the study assesses the performance differences between these models under various traffic densities and ICV counts in a three-lane highway scenario. The use of XAI feature representations enhances the transparency and interpretability of the models, providing insights into the decision-making process. XAI helps to elucidate how the models arrive at their decisions, improving trust and reliability in automated systems. The research reveals that although the DQN model demonstrates initial superior performance in the early phases of experimentation, the DDPG model outperforms in crucial performance metrics such as average fleet speed, headway, and stability during later stages of training. The DDPG model maintains better control over fleet speed and vehicle spacing in both low-density and high-density traffic environments, showcasing its superior adaptability and efficiency. These findings highlight the DDPG model's enhanced capability to manage dynamic and complex driving environments, attributed to its refined policy learning approach which adeptly balances exploration and exploitation. The novel reward function significantly promotes cooperative lane-changing behaviors among ICVs, optimizing lane change decisions and improving overall traffic flow efficiency. This study not only provides valuable technical support for lane-changing decisions in smart vehicular networks but also lays a theoretical and empirical foundation for the advancement of future ITS. The insights gained from comparing DQN and DDPG models contribute to the ongoing discussion on effective deep learning strategies for real-world ITS applications, potentially guiding future developments in autonomous driving technologies.</div></div>","PeriodicalId":29968,"journal":{"name":"Internet of Things","volume":"31 ","pages":"Article 101552"},"PeriodicalIF":6.0000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet of Things","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2542660525000654","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
With the rapid advancement of intelligent connected vehicles (ICVs) technology, efficient and safe vehicular lane-changing decisions have become a focal point of interest for intelligent transportation systems (ITS). This paper investigates the application of explainable artificial intelligence (XAI) techniques to deep reinforcement learning algorithms, specifically deep Q-networks (DQN) and deep deterministic policy gradient (DDPG), for lane-changing decisions in industrial internet of things (IIoT) vehicles. By integrating innovative reward functions, the study assesses the performance differences between these models under various traffic densities and ICV counts in a three-lane highway scenario. The use of XAI feature representations enhances the transparency and interpretability of the models, providing insights into the decision-making process. XAI helps to elucidate how the models arrive at their decisions, improving trust and reliability in automated systems. The research reveals that although the DQN model demonstrates initial superior performance in the early phases of experimentation, the DDPG model outperforms in crucial performance metrics such as average fleet speed, headway, and stability during later stages of training. The DDPG model maintains better control over fleet speed and vehicle spacing in both low-density and high-density traffic environments, showcasing its superior adaptability and efficiency. These findings highlight the DDPG model's enhanced capability to manage dynamic and complex driving environments, attributed to its refined policy learning approach which adeptly balances exploration and exploitation. The novel reward function significantly promotes cooperative lane-changing behaviors among ICVs, optimizing lane change decisions and improving overall traffic flow efficiency. This study not only provides valuable technical support for lane-changing decisions in smart vehicular networks but also lays a theoretical and empirical foundation for the advancement of future ITS. The insights gained from comparing DQN and DDPG models contribute to the ongoing discussion on effective deep learning strategies for real-world ITS applications, potentially guiding future developments in autonomous driving technologies.
期刊介绍:
Internet of Things; Engineering Cyber Physical Human Systems is a comprehensive journal encouraging cross collaboration between researchers, engineers and practitioners in the field of IoT & Cyber Physical Human Systems. The journal offers a unique platform to exchange scientific information on the entire breadth of technology, science, and societal applications of the IoT.
The journal will place a high priority on timely publication, and provide a home for high quality.
Furthermore, IOT is interested in publishing topical Special Issues on any aspect of IOT.