{"title":"模型不确定情况下鲁棒性标题压缩(ROHC)的强化学习","authors":"Shusen Jing;Songyang Zhang;Zhi Ding","doi":"10.1109/TMLCN.2024.3409200","DOIUrl":null,"url":null,"abstract":"Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1033-1044"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547320","citationCount":"0","resultStr":"{\"title\":\"Reinforcement Learning for Robust Header Compression (ROHC) Under Model Uncertainty\",\"authors\":\"Shusen Jing;Songyang Zhang;Zhi Ding\",\"doi\":\"10.1109/TMLCN.2024.3409200\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.\",\"PeriodicalId\":100641,\"journal\":{\"name\":\"IEEE Transactions on Machine Learning in Communications and Networking\",\"volume\":\"2 \",\"pages\":\"1033-1044\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547320\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Machine Learning in Communications and Networking\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10547320/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Machine Learning in Communications and Networking","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10547320/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Reinforcement Learning for Robust Header Compression (ROHC) Under Model Uncertainty
Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.