Pub Date : 2026-01-27DOI: 10.1109/JSAC.2025.3636429
{"title":"IEEE Communications Society Information","authors":"","doi":"10.1109/JSAC.2025.3636429","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3636429","url":null,"abstract":"","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 12","pages":"C3-C3"},"PeriodicalIF":17.2,"publicationDate":"2026-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11366032","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146049300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-21DOI: 10.1109/JSAC.2025.3611419
Seungnyun Kim;Subham Saha;Seokhyun Jeong;Byonghyo Shim;Moe Z. Win
Beam management is an essential operation of next-generation (xG) wireless networks to compensate severe signal attenuation and ensure reliable communications over millimeter wave (mmWave) and terahertz (THz) bands. The primary objective of beam management is to determine beam directions that are properly aligned with the signal propagation paths. Conventional beam management techniques typically rely on geometric channel parameters such as angles, delays, and path gains. However, due to their lack of contextual awareness of the surrounding environment, these techniques fall short in handling the piecewise continuous changes in the geometric channel parameters caused by sudden obstructions or variations in scatterers. In this paper, we propose a novel beam management framework that utilizes environmental information extracted from sensor data (e.g., images) and pilot measurements to optimize beam directions. The main idea of the proposed scheme is to utilize visual channel parameters, including the positions of user equipment (UE), reflection points, and scatterers, which provide a direct visualization of the propagation environment. By tracking the visual channel parameters, dynamic changes in scatterers and propagation paths can be monitored over time. To analyze multimodal data and track these parameters, we exploit large multimodal model (LMM), a generative artificial intelligence (AI) model specialized in extracting correlated features across multimodal data and generating subsequent data. Simulation results show that the proposed scheme can accurately track the visual channel parameters and enhance data rate.
{"title":"Large Multimodal Model-Based Environment-Aware Beam Management","authors":"Seungnyun Kim;Subham Saha;Seokhyun Jeong;Byonghyo Shim;Moe Z. Win","doi":"10.1109/JSAC.2025.3611419","DOIUrl":"10.1109/JSAC.2025.3611419","url":null,"abstract":"Beam management is an essential operation of next-generation (xG) wireless networks to compensate severe signal attenuation and ensure reliable communications over millimeter wave (mmWave) and terahertz (THz) bands. The primary objective of beam management is to determine beam directions that are properly aligned with the signal propagation paths. Conventional beam management techniques typically rely on geometric channel parameters such as angles, delays, and path gains. However, due to their lack of contextual awareness of the surrounding environment, these techniques fall short in handling the piecewise continuous changes in the geometric channel parameters caused by sudden obstructions or variations in scatterers. In this paper, we propose a novel beam management framework that utilizes environmental information extracted from sensor data (e.g., images) and pilot measurements to optimize beam directions. The main idea of the proposed scheme is to utilize visual channel parameters, including the positions of user equipment (UE), reflection points, and scatterers, which provide a direct visualization of the propagation environment. By tracking the visual channel parameters, dynamic changes in scatterers and propagation paths can be monitored over time. To analyze multimodal data and track these parameters, we exploit large multimodal model (LMM), a generative artificial intelligence (AI) model specialized in extracting correlated features across multimodal data and generating subsequent data. Simulation results show that the proposed scheme can accurately track the visual channel parameters and enhance data rate.","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"44 ","pages":"991-1007"},"PeriodicalIF":17.2,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-06DOI: 10.1109/JSAC.2025.3625035
{"title":"IEEE Journal on Selected Areas in Communications Publication Information","authors":"","doi":"10.1109/JSAC.2025.3625035","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3625035","url":null,"abstract":"","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 11","pages":"C2-C2"},"PeriodicalIF":17.2,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11230337","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145449381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-06DOI: 10.1109/JSAC.2025.3583353
Carlos Cordeiro;Edward Knightly;Giovanni Geraci;Joerg Widmer;Malcolm Smith;V. K. Jones
{"title":"Guest Editorial: The Future of Wi-Fi and Wireless Technologies in Unlicensed Spectra","authors":"Carlos Cordeiro;Edward Knightly;Giovanni Geraci;Joerg Widmer;Malcolm Smith;V. K. Jones","doi":"10.1109/JSAC.2025.3583353","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3583353","url":null,"abstract":"","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 11","pages":"3612-3615"},"PeriodicalIF":17.2,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11230360","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145449379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-06DOI: 10.1109/JSAC.2025.3625037
{"title":"IEEE Communications Society Information","authors":"","doi":"10.1109/JSAC.2025.3625037","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3625037","url":null,"abstract":"","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 11","pages":"C3-C3"},"PeriodicalIF":17.2,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11230362","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145449325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-28DOI: 10.1109/JSAC.2025.3623184
Sige Liu;Nan Li;Yansha Deng;Tony Q. S. Quek
The rapid progress of artificial intelligence (AI) and computer vision (CV) has facilitated the development of computation-intensive applications like Visual Question Answering (VQA), which integrates visual perception and natural language processing to generate answers. To overcome the limitations of traditional VQA constrained by local computation resources, edge computing has been incorporated to provide extra computation capability at the edge side. Meanwhile, this brings new communication challenges between the local and edge, including limited bandwidth, channel noise, and multipath effects, which degrade VQA performance and user quality of experience (QoE), particularly during the transmission of large high-resolution images. To overcome these bottlenecks, we propose a goal-oriented semantic communication (GSC) framework that focuses on effectively extracting and transmitting semantic information most relevant to the VQA goals, improving the answering accuracy and enhancing the effectiveness and efficiency. The objective is to maximize the answering accuracy, and we propose a bounding box (BBox)-based image semantic extraction and ranking approach to prioritize the semantic information based on the goal of questions. We then extend it by incorporating a scene graphs (SG)-based approach to handle questions with complex relationships. Experimental results demonstrate that our GSC framework improves answering accuracy by up to 49% under AWGN channels and 59% under Rayleigh channels while reducing total latency by up to 65% compared to traditional bit-oriented transmission.
{"title":"Goal-Oriented Semantic Communication for Wireless Visual Question Answering","authors":"Sige Liu;Nan Li;Yansha Deng;Tony Q. S. Quek","doi":"10.1109/JSAC.2025.3623184","DOIUrl":"10.1109/JSAC.2025.3623184","url":null,"abstract":"The rapid progress of artificial intelligence (AI) and computer vision (CV) has facilitated the development of computation-intensive applications like Visual Question Answering (VQA), which integrates visual perception and natural language processing to generate answers. To overcome the limitations of traditional VQA constrained by local computation resources, edge computing has been incorporated to provide extra computation capability at the edge side. Meanwhile, this brings new communication challenges between the local and edge, including limited bandwidth, channel noise, and multipath effects, which degrade VQA performance and user quality of experience (QoE), particularly during the transmission of large high-resolution images. To overcome these bottlenecks, we propose a goal-oriented semantic communication (GSC) framework that focuses on effectively extracting and transmitting semantic information most relevant to the VQA goals, improving the answering accuracy and enhancing the effectiveness and efficiency. The objective is to maximize the answering accuracy, and we propose a bounding box (BBox)-based image semantic extraction and ranking approach to prioritize the semantic information based on the goal of questions. We then extend it by incorporating a scene graphs (SG)-based approach to handle questions with complex relationships. Experimental results demonstrate that our GSC framework improves answering accuracy by up to 49% under AWGN channels and 59% under Rayleigh channels while reducing total latency by up to 65% compared to traditional bit-oriented transmission.","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 12","pages":"4247-4261"},"PeriodicalIF":17.2,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145381422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-27DOI: 10.1109/JSAC.2025.3623161
Halar Haleem;Igor Bisio;Chiara Garibotto;Fabio Lavagetto;Andrea Sciarrone;Nafeeul Alam Walee;Atef Mohamed Shalan;Lei Chen;Yiming Ji
Drones are integral to various applications, out of which traffic surveillance is an important application. However, their operational efficiency is limited by battery life, which restricts their capacity for extended critical missions. Additionally, in remote or high-interference areas, the bandwidth for drone communication is often limited, leading to a decrease in the quality of images transmitted to the base station. This paper aims to address such challenges by having drones transmit video data in real-time at lower resolutions for traffic monitoring. This approach conserves energy and optimizes transmission. However, it adversely affects object detection accuracy at the base station due to compromised data quality. To address this issue, we incorporate Generative Adversarial Networks (GANs) to improve LR images, restoring their quality for precise object detection. Results indicate that the accuracy of traffic analytics achieved with GAN-enhanced images is comparable to that obtained with high-resolution data transmission. Consequently, our approach allows a fundamental trade-off among drone energy consumption, transmission time, flight time, and object detection accuracy, enabling robust detection performance while conserving energy and enhancing operational capabilities.
{"title":"Toward Intelligent Traffic Monitoring System Exploiting GANs-Based Models for Real-Time UAV Data","authors":"Halar Haleem;Igor Bisio;Chiara Garibotto;Fabio Lavagetto;Andrea Sciarrone;Nafeeul Alam Walee;Atef Mohamed Shalan;Lei Chen;Yiming Ji","doi":"10.1109/JSAC.2025.3623161","DOIUrl":"10.1109/JSAC.2025.3623161","url":null,"abstract":"Drones are integral to various applications, out of which traffic surveillance is an important application. However, their operational efficiency is limited by battery life, which restricts their capacity for extended critical missions. Additionally, in remote or high-interference areas, the bandwidth for drone communication is often limited, leading to a decrease in the quality of images transmitted to the base station. This paper aims to address such challenges by having drones transmit video data in real-time at lower resolutions for traffic monitoring. This approach conserves energy and optimizes transmission. However, it adversely affects object detection accuracy at the base station due to compromised data quality. To address this issue, we incorporate Generative Adversarial Networks (GANs) to improve LR images, restoring their quality for precise object detection. Results indicate that the accuracy of traffic analytics achieved with GAN-enhanced images is comparable to that obtained with high-resolution data transmission. Consequently, our approach allows a fundamental trade-off among drone energy consumption, transmission time, flight time, and object detection accuracy, enabling robust detection performance while conserving energy and enhancing operational capabilities.","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 12","pages":"4277-4293"},"PeriodicalIF":17.2,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145381441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-23DOI: 10.1109/JSAC.2025.3615342
{"title":"IEEE Communications Society Information","authors":"","doi":"10.1109/JSAC.2025.3615342","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3615342","url":null,"abstract":"","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 10","pages":"C3-C3"},"PeriodicalIF":17.2,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11214504","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145351980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-20DOI: 10.1109/JSAC.2025.3623175
Songjie Xie;Hengtao He;Shenghui Song;Jun Zhang;Khaled B. Letaief
Task-oriented communication presents a promising approach to improve the communication efficiency of edge inference systems by optimizing learning-based modules to extract and transmit relevant task information. However, real-time applications face practical challenges, such as incomplete coverage and potential malfunctions of edge servers. This situation necessitates cross-model communication between different inference systems, enabling edge devices from one service provider to collaborate effectively with edge servers from another. Independent optimization of diverse edge systems often leads to incoherent feature spaces, which hinders the cross-model inference for existing task-oriented communication. To facilitate and achieve effective cross-model task-oriented communication, this study introduces a novel framework that utilizes shared anchor data across diverse systems. This approach addresses the challenge of feature alignment in both server-based and on-device scenarios. In particular, by leveraging the linear invariance of visual features, we propose efficient server-based feature alignment techniques to estimate linear transformations using encoded anchor data features. For on-device alignment, we exploit the angle-preserving nature of visual features and propose to encode relative representations with anchor data to streamline cross-model communication without additional alignment procedures during the inference. The experimental results on computer vision benchmarks demonstrate the superior performance of the proposed feature alignment approaches in cross-model task-oriented communications. The runtime and computation overhead analysis further confirm the effectiveness of the proposed feature alignment approaches in real-time applications.
{"title":"Toward Real-Time Edge AI: Model-Agnostic Task-Oriented Communication With Visual Feature Alignment","authors":"Songjie Xie;Hengtao He;Shenghui Song;Jun Zhang;Khaled B. Letaief","doi":"10.1109/JSAC.2025.3623175","DOIUrl":"https://doi.org/10.1109/JSAC.2025.3623175","url":null,"abstract":"Task-oriented communication presents a promising approach to improve the communication efficiency of edge inference systems by optimizing learning-based modules to extract and transmit relevant task information. However, real-time applications face practical challenges, such as incomplete coverage and potential malfunctions of edge servers. This situation necessitates cross-model communication between different inference systems, enabling edge devices from one service provider to collaborate effectively with edge servers from another. Independent optimization of diverse edge systems often leads to incoherent feature spaces, which hinders the cross-model inference for existing task-oriented communication. To facilitate and achieve effective cross-model task-oriented communication, this study introduces a novel framework that utilizes shared anchor data across diverse systems. This approach addresses the challenge of feature alignment in both server-based and on-device scenarios. In particular, by leveraging the linear invariance of visual features, we propose efficient server-based feature alignment techniques to estimate linear transformations using encoded anchor data features. For on-device alignment, we exploit the angle-preserving nature of visual features and propose to encode relative representations with anchor data to streamline cross-model communication without additional alignment procedures during the inference. The experimental results on computer vision benchmarks demonstrate the superior performance of the proposed feature alignment approaches in cross-model task-oriented communications. The runtime and computation overhead analysis further confirm the effectiveness of the proposed feature alignment approaches in real-time applications.","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"43 12","pages":"4262-4276"},"PeriodicalIF":17.2,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146049289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Respiration monitoring via radio signals enables contactless health sensing but suffers from interference caused by nearby motion. We propose a robust respiration sensing framework using Cell-free Massive MIMO (CF-mMIMO), which leverages spatial macro-diversity for interference resilience. Specifically, we analyze respiration sensing in single-antenna channels using Power Spectral Density (PSD) to reveal the impact of interference on the breathing channel’s movement spectrum. Based on this, we introduce a new metric, Sensing-Signal-to-Interference Ratio (SSIR), to evaluate local channel quality without requiring ground truth. Then, we design a Weighted Antenna Combining (WAC) method to prioritize reliable sensing links and suppress distortion. Experimental validation using a 64-antenna CF-mMIMO testbed with 100 Orthogonal Frequency-Division Multiplexing (OFDM) subcarriers over an 18 MHz bandwidth confirms the framework’s robustness. In the presence of interference, the WAC method achieves a mean waveform correlation of 0.81 with ground truth, significantly outperforming single-antenna (0.52), averaging-based methods (0.53), and existing Wi-Fi approaches. Finally, we analyze the impact of time, frequency, and spatial resource allocation on both communication and sensing performance. Results show that increasing bandwidth and antenna count benefits both communication and sensing. With a sufficient number of antennas, respiration sensing remains accurate even with long coherence times (1 second) and narrow bandwidths (3 subcarriers), enabling its integration into communication systems with negligible overhead, making it practically “for free”. This makes CF-mMIMO a promising architecture for robust and scalable Integrated Sensing and Communication (ISAC) health monitoring.
{"title":"Fundamentals and Experiments of Robust Respiration Sensing via Cell-Free Massive MIMO","authors":"Haoqiu Xiong;Robbert Beerten;Qing Zhang;Yang Miao;Zhuangzhuang Cui;Sofie Pollin","doi":"10.1109/JSAC.2025.3617012","DOIUrl":"10.1109/JSAC.2025.3617012","url":null,"abstract":"Respiration monitoring via radio signals enables contactless health sensing but suffers from interference caused by nearby motion. We propose a robust respiration sensing framework using Cell-free Massive MIMO (CF-mMIMO), which leverages spatial macro-diversity for interference resilience. Specifically, we analyze respiration sensing in single-antenna channels using Power Spectral Density (PSD) to reveal the impact of interference on the breathing channel’s movement spectrum. Based on this, we introduce a new metric, Sensing-Signal-to-Interference Ratio (SSIR), to evaluate local channel quality without requiring ground truth. Then, we design a Weighted Antenna Combining (WAC) method to prioritize reliable sensing links and suppress distortion. Experimental validation using a 64-antenna CF-mMIMO testbed with 100 Orthogonal Frequency-Division Multiplexing (OFDM) subcarriers over an 18 MHz bandwidth confirms the framework’s robustness. In the presence of interference, the WAC method achieves a mean waveform correlation of 0.81 with ground truth, significantly outperforming single-antenna (0.52), averaging-based methods (0.53), and existing Wi-Fi approaches. Finally, we analyze the impact of time, frequency, and spatial resource allocation on both communication and sensing performance. Results show that increasing bandwidth and antenna count benefits both communication and sensing. With a sufficient number of antennas, respiration sensing remains accurate even with long coherence times (1 second) and narrow bandwidths (3 subcarriers), enabling its integration into communication systems with negligible overhead, making it practically “for free”. This makes CF-mMIMO a promising architecture for robust and scalable Integrated Sensing and Communication (ISAC) health monitoring.","PeriodicalId":73294,"journal":{"name":"IEEE journal on selected areas in communications : a publication of the IEEE Communications Society","volume":"44 ","pages":"959-974"},"PeriodicalIF":17.2,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145215671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}