Design of a Genetic Algorithm Based Dynamic Learning Method for Improved Channel Modelling in mmWave Radios via Temporal Breakpoint Analysis

IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Traitement Du Signal Pub Date : 2023-08-31 DOI:10.18280/ts.400420
A. Bhoi, V. Hendre
{"title":"Design of a Genetic Algorithm Based Dynamic Learning Method for Improved Channel Modelling in mmWave Radios via Temporal Breakpoint Analysis","authors":"A. Bhoi, V. Hendre","doi":"10.18280/ts.400420","DOIUrl":null,"url":null,"abstract":"Dynamic channel modelling allows communication interfaces to integrate continuous learning operations for incremental BER reductions. These models scan temporal BER patterns, and then tune internal-channel parameters in order to improving communication efficiency under real-time traffic scenarios. But these models showcase high complexity, thus cannot be scaled to large-scale network deployments. Moreover, these models are not flexible, and do not support denser channel models, which restricts their applicability under real-time scenarios. To overcome these issues, this text proposes design of a novel dynamic learning method for improved channel modelling in Phased array antennas mm Wave radios via temporal breakpoint analysis. The model initially collects information about channel BER and uses a Grey Wolf Optimization (GWO) technique to improve its internal model parameters. These parameters are further tuned via a novel breakpoint model, which enables for continuous and light-weighted tuning of channel modelling parameters. This allows the model to incrementally reduce BER even under denser noise levels. The model is further cascaded with a Q-Learning based optimization process, which assists in improving channel modelling efficiency for large-scale networks. Due to these integrations, the model is capable of reducing Bit Error Rate (BER) by 8.3% when compared with standard channel modelling techniques that use Convolutional Neural Networks (CNNs), Sparse Bayesian Learning, etc. These methods were selected for comparison due to their higher efficiency and scalability when applied to real-time communication scenarios. The model also showcased 6.5% lower computational delay due to linear processing operations. It was able to achieve 10.4% better channel coverage, 8.5% higher throughput, and 4.9% higher channel estimation accuracy, which makes it useful for a wide","PeriodicalId":49430,"journal":{"name":"Traitement Du Signal","volume":" ","pages":""},"PeriodicalIF":1.2000,"publicationDate":"2023-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Traitement Du Signal","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.18280/ts.400420","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Dynamic channel modelling allows communication interfaces to integrate continuous learning operations for incremental BER reductions. These models scan temporal BER patterns, and then tune internal-channel parameters in order to improving communication efficiency under real-time traffic scenarios. But these models showcase high complexity, thus cannot be scaled to large-scale network deployments. Moreover, these models are not flexible, and do not support denser channel models, which restricts their applicability under real-time scenarios. To overcome these issues, this text proposes design of a novel dynamic learning method for improved channel modelling in Phased array antennas mm Wave radios via temporal breakpoint analysis. The model initially collects information about channel BER and uses a Grey Wolf Optimization (GWO) technique to improve its internal model parameters. These parameters are further tuned via a novel breakpoint model, which enables for continuous and light-weighted tuning of channel modelling parameters. This allows the model to incrementally reduce BER even under denser noise levels. The model is further cascaded with a Q-Learning based optimization process, which assists in improving channel modelling efficiency for large-scale networks. Due to these integrations, the model is capable of reducing Bit Error Rate (BER) by 8.3% when compared with standard channel modelling techniques that use Convolutional Neural Networks (CNNs), Sparse Bayesian Learning, etc. These methods were selected for comparison due to their higher efficiency and scalability when applied to real-time communication scenarios. The model also showcased 6.5% lower computational delay due to linear processing operations. It was able to achieve 10.4% better channel coverage, 8.5% higher throughput, and 4.9% higher channel estimation accuracy, which makes it useful for a wide
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于遗传算法的毫米波信道建模改进动态学习方法设计
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Traitement Du Signal
Traitement Du Signal 工程技术-工程:电子与电气
自引率
21.10%
发文量
162
审稿时长
>12 weeks
期刊介绍: The TS provides rapid dissemination of original research in the field of signal processing, imaging and visioning. Since its founding in 1984, the journal has published articles that present original research results of a fundamental, methodological or applied nature. The editorial board welcomes articles on the latest and most promising results of academic research, including both theoretical results and case studies. The TS welcomes original research papers, technical notes and review articles on various disciplines, including but not limited to: Signal processing Imaging Visioning Control Filtering Compression Data transmission Noise reduction Deconvolution Prediction Identification Classification.
期刊最新文献
Hierarchical Spatial Feature-CNN Employing Grad-CAM for Enhanced Segmentation and Classification in Alzheimer's and Parkinson's Disease Diagnosis via MRI Massage Acupoint Positioning Method of Human Body Images Based on Transfer Learning Exploring the Application of Deep Learning in Multi-View Image Fusion in Complex Environments A Hybrid Diabetic Retinopathy Neural Network Model for Early Diabetic Retinopathy Detection and Classification of Fundus Images Leveraging Tripartite Tier Convolutional Neural Network for Human Emotion Recognition: A Multimodal Data Approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1