通过实时电子商务应用中无人机覆盖的云强化学习,实现利润最大化

IF 3.5 2区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Simulation Modelling Practice and Theory Pub Date : 2024-05-25 DOI:10.1016/j.simpat.2024.102970
Haythem Bany Salameh , Ghaleb Elrefae , Mohannad Alhafnawi , Yaser Jararweh , Ayat Alkhdour , Sharief Abdel-Razeq
{"title":"通过实时电子商务应用中无人机覆盖的云强化学习,实现利润最大化","authors":"Haythem Bany Salameh ,&nbsp;Ghaleb Elrefae ,&nbsp;Mohannad Alhafnawi ,&nbsp;Yaser Jararweh ,&nbsp;Ayat Alkhdour ,&nbsp;Sharief Abdel-Razeq","doi":"10.1016/j.simpat.2024.102970","DOIUrl":null,"url":null,"abstract":"<div><p>Real-time e-business applications are vital for operational efficiency, but connectivity challenges persist, particularly in remote or crowded areas. Drone Base Station (DBS) architecture, proposed for Beyond fifth Generation (B5G) and Sixth Generation (6G) multi-cell networks, offers on-demand hotspot coverage, addressing connectivity gaps in remote or crowded environments. DBSs provide a promising solution to meet the demanding requirements of high data rates, real-time responsiveness, low latency, and extended network coverage, particularly for real-time e-business applications. A critical challenge in this context involves efficiently allocating the needed number of DBSs to the different hotspot service areas, referred to as cells, to optimize the operator’s total profit under unpredictable user demands, varying area-specific service costs, and price dependence real-time e-service. The objective is to achieve the highest total revenue while minimizing the cost (cost savings) throughout the multi-cell system. This challenge is formulated as a profit-maximization discount return problem, integrating the coverage constraint, the variable cell-dependent operational cost, the e-service-based price and the uncertain demands of users across cells. Traditional optimization methods fail due to environmental uncertainty, which leads to the need to reformulate the problem as a Markov Decision Problem (MDP). We introduce a cloud-based Reinforcement Learning (RL) algorithm for DBS dispatch to address the MDP formulation. This algorithm dynamically adjusts to uncertain per-cell user distributions, considering variable operating costs and service-dependent prices across cells. Through extensive evaluation, the RL-based dispatch approach is compared with reference drone dispatch algorithms, demonstrating superior performance in maximizing operator profit through <em>cost savings</em> by optimizing DBS dispatch decisions based on learned user behaviors, variable operational costs, and e-service types.</p></div>","PeriodicalId":49518,"journal":{"name":"Simulation Modelling Practice and Theory","volume":null,"pages":null},"PeriodicalIF":3.5000,"publicationDate":"2024-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Maximizing profitability through cloud-enabled Reinforcement Learning for UAV coverage in real-time e-business applications\",\"authors\":\"Haythem Bany Salameh ,&nbsp;Ghaleb Elrefae ,&nbsp;Mohannad Alhafnawi ,&nbsp;Yaser Jararweh ,&nbsp;Ayat Alkhdour ,&nbsp;Sharief Abdel-Razeq\",\"doi\":\"10.1016/j.simpat.2024.102970\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Real-time e-business applications are vital for operational efficiency, but connectivity challenges persist, particularly in remote or crowded areas. Drone Base Station (DBS) architecture, proposed for Beyond fifth Generation (B5G) and Sixth Generation (6G) multi-cell networks, offers on-demand hotspot coverage, addressing connectivity gaps in remote or crowded environments. DBSs provide a promising solution to meet the demanding requirements of high data rates, real-time responsiveness, low latency, and extended network coverage, particularly for real-time e-business applications. A critical challenge in this context involves efficiently allocating the needed number of DBSs to the different hotspot service areas, referred to as cells, to optimize the operator’s total profit under unpredictable user demands, varying area-specific service costs, and price dependence real-time e-service. The objective is to achieve the highest total revenue while minimizing the cost (cost savings) throughout the multi-cell system. This challenge is formulated as a profit-maximization discount return problem, integrating the coverage constraint, the variable cell-dependent operational cost, the e-service-based price and the uncertain demands of users across cells. Traditional optimization methods fail due to environmental uncertainty, which leads to the need to reformulate the problem as a Markov Decision Problem (MDP). We introduce a cloud-based Reinforcement Learning (RL) algorithm for DBS dispatch to address the MDP formulation. This algorithm dynamically adjusts to uncertain per-cell user distributions, considering variable operating costs and service-dependent prices across cells. Through extensive evaluation, the RL-based dispatch approach is compared with reference drone dispatch algorithms, demonstrating superior performance in maximizing operator profit through <em>cost savings</em> by optimizing DBS dispatch decisions based on learned user behaviors, variable operational costs, and e-service types.</p></div>\",\"PeriodicalId\":49518,\"journal\":{\"name\":\"Simulation Modelling Practice and Theory\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":3.5000,\"publicationDate\":\"2024-05-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Simulation Modelling Practice and Theory\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1569190X24000844\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Simulation Modelling Practice and Theory","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1569190X24000844","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

摘要

实时电子商务应用对提高运营效率至关重要,但连接方面的挑战依然存在,尤其是在偏远或拥挤地区。为超越第五代(B5G)和第六代(6G)多蜂窝网络提出的无人机基站(DBS)架构可提供按需热点覆盖,解决偏远或拥挤环境中的连接缺口。DBS 为满足高数据速率、实时响应、低延迟和扩展网络覆盖等苛刻要求,尤其是实时电子商务应用的要求,提供了一种前景广阔的解决方案。在这种情况下,一个关键的挑战是在不可预测的用户需求、不同区域的服务成本以及实时电子服务的价格依赖性条件下,有效地将所需数量的 DBS 分配到不同的热点服务区域(称为小区),以优化运营商的总利润。目标是在整个多小区系统中实现最高的总收入,同时最大限度地降低成本(节约成本)。这一挑战被表述为一个利润最大化折扣回报问题,其中整合了覆盖范围约束、与小区相关的可变运营成本、基于电子服务的价格以及各小区用户的不确定需求。传统的优化方法因环境的不确定性而失效,因此需要将该问题重新表述为马尔可夫决策问题(Markov Decision Problem,MDP)。我们为 DBS 调度引入了基于云的强化学习(RL)算法,以解决 MDP 问题。该算法可根据不确定的每小区用户分布进行动态调整,同时考虑到各小区的可变运营成本和与服务相关的价格。通过广泛的评估,将基于 RL 的调度方法与参考的无人机调度算法进行了比较,结果表明,基于学习到的用户行为、可变运营成本和电子服务类型优化 DBS 调度决策,在通过节约成本实现运营商利润最大化方面表现出色。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Maximizing profitability through cloud-enabled Reinforcement Learning for UAV coverage in real-time e-business applications

Real-time e-business applications are vital for operational efficiency, but connectivity challenges persist, particularly in remote or crowded areas. Drone Base Station (DBS) architecture, proposed for Beyond fifth Generation (B5G) and Sixth Generation (6G) multi-cell networks, offers on-demand hotspot coverage, addressing connectivity gaps in remote or crowded environments. DBSs provide a promising solution to meet the demanding requirements of high data rates, real-time responsiveness, low latency, and extended network coverage, particularly for real-time e-business applications. A critical challenge in this context involves efficiently allocating the needed number of DBSs to the different hotspot service areas, referred to as cells, to optimize the operator’s total profit under unpredictable user demands, varying area-specific service costs, and price dependence real-time e-service. The objective is to achieve the highest total revenue while minimizing the cost (cost savings) throughout the multi-cell system. This challenge is formulated as a profit-maximization discount return problem, integrating the coverage constraint, the variable cell-dependent operational cost, the e-service-based price and the uncertain demands of users across cells. Traditional optimization methods fail due to environmental uncertainty, which leads to the need to reformulate the problem as a Markov Decision Problem (MDP). We introduce a cloud-based Reinforcement Learning (RL) algorithm for DBS dispatch to address the MDP formulation. This algorithm dynamically adjusts to uncertain per-cell user distributions, considering variable operating costs and service-dependent prices across cells. Through extensive evaluation, the RL-based dispatch approach is compared with reference drone dispatch algorithms, demonstrating superior performance in maximizing operator profit through cost savings by optimizing DBS dispatch decisions based on learned user behaviors, variable operational costs, and e-service types.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Simulation Modelling Practice and Theory
Simulation Modelling Practice and Theory 工程技术-计算机:跨学科应用
CiteScore
9.80
自引率
4.80%
发文量
142
审稿时长
21 days
期刊介绍: The journal Simulation Modelling Practice and Theory provides a forum for original, high-quality papers dealing with any aspect of systems simulation and modelling. The journal aims at being a reference and a powerful tool to all those professionally active and/or interested in the methods and applications of simulation. Submitted papers will be peer reviewed and must significantly contribute to modelling and simulation in general or use modelling and simulation in application areas. Paper submission is solicited on: • theoretical aspects of modelling and simulation including formal modelling, model-checking, random number generators, sensitivity analysis, variance reduction techniques, experimental design, meta-modelling, methods and algorithms for validation and verification, selection and comparison procedures etc.; • methodology and application of modelling and simulation in any area, including computer systems, networks, real-time and embedded systems, mobile and intelligent agents, manufacturing and transportation systems, management, engineering, biomedical engineering, economics, ecology and environment, education, transaction handling, etc.; • simulation languages and environments including those, specific to distributed computing, grid computing, high performance computers or computer networks, etc.; • distributed and real-time simulation, simulation interoperability; • tools for high performance computing simulation, including dedicated architectures and parallel computing.
期刊最新文献
Machine learning-assisted microscopic public transportation simulation: Two coupling strategies A novel energy-efficient and cost-effective task offloading approach for UAV-enabled MEC with LEO enhancement in Internet of Remote Things networks An AI-driven solution to prevent adversarial attacks on mobile Vehicle-to-Microgrid services Advancements in traffic simulation for enhanced road safety: A review Investigation on directional rock fracture mechanism under instantaneous expansion from the perspective of damage mechanics: A 3-D simulation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1