首页 > 最新文献

International journal of smart computing and artificial intelligence最新文献

英文 中文
An Extension of Particle Swarm Optimization to Identify Multiple Peaks using Re-diversification in Static and Dynamic Environments 静态和动态环境下利用再多样化识别多峰的粒子群算法的扩展
Pub Date : 2023-01-01 DOI: 10.52731/ijscai.v7.i2.793
Stephen Raharja, Toshiharu Sugawara
We propose an extension of the particle swarm optimization (PSO) algorithm for each particle to store multiple global optima internally for identifying multiple (top-k) peaks in static and dynamic environments. We then applied this technique to search and rescue problems of rescuing potential survivors urgently in life-threatening disaster scenarios. With the rapid development of robotics andcomputer technology, aerial drones can be programmed to implement search algorithms that locate potential survivors and relay their positions to rescue teams. We model an environment of a disaster area with potential survivors using randomizedbivariate normal distributions. We extended the Clerk-Kennedy PSO algorithm as top-k PSO by considering individual drones as particles, where each particle remembers a set of global optima to identify the top-k peaks. By comparing several otheralgorithms, including the canonical PSO, Clerk-Kennedy PSO, and NichePSO, we evaluated our proposed algorithm in static and dynamic environments. The experimental results show that the proposed algorithm was able to identify the top-kpeaks (optima) with a higher success rate than the baseline methods, although the rate gradually decreased with increasing movement speed of the peaks in dynamic environments.
我们提出了粒子群优化(PSO)算法的扩展,在每个粒子内部存储多个全局最优解,以识别静态和动态环境中的多个(top-k)峰值。然后,我们将这项技术应用于在危及生命的灾难场景中紧急营救潜在幸存者的搜索和救援问题。随着机器人技术和计算机技术的快速发展,空中无人机可以被编程来执行搜索算法,定位潜在的幸存者并将他们的位置传递给救援队。我们用随机双变量正态分布模拟了一个有潜在幸存者的灾区环境。我们将Clerk-Kennedy PSO算法扩展为top-k PSO算法,将单个无人机视为粒子,其中每个粒子记住一组全局最优值来识别top-k峰值。通过比较几种其他算法,包括canonical PSO、Clerk-Kennedy PSO和NichePSO,我们在静态和动态环境中评估了我们提出的算法。实验结果表明,尽管在动态环境中,随着峰值移动速度的增加,该算法的识别成功率会逐渐降低,但与基线方法相比,该算法能够以更高的成功率识别顶峰值。
{"title":"An Extension of Particle Swarm Optimization to Identify Multiple Peaks using Re-diversification in Static and Dynamic Environments","authors":"Stephen Raharja, Toshiharu Sugawara","doi":"10.52731/ijscai.v7.i2.793","DOIUrl":"https://doi.org/10.52731/ijscai.v7.i2.793","url":null,"abstract":"We propose an extension of the particle swarm optimization (PSO) algorithm for each particle to store multiple global optima internally for identifying multiple (top-k) peaks in static and dynamic environments. We then applied this technique to search and rescue problems of rescuing potential survivors urgently in life-threatening disaster scenarios. With the rapid development of robotics andcomputer technology, aerial drones can be programmed to implement search algorithms that locate potential survivors and relay their positions to rescue teams. We model an environment of a disaster area with potential survivors using randomizedbivariate normal distributions. We extended the Clerk-Kennedy PSO algorithm as top-k PSO by considering individual drones as particles, where each particle remembers a set of global optima to identify the top-k peaks. By comparing several otheralgorithms, including the canonical PSO, Clerk-Kennedy PSO, and NichePSO, we evaluated our proposed algorithm in static and dynamic environments. The experimental results show that the proposed algorithm was able to identify the top-kpeaks (optima) with a higher success rate than the baseline methods, although the rate gradually decreased with increasing movement speed of the peaks in dynamic environments.","PeriodicalId":495454,"journal":{"name":"International journal of smart computing and artificial intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135102880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Multi-Agent Reinforcement Learning for Beer Game by Reward Design Based on Payment Mechanism 基于支付机制的奖励设计改进啤酒博弈多智能体强化学习
Pub Date : 2023-01-01 DOI: 10.52731/ijscai.v7.i2.789
Masaaki Hori, Toshihiro Matsui
Supply chain management aims to maximize profits among supply chain partners by managing the flow of information and products. Multiagent reinforcement learning in artificial intelligence research fields has been applied to supply chain management. The beer game is an example problem in supply chain management and has also been studied as a cooperation problem in multiagent systems. In the previous study, a solution method SRDQN that is based on deep reinforcement learning and reward shaping has been applied to the beer game. By introducing a single reinforcement learning agent with SRDQN as a participant in the beer game, the cost of beer inventory was reduced. However, the previous study has not addressed the case of multiagent reinforcement learning due to the difficulties in cooperation among agents. To address the multiagent cases, we apply a reward shaping technique RDPM based on mechanism design to SRDQN and improve cooperative policies in multiagent reinforcement learning. Furthermore, we propose two reward design methods with modifications to the state value function designs in RDPM to address various consumer demands for beers in the supply chain. And then we empirically evaluate the effectiveness of the proposed approaches.
供应链管理的目的是通过管理信息和产品的流动,使供应链合作伙伴之间的利润最大化。人工智能研究领域中的多智能体强化学习已经应用到供应链管理中。啤酒博弈是供应链管理中的一个典型问题,也是多智能体系统中的一个合作问题。在之前的研究中,我们将一种基于深度强化学习和奖励塑造的SRDQN求解方法应用于啤酒博弈。通过引入SRDQN作为啤酒博弈参与者的单个强化学习智能体,降低了啤酒库存成本。然而,由于智能体之间的合作困难,以往的研究并没有解决多智能体强化学习的情况。为了解决多智能体案例,我们将基于机制设计的奖励塑造技术RDPM应用于SRDQN,并改进了多智能体强化学习中的合作策略。此外,我们提出了两种奖励设计方法,修改了RDPM中的状态价值函数设计,以解决供应链中消费者对啤酒的各种需求。然后对所提出方法的有效性进行了实证评价。
{"title":"Improving Multi-Agent Reinforcement Learning for Beer Game by Reward Design Based on Payment Mechanism","authors":"Masaaki Hori, Toshihiro Matsui","doi":"10.52731/ijscai.v7.i2.789","DOIUrl":"https://doi.org/10.52731/ijscai.v7.i2.789","url":null,"abstract":"Supply chain management aims to maximize profits among supply chain partners by managing the flow of information and products. Multiagent reinforcement learning in artificial intelligence research fields has been applied to supply chain management. The beer game is an example problem in supply chain management and has also been studied as a cooperation problem in multiagent systems. In the previous study, a solution method SRDQN that is based on deep reinforcement learning and reward shaping has been applied to the beer game. By introducing a single reinforcement learning agent with SRDQN as a participant in the beer game, the cost of beer inventory was reduced. However, the previous study has not addressed the case of multiagent reinforcement learning due to the difficulties in cooperation among agents. To address the multiagent cases, we apply a reward shaping technique RDPM based on mechanism design to SRDQN and improve cooperative policies in multiagent reinforcement learning. Furthermore, we propose two reward design methods with modifications to the state value function designs in RDPM to address various consumer demands for beers in the supply chain. And then we empirically evaluate the effectiveness of the proposed approaches.","PeriodicalId":495454,"journal":{"name":"International journal of smart computing and artificial intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135448375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Abstractive Summarization by Transfer Learning with Adaptive Document Selection 基于自适应文档选择的迁移学习改进抽象摘要
Pub Date : 2023-01-01 DOI: 10.52731/ijscai.v7.i2.701
Masato Shirai, Kei Wakabayashi
ive document summarization based on neural networks is a promising approach to generate a flexible summary but requires a large amount of training data.While transfer learning can address this issue, there is a potential concern about the negative transfer effect that deteriorates the performance when we use training documents irrelevant to the target domain, which has not been explicitly explored in document summarization tasks.In this paper, we propose a method that selects training documents from the source domain that are expected to be useful for the target summarization.The proposed method is based on the similarity of word distributions between each source document and a set of target documents.We further propose an adaptive approach that builds a custom-made summarization model for each test document by selecting source documents similar to the test document.In the experiment, we confirmed that the negative transfer actually happens also in the document summarization tasks.Additionally, we show that the proposed method effectively avoids the negative transfer issue and improves summarization performance.
基于神经网络的Ive文档摘要是一种很有前途的生成灵活摘要的方法,但需要大量的训练数据。虽然迁移学习可以解决这个问题,但当我们使用与目标领域无关的训练文档时,存在潜在的负面迁移效应,这种效应会降低性能,这在文档摘要任务中尚未得到明确的探讨。在本文中,我们提出了一种从源域中选择对目标摘要有用的训练文档的方法。该方法基于每个源文档和一组目标文档之间单词分布的相似性。我们进一步提出了一种自适应方法,通过选择与测试文档相似的源文档,为每个测试文档构建定制的摘要模型。在实验中,我们证实了负迁移实际上也发生在文档摘要任务中。此外,我们还表明,该方法有效地避免了负迁移问题,提高了摘要性能。
{"title":"Improving Abstractive Summarization by Transfer Learning with Adaptive Document Selection","authors":"Masato Shirai, Kei Wakabayashi","doi":"10.52731/ijscai.v7.i2.701","DOIUrl":"https://doi.org/10.52731/ijscai.v7.i2.701","url":null,"abstract":"ive document summarization based on neural networks is a promising approach to generate a flexible summary but requires a large amount of training data.While transfer learning can address this issue, there is a potential concern about the negative transfer effect that deteriorates the performance when we use training documents irrelevant to the target domain, which has not been explicitly explored in document summarization tasks.In this paper, we propose a method that selects training documents from the source domain that are expected to be useful for the target summarization.The proposed method is based on the similarity of word distributions between each source document and a set of target documents.We further propose an adaptive approach that builds a custom-made summarization model for each test document by selecting source documents similar to the test document.In the experiment, we confirmed that the negative transfer actually happens also in the document summarization tasks.Additionally, we show that the proposed method effectively avoids the negative transfer issue and improves summarization performance.","PeriodicalId":495454,"journal":{"name":"International journal of smart computing and artificial intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135101820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International journal of smart computing and artificial intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1