{"title":"Reinforcement learning-based control for offshore crane load-landing operations","authors":"Khaled Said Ahmed Maamoun, H. Karimi","doi":"10.20517/ces.2022.28","DOIUrl":null,"url":null,"abstract":"Offshore crane operations are frequently carried out under adverse weather conditions. While offshore cranes attempt to finish the load-landing or lifting operation, the impact between the loads and the vessels is critical, as it can cause serious injuries and extensive damage. Multiple offshore crane operations, including load-landing operations, have used reinforcement learning (RL) to control their activities. In this paper, the Q-learning algorithm is used to develop optimal control sequences for the offshore crane’s actuators to minimize the impact velocity between the crane’s load and the moving vessel. To expand the RL environment, a mathematical model is constructed for the dynamical analysis utilizing the Denavit–Hartenberg (DH) technique and the Lagrange approach. The Double Q-learning algorithm is used to locate the bias that is common in Q-learning algorithms. The average return feature is studied to assess the performance of the Q-learning algorithm. Furthermore, the trained control sequence was tested on a separate sample of episodes, and the hypothesis that, unlike supervised learning, reinforcement learning cannot have a global optimal control sequence but only a local one, was confirmed in this application domain.","PeriodicalId":72652,"journal":{"name":"Complex engineering systems (Alhambra, Calif.)","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex engineering systems (Alhambra, Calif.)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.20517/ces.2022.28","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Offshore crane operations are frequently carried out under adverse weather conditions. While offshore cranes attempt to finish the load-landing or lifting operation, the impact between the loads and the vessels is critical, as it can cause serious injuries and extensive damage. Multiple offshore crane operations, including load-landing operations, have used reinforcement learning (RL) to control their activities. In this paper, the Q-learning algorithm is used to develop optimal control sequences for the offshore crane’s actuators to minimize the impact velocity between the crane’s load and the moving vessel. To expand the RL environment, a mathematical model is constructed for the dynamical analysis utilizing the Denavit–Hartenberg (DH) technique and the Lagrange approach. The Double Q-learning algorithm is used to locate the bias that is common in Q-learning algorithms. The average return feature is studied to assess the performance of the Q-learning algorithm. Furthermore, the trained control sequence was tested on a separate sample of episodes, and the hypothesis that, unlike supervised learning, reinforcement learning cannot have a global optimal control sequence but only a local one, was confirmed in this application domain.