{"title":"Human Cognitive Learning in Shared Control via Differential Game With Bounded Rationality and Incomplete Information","authors":"Huai-Ning Wu;Xiao-Yan Jiang;Mi Wang","doi":"10.1109/TAI.2024.3415549","DOIUrl":null,"url":null,"abstract":"Since human beings are of limited reasoning ability as well as the machines do not usually know human intentions, how to learn human cognitive levels in shared control to enhance the machines’ intelligence is a challenging issue. In this study, this issue is addressed in the context of human–machine shared control for a class of human-in-the-loop (HiTL) systems based on a differential game with bounded rationality and incomplete information. Initially, we formulate the human–machine shared control problem as a two-player nonzero-sum linear quadratic dynamic game (LQDG), where the weighting matrix of the cost function representing the human intention is unknown for the machine. To model the human bounded rationality, the level-\n<inline-formula><tex-math>$\\boldsymbol{k}$</tex-math></inline-formula>\n (LK) approach is employed to set up the LK control policies of two players performing the corresponding steps of strategic thinking. To infer the human intention, an online adaptive inverse optimal control (IOC) algorithm is then developed by using the system state data, so that the control policies of different cognitive levels can be computed. In addition, a reinforcement learning method is proposed for the machine to identify the distribution of the human cognitive levels while providing a proactive collaborative control to assist the human in a probabilistic switching way. Finally, simulation results on a cooperative shared control driver assistance system (DAS) illustrate the efficacy of the proposed approach.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 10","pages":"5141-5152"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10562193/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Since human beings are of limited reasoning ability as well as the machines do not usually know human intentions, how to learn human cognitive levels in shared control to enhance the machines’ intelligence is a challenging issue. In this study, this issue is addressed in the context of human–machine shared control for a class of human-in-the-loop (HiTL) systems based on a differential game with bounded rationality and incomplete information. Initially, we formulate the human–machine shared control problem as a two-player nonzero-sum linear quadratic dynamic game (LQDG), where the weighting matrix of the cost function representing the human intention is unknown for the machine. To model the human bounded rationality, the level-
$\boldsymbol{k}$
(LK) approach is employed to set up the LK control policies of two players performing the corresponding steps of strategic thinking. To infer the human intention, an online adaptive inverse optimal control (IOC) algorithm is then developed by using the system state data, so that the control policies of different cognitive levels can be computed. In addition, a reinforcement learning method is proposed for the machine to identify the distribution of the human cognitive levels while providing a proactive collaborative control to assist the human in a probabilistic switching way. Finally, simulation results on a cooperative shared control driver assistance system (DAS) illustrate the efficacy of the proposed approach.