{"title":"改进学习环境中风险感知控制器的分布鲁棒性","authors":"A. Hakobyan, Insoon Yang","doi":"10.1109/CDC45484.2021.9682981","DOIUrl":null,"url":null,"abstract":"This paper is concerned with designing a risk-aware controller in an unknown and dynamic environment. In our method, the evolution of the environment state is learned using observational data via Gaussian process regression (GPR). Unfortunately, these learning results provide imperfect distribution information about the environment. To address such distribution errors, we propose a risk-constrained model predictive control (MPC) method that exploits techniques from modern distributionally robust optimization (DRO). To resolve the infinite dimensionality issue inherent in DRO, we derive a tractable semidefinite programming (SDP) problem that upper-bounds the original MPC problem. Furthermore, the SDP problem is reduced to a quadratic program when the constraint function has a decomposable form. The performance and the utility of our method are demonstrated through an autonomous driving problem, and the results show that our controller preserves safety despite errors in learning the behaviors of surrounding vehicles.","PeriodicalId":229089,"journal":{"name":"2021 60th IEEE Conference on Decision and Control (CDC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Toward Improving the Distributional Robustness of Risk-Aware Controllers in Learning-Enabled Environments\",\"authors\":\"A. Hakobyan, Insoon Yang\",\"doi\":\"10.1109/CDC45484.2021.9682981\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper is concerned with designing a risk-aware controller in an unknown and dynamic environment. In our method, the evolution of the environment state is learned using observational data via Gaussian process regression (GPR). Unfortunately, these learning results provide imperfect distribution information about the environment. To address such distribution errors, we propose a risk-constrained model predictive control (MPC) method that exploits techniques from modern distributionally robust optimization (DRO). To resolve the infinite dimensionality issue inherent in DRO, we derive a tractable semidefinite programming (SDP) problem that upper-bounds the original MPC problem. Furthermore, the SDP problem is reduced to a quadratic program when the constraint function has a decomposable form. The performance and the utility of our method are demonstrated through an autonomous driving problem, and the results show that our controller preserves safety despite errors in learning the behaviors of surrounding vehicles.\",\"PeriodicalId\":229089,\"journal\":{\"name\":\"2021 60th IEEE Conference on Decision and Control (CDC)\",\"volume\":\"31 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 60th IEEE Conference on Decision and Control (CDC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CDC45484.2021.9682981\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 60th IEEE Conference on Decision and Control (CDC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CDC45484.2021.9682981","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Toward Improving the Distributional Robustness of Risk-Aware Controllers in Learning-Enabled Environments
This paper is concerned with designing a risk-aware controller in an unknown and dynamic environment. In our method, the evolution of the environment state is learned using observational data via Gaussian process regression (GPR). Unfortunately, these learning results provide imperfect distribution information about the environment. To address such distribution errors, we propose a risk-constrained model predictive control (MPC) method that exploits techniques from modern distributionally robust optimization (DRO). To resolve the infinite dimensionality issue inherent in DRO, we derive a tractable semidefinite programming (SDP) problem that upper-bounds the original MPC problem. Furthermore, the SDP problem is reduced to a quadratic program when the constraint function has a decomposable form. The performance and the utility of our method are demonstrated through an autonomous driving problem, and the results show that our controller preserves safety despite errors in learning the behaviors of surrounding vehicles.