{"title":"面向驾驶员活动分类的动作与目标交互识别","authors":"Patrick Weyers, David Schiebener, A. Kummert","doi":"10.1109/ITSC.2019.8917139","DOIUrl":null,"url":null,"abstract":"Knowing what the driver is doing inside a vehicle is essential information for all stages of vehicle automation. For example it can be used for adaptive warning strategies in combination with an advanced driver assistance systems system, for predicting the response time to take back the control of a partially automated vehicle, or ensuring the driver is ready to manually drive a highly automated vehicle in the future. We present a system for driver activity recognition based on image sequences of an in-cabin time-of-flight camera. Our dataset includes actions such as entering and leaving a car or driver object interactions such as using a phone or drinking. In the first stage, we localize body key points of the driver. In the second stage, we extract image regions around the localized hands. These regions and the determined 3D body key points are used as the input to a recurrent neural network for driver activity recognition. With a mean average precision of 0.85 we reach better classification rates than approaches relying only on body key points or images.","PeriodicalId":6717,"journal":{"name":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","volume":"56 1","pages":"4336-4341"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":"{\"title\":\"Action and Object Interaction Recognition for Driver Activity Classification\",\"authors\":\"Patrick Weyers, David Schiebener, A. Kummert\",\"doi\":\"10.1109/ITSC.2019.8917139\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Knowing what the driver is doing inside a vehicle is essential information for all stages of vehicle automation. For example it can be used for adaptive warning strategies in combination with an advanced driver assistance systems system, for predicting the response time to take back the control of a partially automated vehicle, or ensuring the driver is ready to manually drive a highly automated vehicle in the future. We present a system for driver activity recognition based on image sequences of an in-cabin time-of-flight camera. Our dataset includes actions such as entering and leaving a car or driver object interactions such as using a phone or drinking. In the first stage, we localize body key points of the driver. In the second stage, we extract image regions around the localized hands. These regions and the determined 3D body key points are used as the input to a recurrent neural network for driver activity recognition. With a mean average precision of 0.85 we reach better classification rates than approaches relying only on body key points or images.\",\"PeriodicalId\":6717,\"journal\":{\"name\":\"2019 IEEE Intelligent Transportation Systems Conference (ITSC)\",\"volume\":\"56 1\",\"pages\":\"4336-4341\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"15\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Intelligent Transportation Systems Conference (ITSC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITSC.2019.8917139\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITSC.2019.8917139","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Action and Object Interaction Recognition for Driver Activity Classification
Knowing what the driver is doing inside a vehicle is essential information for all stages of vehicle automation. For example it can be used for adaptive warning strategies in combination with an advanced driver assistance systems system, for predicting the response time to take back the control of a partially automated vehicle, or ensuring the driver is ready to manually drive a highly automated vehicle in the future. We present a system for driver activity recognition based on image sequences of an in-cabin time-of-flight camera. Our dataset includes actions such as entering and leaving a car or driver object interactions such as using a phone or drinking. In the first stage, we localize body key points of the driver. In the second stage, we extract image regions around the localized hands. These regions and the determined 3D body key points are used as the input to a recurrent neural network for driver activity recognition. With a mean average precision of 0.85 we reach better classification rates than approaches relying only on body key points or images.