Zhaoxin Su , Gang Huang , Zhou Zhou , Yongfu Li , Sanyuan Zhang , Wei Hua
{"title":"Improving generative trajectory prediction via collision-free modeling and goal scene reconstruction","authors":"Zhaoxin Su , Gang Huang , Zhou Zhou , Yongfu Li , Sanyuan Zhang , Wei Hua","doi":"10.1016/j.patrec.2024.12.004","DOIUrl":null,"url":null,"abstract":"<div><div>In the context of bird’s-eye view traffic scenarios, accurate prediction of future trajectories for various traffic agents (e.g., pedestrians, vehicles) is crucial for driving safety and decision planning. In this paper, we present a generative trajectory prediction framework that incorporates both collision-free modeling and additional reconstruction for future scene context. For the social encoder, we leverage the collision prior by incorporating collision-free constraints (CFC). We construct a social model composed of multiple graphs, where each graph points to the collision prior calculated from uniform direction sampling. In the scene encoder, we employ an attention module to establish connections between trajectory motion and all scene image pixels. Additionally, we reconstruct a goal response map (GRM) aligned with the intended one, thereby enhancing the scene representations. Experiments conducted on nuScenes and ETH/UCY datasets demonstrate the superiority of the proposed framework, achieving a 13.8% reduction in off-road rate on nuScenes and an average 13.2% reduction in collision rate on ETH/UCY datasets.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"188 ","pages":"Pages 117-124"},"PeriodicalIF":3.9000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition Letters","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167865524003593","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
In the context of bird’s-eye view traffic scenarios, accurate prediction of future trajectories for various traffic agents (e.g., pedestrians, vehicles) is crucial for driving safety and decision planning. In this paper, we present a generative trajectory prediction framework that incorporates both collision-free modeling and additional reconstruction for future scene context. For the social encoder, we leverage the collision prior by incorporating collision-free constraints (CFC). We construct a social model composed of multiple graphs, where each graph points to the collision prior calculated from uniform direction sampling. In the scene encoder, we employ an attention module to establish connections between trajectory motion and all scene image pixels. Additionally, we reconstruct a goal response map (GRM) aligned with the intended one, thereby enhancing the scene representations. Experiments conducted on nuScenes and ETH/UCY datasets demonstrate the superiority of the proposed framework, achieving a 13.8% reduction in off-road rate on nuScenes and an average 13.2% reduction in collision rate on ETH/UCY datasets.
期刊介绍:
Pattern Recognition Letters aims at rapid publication of concise articles of a broad interest in pattern recognition.
Subject areas include all the current fields of interest represented by the Technical Committees of the International Association of Pattern Recognition, and other developing themes involving learning and recognition.