The objective of this paper is to deconstruct driving behaviors in interactions with pedestrians at uncontrolled crosswalks. Trajectory data are used to extract variables describing driver–pedestrian interactions, including position, acceleration, velocity, yaw rate, and interaction risk. Driving behavior is modeled as utility-driven, intelligent, and rational decision-making within the framework of a finite-state Markov decision process (MDP). The vanilla generative adversarial imitation learning (GAIL) framework is improved to reconstruct a human-like driving behavior model where the utility function is defined as the deviation between the agent’s behavior distribution and that of human drivers. Maximizing this utility through a deep reinforcement learning (RL) approach drives agents to progressively clone the behavioral policies of human drivers in the real world. The behavioral policy is formulated as a pre-trained driving behavior model and validated on a simulation platform for its ability in reproducing human driving behavior. Experimental results show that the model successfully reproduces the rationality of human drivers and generates human-like interaction trajectories in the simulation environment. Transfer experiments further demonstrate the generalizability of the pre-tained behavioral model. The interaction policy map and the state-value map are visualized to elucidate the generative mechanisms underlying human-like trajectories by revealing risk- and context-dependent layered patterns and latent behavioral preferences. This work contributes to the advancement of human-like behavioral models, thereby enhancing the fidelity of traffic microsimulation and improving behavior modeling in complex driver–pedestrian interactions.
扫码关注我们
求助内容:
应助结果提醒方式:
