Driving Scene Synthesis on Free-form Trajectories with Generative Prior


Zeyu Yang1*, Zijie Pan1*, Yuankun Yang1*, Xiatian Zhu2, Li Zhang1
1 Fudan University, 2 University of Surrey
*Equally contributed.

Paper Code



Abstract



Driving scene synthesis along free-form trajectories is essential for driving simulations to enable closed-loop evaluation of end-to-end driving policies. While existing methods excel at novel view synthesis on recorded trajectories, they face challenges with novel trajectories due to limited views of driving videos and the vastness of driving environments. To tackle this challenge, we propose a novel free-form driving view synthesis approach, dubbed DriveX, by leveraging video generative prior to optimize a 3D model across a variety of trajectories. Concretely, we crafted an inverse problem that enables a video diffusion model to be utilized as a prior for many-trajectory optimization of a parametric 3D model (e.g., Gaussian splatting). To seamlessly use the generative prior, we iteratively conduct this process during optimization. Our resulting model can produce high-fidelity virtual driving environments outside the recorded trajectory, enabling free-form trajectory driving simulation. Beyond real driving scenes, DriveX can also be utilized to simulate virtual driving worlds from AI-generated videos.



Method Overview









Driving scene synthesis on novel trajectory


StreetGaussian
DriveX (Ours)


Comparisons on Waymo along novel trajectory


EmerNeRF
PVG
StreetGaussian
DriveX (Ours)


Comparison with DriveDreamer4D


Recorded trajectory
DriveDreamer4D
DriveX (Ours)


Driving scene synthesis along novel trajectory from AI-generated videos


Original trajectory
Novel traj. 1
Novel traj. 2




Acknowledgements: The website template was borrowed from Lior Yariv. Image sliders are based on dics.