Neural Rendering for Autonomous Driving
Doctoral thesis, 2026
The first contribution is NeuRAD, a neural simulator that jointly reconstructs camera and lidar data from recorded driving sequences. By explicitly modeling key sensor characteristics, including rolling shutter effects, beam divergence, and non-returning lidar rays, NeuRAD achieves state-of-the-art rendering fidelity across multiple autonomous driving datasets.
The second contribution is SplatAD, which replaces the volumetric representation of NeuRAD with 3D Gaussian primitives and purpose-built rasterization algorithms. This yields real-time rendering of both camera and lidar data at competitive fidelity, while reducing training and inference times by an order of magnitude, making large-scale simulation substantially more practical.
The third contribution is IDSplat, which eliminates the reliance on human-annotated 3D bounding boxes required by both NeuRAD and SplatAD. By combining vision foundation models with classical matching and estimation techniques, IDSplat achieves annotation-free dynamic scene reconstruction that generalizes zero-shot to new datasets.
The fourth contribution is a highly parallelizable GPU algorithm for constructing large-scale 3D Voronoi and power diagrams, addressing a key computational bottleneck in mesh-based neural rendering and enabling larger, higher-fidelity scene representations.
Finally, the thesis investigates the real-to-sim gap: the discrepancy between how autonomous systems perceive real and rendered sensor data. Through large-scale evaluation, correlations between rendering quality and downstream perception performance are identified, and fine-tuning strategies are developed that improve model robustness to simulation artifacts without compromising real-world performance.
Together, these contributions advance the state of neural simulation for autonomous driving, bringing scalable, high-fidelity virtual testing closer to practical deployment.
Sensor simulation
Gaussian splatting
Autonomous driving
Neural rendering
Neural radiance fields
Novel view synthesis
Author
Carl Lindström
Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering
NeuRAD: Neural Rendering for Autonomous Driving
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,;(2024)p. 14895-14904
Paper in proceeding
Are NeRFs ready for autonomous driving? Towards closing the real-to-simulation gap
IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops,;(2024)p. 4461-4471
Paper in proceeding
SplatAD: Real-Time Lidar and Camera Rendering with 3D Gaussian Splatting for Autonomous Driving
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,;(2025)
Paper in proceeding
Lindström, C., Rafidashti, M., Fatemi, M., Hammarstrand, L., Oswald, M. R., Svensson, L. (2026). IDSplat: Instance-Decomposed 3D Gaussian Splatting for Driving Scenes.
Taveira, B., Lindström, C., Fatemi, M., Hammarstrand, L., Kahl, F. (2026). Scalable GPU Construction of 3D Voronoi and Power Diagrams.
This thesis takes a different approach: learning virtual environments from data recorded by real vehicles on real roads. Camera images and laser scans from an ordinary drive are used to reconstruct an interactive virtual replica of the scene, in which "what if" questions can be explored and realistic sensor observations generated for hypothetical scenarios.
We address three challenges. First, we develop methods that reconstruct scenes for both cameras and lidars, carefully modeling the characteristics of each sensor. Second, we make reconstruction fast enough to scale to thousands of scenes, and remove the need for manual annotations. Third, we study how well conclusions drawn from simulation transfer to the real world and develop strategies to close that gap. Together, these contributions bring us closer to a future where self-driving systems can be tested exhaustively in virtual environments built from real-world data.
Self-supervised learning for multimodal perception systems
Wallenberg AI, Autonomous Systems and Software Program, 2023-08-01 -- 2027-08-01.
Areas of Advance
Information and Communication Technology
Transport
Subject Categories (SSIF 2025)
Computer Vision and learning System
Computer graphics and computer vision
Artificial Intelligence
DOI
10.63959/chalmers.dt/5865
ISBN
978-91-8103-408-0
Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 5865
Publisher
Chalmers
EE, Hörsalsvägen 11
Opponent: Associate Professor Arno Solin, Department of Computer Science, Aalto University, Finland