Neural Rendering for Autonomous Driving
Doctoral thesis, 2026

Autonomous driving holds the potential to fundamentally transform transportation, but realizing that potential requires rigorous testing across a vast and diverse range of scenarios. Exhaustive real-world testing is neither safe for exploring edge cases nor practical at the scale required for adequate coverage, making high-fidelity simulation an essential component of the development and validation pipeline. This thesis addresses the challenge of building digital twins for autonomous driving: data-driven virtual replicas of real-world scenes that faithfully reproduce sensor observations while remaining editable for counterfactual testing.
The first contribution is NeuRAD, a neural simulator that jointly reconstructs camera and lidar data from recorded driving sequences. By explicitly modeling key sensor characteristics, including rolling shutter effects, beam divergence, and non-returning lidar rays, NeuRAD achieves state-of-the-art rendering fidelity across multiple autonomous driving datasets.
The second contribution is SplatAD, which replaces the volumetric representation of NeuRAD with 3D Gaussian primitives and purpose-built rasterization algorithms. This yields real-time rendering of both camera and lidar data at competitive fidelity, while reducing training and inference times by an order of magnitude, making large-scale simulation substantially more practical.
The third contribution is IDSplat, which eliminates the reliance on human-annotated 3D bounding boxes required by both NeuRAD and SplatAD. By combining vision foundation models with classical matching and estimation techniques, IDSplat achieves annotation-free dynamic scene reconstruction that generalizes zero-shot to new datasets.
The fourth contribution is a highly parallelizable GPU algorithm for constructing large-scale 3D Voronoi and power diagrams, addressing a key computational bottleneck in mesh-based neural rendering and enabling larger, higher-fidelity scene representations.
Finally, the thesis investigates the real-to-sim gap: the discrepancy between how autonomous systems perceive real and rendered sensor data. Through large-scale evaluation, correlations between rendering quality and downstream perception performance are identified, and fine-tuning strategies are developed that improve model robustness to simulation artifacts without compromising real-world performance.
Together, these contributions advance the state of neural simulation for autonomous driving, bringing scalable, high-fidelity virtual testing closer to practical deployment.

Sensor simulation

Gaussian splatting

Autonomous driving

Neural rendering

Neural radiance fields

Novel view synthesis

EE, Hörsalsvägen 11
Opponent: Associate Professor Arno Solin, Department of Computer Science, Aalto University, Finland

Author

Carl Lindström

Chalmers, Electrical Engineering, Signal Processing and Biomedical Engineering

NeuRAD: Neural Rendering for Autonomous Driving

Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,;(2024)p. 14895-14904

Paper in proceeding

Are NeRFs ready for autonomous driving? Towards closing the real-to-simulation gap

IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops,;(2024)p. 4461-4471

Paper in proceeding

SplatAD: Real-Time Lidar and Camera Rendering with 3D Gaussian Splatting for Autonomous Driving

Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,;(2025)

Paper in proceeding

Lindström, C., Rafidashti, M., Fatemi, M., Hammarstrand, L., Oswald, M. R., Svensson, L. (2026). IDSplat: Instance-Decomposed 3D Gaussian Splatting for Driving Scenes.

Taveira, B., Lindström, C., Fatemi, M., Hammarstrand, L., Kahl, F. (2026). Scalable GPU Construction of 3D Voronoi and Power Diagrams.

Self-driving cars have the potential to save lives and transform how we move through cities. But before a vehicle can navigate public roads on its own, it must be tested across a wide range of situations. The fundamental problem is coverage: safety-critical events may occur only once in millions of kilometers of driving, yet the system must handle them correctly every time. Simulation offers a solution, but traditional simulators rely on virtual environments built by hand, which is expensive and struggles to capture real-world complexity.

This thesis takes a different approach: learning virtual environments from data recorded by real vehicles on real roads. Camera images and laser scans from an ordinary drive are used to reconstruct an interactive virtual replica of the scene, in which "what if" questions can be explored and realistic sensor observations generated for hypothetical scenarios.

We address three challenges. First, we develop methods that reconstruct scenes for both cameras and lidars, carefully modeling the characteristics of each sensor. Second, we make reconstruction fast enough to scale to thousands of scenes, and remove the need for manual annotations. Third, we study how well conclusions drawn from simulation transfer to the real world and develop strategies to close that gap. Together, these contributions bring us closer to a future where self-driving systems can be tested exhaustively in virtual environments built from real-world data.

Self-supervised learning for multimodal perception systems

Wallenberg AI, Autonomous Systems and Software Program, 2023-08-01 -- 2027-08-01.

Areas of Advance

Information and Communication Technology

Transport

Subject Categories (SSIF 2025)

Computer Vision and learning System

Computer graphics and computer vision

Artificial Intelligence

DOI

10.63959/chalmers.dt/5865

ISBN

978-91-8103-408-0

Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 5865

Publisher

Chalmers

EE, Hörsalsvägen 11

Online

Opponent: Associate Professor Arno Solin, Department of Computer Science, Aalto University, Finland

More information

Latest update

5/11/2026