Instance-Level Safety-Aware Fidelity of Synthetic Data and its Calibration
Paper in proceeding, 2024

Modeling and calibrating the fidelity of synthetic data is paramount in shaping the future of safe and reliable self-driving technology by offering a cost-effective and scalable alternative to real-world data collection. We focus on its role in safety-critical applications, introducing four types of instance-level fidelity that go beyond mere visual input characteristics. The aim is to ensure that applying testing on synthetic data can reveal real-world safety issues, and the absence of safety-critical issues when testing under synthetic data can provide a strong safety guarantee in real-world behavior. We suggest an optimization method to refine the synthetic data generator, reducing fidelity gaps identified by deep learning components. Experiments show this tuning enhances the correlation between safety-critical errors in synthetic and real data.

Author

Chih-Hong Cheng

University of Gothenburg

Software Engineering 2

Paul Stockel

University of Hildesheim

Xingyu Zhao

The University of Warwick

IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC

21530009 (ISSN) 21530017 (eISSN)

2354-2361
9798331505929 (ISBN)

27th IEEE International Conference on Intelligent Transportation Systems, ITSC 2024
Edmonton, Canada,

Subject Categories (SSIF 2025)

Computer Sciences

Computer Systems

DOI

10.1109/ITSC58415.2024.10920032

More information

Latest update

4/14/2025