Handling missing values in clinical machine learning: Insights from an expert study
Preprint, 2025

Inherently interpretable machine learning (IML) models offer valuable support for clinical decision-making but face challenges when features contain missing values. Traditional approaches, such as imputation or discarding incomplete records, are often impractical in scenarios where data is missing at test time. We surveyed 55 clinicians from 29 French trauma centers, collecting 20 complete responses to study their interaction with three IML models in a real-world clinical setting for predicting hemorrhagic shock with missing values. Our findings reveal that while clinicians recognize the value of interpretability and are familiar with common IML approaches, traditional imputation techniques often conflict with their intuition. Instead of imputing unobserved values, they rely on observed features combined with medical intuition and experience. As a result, methods that natively handle missing values are preferred. These findings underscore the need to integrate clinical reasoning into future IML models to enhance human-computer interaction—supporting not only communication with patients but also collaborative decision-making among clinicians.

user-studies

missing values

evaluation with clinicans

human-computer interaction

interpretable machine learning

Författare

Lena Stempfle

Chalmers, Data- och informationsteknik, Data Science och AI

Arthur James

Sorbonne Université

Julie Josse

Institut National de la Sante et de la Recherche Medicale (Inserm)

Tobias Gauss

Université Grenoble Alpes

Fredrik Johansson

Data Science och AI 3

Styrkeområden

Informations- och kommunikationsteknik

Ämneskategorier (SSIF 2025)

Datavetenskap (datalogi)

Fundament

Grundläggande vetenskaper

Infrastruktur

Chalmers e-Commons (inkl. C3SE, 2020-)

Mer information

Senast uppdaterat

2025-09-03