Interpretable Machine Learning for Prediction with Missing Values at Test Time
Doktorsavhandling, 2025
We first leverage missingness patterns by introducing Sharing Pattern Submodels, where a separate interpretable submodel is trained for each unique missingness pattern, with parameters shared across submodels via sparsity to enhance generalization. Next, we investigate training models that rarely require the values of missing (or imputed) features at test time. We introduce MINTY, a linear rule-based model that avoids imputation by allowing logical substitutions for missing features. We then generalize this idea through a missingness-avoiding framework, which extends to multiple model classes, including decision trees, sparse linear models, and ensembles, by incorporating classifier-specific regularization terms into their learning objectives to discourage reliance on missing values. To support the development of clinically valuable models, we conducted a clinician survey revealing that medical professionals favor models that natively handle missingness. Finally, we explore interpretable patient history representations for modeling policies in sequential clinical decision-making, shifting the focus from missingness to temporal modeling. Collectively, this work establishes methods for interpretable machine learning with test-time missingness, supported by both technical innovations and human-centered insights, to enable transparent and practical decision support.
Missing Values
Decision Making
Healthcare
Machine learning
Interpretability
Författare
Lena Stempfle
Data Science och AI 3
Prediction Models That Learn to Avoid Missing Values
Proceedings of Machine Learning Research,;Vol. 267(2025)
Paper i proceeding
Sharing Pattern Submodels for Prediction with Missing Values
Proceedings of the AAAI Conference on Artificial Intelligence,;Vol. 37(2023)p. 9882-9890
Paper i proceeding
MINTY: Rule-based Models that Minimize the Need for Imputing Features with Missing Values
Proceedings of the 38th Conference on Uncertainty in Artificial Intelligence, UAI 2022,;Vol. 238(2024)p. 964-972
Paper i proceeding
How Should We Represent History in Interpretable Models of Clinical Policies?
Proceedings of Machine Learning Research,;Vol. 259(2024)p. 714-734
Paper i proceeding
In safety-critical fields like healthcare, clinicians often must make decisions based on incomplete records–lab results may be delayed, or certain tests never ordered. A common strategy to deal with missing data is imputation, where gaps are filled using averages, zeros, or statistical estimates. However, this can introduce bias and make predictions harder to interpret and trust.
This thesis explores new approaches to building models that work reliably even when some data is missing, without relying on imputation. The goal is to maintain or improve predictive performance while keeping the models interpretable and aligned with how clinicians reason. One method trains separate models for different patterns of missing data, for example, grouping patients with similar available measurements, while still sharing useful information across groups to generalize well. Another key contribution is a rule-based method that uses logical substitutions instead of imputing missing values. This idea is extended into a broader framework that helps models, like decision trees, sparse linear models, and ensembles, learn to avoid relying on imputed data altogether. This can help clinicians make better decisions using fewer additional measurements by leveraging frequently observed features, ultimately reducing delays and healthcare costs. These models, tested on diverse healthcare datasets, demonstrate strong predictive performance while reducing reliance on filled-in data. A clinician survey confirms a preference for models that explicitly recognize and handle missingness. The thesis also investigates interpretable models for time-series data, common in patient records, to support evolving treatment decisions over time.
Together, these contributions offer practical tools for building decision-making systems that are both interpretable and trustworthy–especially in clinical environments where missing data is the norm, not the exception.
Styrkeområden
Informations- och kommunikationsteknik
Ämneskategorier (SSIF 2025)
Sannolikhetsteori och statistik
Datavetenskap (datalogi)
Datorsystem
Fundament
Grundläggande vetenskaper
Infrastruktur
Chalmers e-Commons (inkl. C3SE, 2020-)
ISBN
978-91-8103-246-8
Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 5704
Utgivare
Chalmers
Room EE (6233), EDIT Building Hörsalsvägen 11, Chalmers University of Technology, Campus Johanneberg
Opponent: Prof. Stefan Feuerriegel, Institute of Artificial Intelligence (AI) in Management, LMU Munich School of Management, Germany