Asymptotic Analysis of Machine Learning Models: Comparison Theorems and Universality
Licentiate thesis, 2023
The papers in this thesis particularly focus on the usage of Gaussian comparison theorems as a methodological tool for the analysis of these problems. In particular, the convex Gaussian min max theorem allows us to study more complex ML optimization problems, by considering alternative models that are simpler to analyze, but asymptotically hold similar properties.
Secondarily, this thesis considers universality, which within the asymptotic context demonstrates that many statistics of ML models are fully determined by lower order statistical moments. This allows us to study surrogate Gaussian models, matching these moments. These surrogate Gaussian models can subsequently be analyzed by means of the Gaussian comparison theorems.
CGMT
Machine Learning
convex gaussian min max theorem
Asymptotic
universality
Author
David Bosch
Chalmers, Computer Science and Engineering (Chalmers), Data Science and AI
Double Descent in Feature Selection: Revisiting LASSO and Basis Pursuit
Thirty-eighth International Conference on Machine Learning, ICML 2021,;(2021)
Paper in proceeding
Random Features Model with General Convex Regularization: A Fine Grained Analysis with Precise Asymptotic Learning Curves
Proceedings of Machine Learning Research,;Vol. 206(2023)p. 11371-11414
Paper in proceeding
Subject Categories
Computer and Information Science
Probability Theory and Statistics
Publisher
Chalmers
CSE EDIT 3128
Opponent: Samet Oymak, Assistant Professor Electrical and Computer Engineering, UC Riverside, USA