we advance the understanding of deep neural networks through the investigation of stochastic continuous-depth neural networks. These can be thought of as deep neural networks (DNN) composed of infinitely many stochastic layers, where each single layer only brings about a gradual change to the output of the preceding layers. We will analyse such stochastic continuous-depth neural networks using tools from stochastic calculus and Bayesian statistics. From that, we will derive practically relevant and novel training algorithms for stochastic DNNs with the aim to capture the uncertainty associated with the predictions of the network. This project is supported by Chalmers AI Research Centre (CHAIR) including a PhD position.
vid Chalmers, Mathematical Sciences, Applied Mathematics and Statistics
Docent vid Chalmers, Mathematical Sciences, Applied Mathematics and Statistics
Funding Chalmers participation during 2020–