Federated Learning of Deep Neural Decision Forests
Paper in proceeding, 2019

Modern technical products have access to a huge amount of data and by utilizing machine learning algorithms this data can be used to improve usability and performance of the products. However, the data is likely to be large in quantity and privacy sensitive, which excludes the possibility of sending and storing all the data centrally. This in turn makes it difficult to train global machine learning models on the combined data of different devices. A decentralized approach known as federated learning solves this problem by letting devices, or clients, update a global model using their own data and only sending changes of the global model, which means that they do not need to communicate privacy sensitive data.

Deep neural decision forests (DNDF), inspired by the versatile algorithm random forests, combine the divide-and-conquer principle together with the property representation learning. In this paper we further develop the concept of DNDF to be more suited for the framework of federated learning. By parameterizing the probability distributions in the prediction nodes of the forest, and include all trees of the forest in the loss function, a gradient of the whole forest can be computed which some/several federated learning algorithms utilize. We demonstrate the inclusion of DNDF in federated learning by an empirical experiment with both homogeneous and heterogeneous data and baseline it against a convolutional neural network with the same architecture as the DNDF. Experimental results show that the modified DNDF, consisting of three to five decision trees, outperform the baseline convolutional neural network.

Parameterization

Federated learning

Deep neural decision forests

Author

Anders Sjöberg

Fraunhofer-Chalmers Centre

Emil Gustavsson

Fraunhofer-Chalmers Centre

Ashok Krishna Chaitanya Koppisetty

Volvo Cars

Mats Jirstrand

Fraunhofer-Chalmers Centre

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

03029743 (ISSN) 16113349 (eISSN)

Vol. 11943 700-710
978-3-030-37599-7 (ISBN)

The Fifth International Conference on Machine Learning, Optimization, and Data Science (LOD))
Siena, Italy,

Subject Categories

Probability Theory and Statistics

Computer Science

DOI

10.1007/978-3-030-37599-7_58

More information

Latest update

8/29/2024