Training Effective Neural Networks on Structured Data with Federated Learning
Description
Federated Learning decreases privacy risks when training Machine Learning (ML) models on distributed data, as it removes the need for sharing and centralizing sensitive data. However, this learning paradigm can also influence the effectiveness of the obtained prediction models. In this paper, we specifically study Neural Networks, as a powerful and popular ML model, and contrast the impact of Federated Learning on the effectiveness compared to a centralized approach - when data is aggregated at one place before processing - to assess to what extent Federated Learning is suited as a replacement. We also analyze the effect of non-independent and identically distributed (non-iid) data on effectiveness and convergence speed (efficiency) of Federated Learning. Based on this, we show in which scenarios (depending on the dataset, the number of nodes in the setting and data distribution) Federated Learning can be successfully employed.
Files
adult.csv
Files
(5.2 GB)
Name | Size | Download all |
---|---|---|
md5:75048147e2e36ef4ff1d8bfde6636a2f
|
3.3 MB | Preview Download |
md5:52b89051b9bd37a91a54e8570b963719
|
19.9 kB | Download |
md5:05f98bb3ed813e7cdc408b6575a72dc2
|
6.2 MB | Preview Download |
md5:88c8ecf6a22fe0034e2cfc61613ac860
|
1.0 GB | Preview Download |
md5:4951cfd3c3e18a8ffb3ff71e819a4361
|
1.0 GB | Preview Download |
md5:428511f444ac80258ad78a0a7a1f48fd
|
1.0 GB | Preview Download |
md5:7e414cb425a1f3e6ada946fd2e502c9e
|
1.0 GB | Preview Download |
md5:b328ec4573311856486f32eace1cec24
|
1.0 GB | Preview Download |