Published January 20, 2025 | Version v1
Journal article Open

Privacy-Preserving Hyperparameter Tuning for Federated Learning

  • 1. Kera Health Platforms Inc, Dakar, Senegal
  • 2. ROR icon RISE Research Institutes of Sweden
  • 3. Bilkent University: Ankara, TR

Description

In this paper, we study the open problem of privacy-preserving hyperparameter (HP) tuning for cross-silo federated learning (FL). We first perform a comprehensive measurement study and benchmark various single-shot HP tuning strategies compatible with privacy-preserving FL pipelines. Our experimental results show that the optimal parameters of the FL server, e.g., the learning rate, can be accurately and efficiently tuned based on the HPs found by each client on its local data. We demonstrate that HP averaging is suitable for iid settings, while density-based clustering can uncover the optimal set of parameters in non-iid ones. Then, to prevent information leakage from the exchange of the clients' local HPs, we design and implement PrivTuna, a novel framework for privacy-preserving HP tuning using multiparty homomorphic encryption. We use PrivTuna to implement privacy-preserving federated averaging and density-based clustering, and we experimentally evaluate its performance demonstrating its computation/communication efficiency and its precision in tuning hyperparameters.

Files

Privacy-Preserving_Hyperparameter_Tuning_for_Federated_Learning.pdf

Files (1.5 MB)