Publications
WORKSHOP (INTERNATIONAL) Measuring Lower Bound of Local Differential Privacy via Adversary Instantiation in Federated Learning
Marin Matsumoto (Ochanomizu University), Tsubasa Takahashi, Seng Pei Liew, Masato Oguchi (Ochanomizu University)
Theory and Practice of Differential Privacy (TPDP 2022)
July 22, 2022
Local differential privacy (LDP) gives a strong privacy guarantee to be used in a distributed setting like federated learning (FL). LDP mechanisms in FL protect a client's gradient by randomizing it on the client; however, how can we interpret the privacy level given by the randomization? Moreover, what types of attacks can we mitigate in practice? To answer these questions, we introduce an empirical privacy test by measuring the lower bounds of LDP. The privacy test estimates how an adversary predicts if a reported randomized gradient was crafted from a raw gradient g1 or g2. We then instantiate six adversaries in FL under LDP to measure empirical LDP at various attack surfaces, including a worst-case attack that reaches the theoretical upper bound of LDP. The empirical privacy test with the adversary instantiations enables us to interpret LDP more intuitively and discuss relaxation of the privacy parameter until a particular instantiated attack surfaces. We also demonstrate numerical observations of the measured privacy in these adversarial settings, and the worst-case attack is not realistic in FL. In the end, we also discuss the possible relaxation of privacy levels in FL under LDP.
Paper : Measuring Lower Bound of Local Differential Privacy via Adversary Instantiation in Federated Learning (external link)