Fed-AugMix: Balancing Privacy and Utility via Data Augmentation

Збережено в:
Бібліографічні деталі
Опубліковано в::arXiv.org (Dec 18, 2024), p. n/a
Автор: Li, Haoyang
Інші автори: Chen, Wei, Zhang, Xiaojin
Опубліковано:
Cornell University Library, arXiv.org
Предмети:
Онлайн доступ:Citation/Abstract
Full text outside of ProQuest
Теги: Додати тег
Немає тегів, Будьте першим, хто поставить тег для цього запису!
Опис
Короткий огляд:Gradient leakage attacks pose a significant threat to the privacy guarantees of federated learning. While distortion-based protection mechanisms are commonly employed to mitigate this issue, they often lead to notable performance degradation. Existing methods struggle to preserve model performance while ensuring privacy. To address this challenge, we propose a novel data augmentation-based framework designed to achieve a favorable privacy-utility trade-off, with the potential to enhance model performance in certain cases. Our framework incorporates the AugMix algorithm at the client level, enabling data augmentation with controllable severity. By integrating the Jensen-Shannon divergence into the loss function, we embed the distortion introduced by AugMix into the model gradients, effectively safeguarding privacy against deep leakage attacks. Moreover, the JS divergence promotes model consistency across different augmentations of the same image, enhancing both robustness and performance. Extensive experiments on benchmark datasets demonstrate the effectiveness and stability of our method in protecting privacy. Furthermore, our approach maintains, and in some cases improves, model performance, showcasing its ability to achieve a robust privacy-utility trade-off.
ISSN:2331-8422
Джерело:Engineering Database