Back to Search Start Over

Sparsity-Preserving Differentially Private Training of Large Embedding Models

Authors :
Ghazi, Badih
Huang, Yangsibo
Kamath, Pritish
Kumar, Ravi
Manurangsi, Pasin
Sinha, Amer
Zhang, Chiyuan
Ghazi, Badih
Huang, Yangsibo
Kamath, Pritish
Kumar, Ravi
Manurangsi, Pasin
Sinha, Amer
Zhang, Chiyuan
Publication Year :
2023

Abstract

As the use of large embedding models in recommendation systems and language applications increases, concerns over user data privacy have also risen. DP-SGD, a training algorithm that combines differential privacy with stochastic gradient descent, has been the workhorse in protecting user privacy without compromising model accuracy by much. However, applying DP-SGD naively to embedding models can destroy gradient sparsity, leading to reduced training efficiency. To address this issue, we present two new algorithms, DP-FEST and DP-AdaFEST, that preserve gradient sparsity during private training of large embedding models. Our algorithms achieve substantial reductions ($10^6 \times$) in gradient size, while maintaining comparable levels of accuracy, on benchmark real-world datasets.<br />Comment: Neural Information Processing Systems (NeurIPS) 2023

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1438499083
Document Type :
Electronic Resource