MemDPT: Differential Privacy for Memory Efficient Language Models

Y Liu, X Peng, J Cao, Y Zhang, C Ma, S Deng… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models have consistently demonstrated remarkable performance across a
wide spectrum of applications. Nonetheless, the deployment of these models can …

Mitigating Noise Detriment in Differentially Private Federated Learning with Model Pre-training

H Jin, Y Zhou, L Cui, QZ Sheng - arXiv preprint arXiv:2408.09478, 2024 - arxiv.org
Pre-training exploits public datasets to pre-train an advanced machine learning model, so
that the model can be easily tuned to adapt to various downstream tasks. Pre-training has …

On the benefits of public representations for private transfer learning under distribution shift

P Thaker, A Setlur, S Wu, V Smith - The Thirty-eighth Annual …, 2023 - openreview.net
Public pretraining is a promising approach to improve differentially private model training.
However, recent work has noted that many positive research results studying this paradigm …

Strategies for Learning From Non-Ideal Sources of Data

C Hou - 2024 - search.proquest.com
Abstract Machine learning (ML) generally performs well when there is unfettered access to
large quantities of clean, relevant data. However, harnessing the benefits of large, clean …

[PDF][PDF] DP-MemArc: Differential Privacy Transfer Learning for Memory Efficient Language Models

Y Liu, X Peng, Y Zhang, X Ke, S Deng, J Cao, C Ma… - researchgate.net
Large language models have repeatedly shown outstanding performance across diverse
applications. However, deploying these models can inadvertently risk user privacy. The …

[PDF][PDF] Mean Estimation Strikes Back! An Efficient Solution to Private Classification

Y Zhu, H Zhang - tpdp.journalprivacyconfidentiality.org
Differentially private (DP) mean estimation is arguably the most comprehensively studied
problems in the DP literature, leading to various efficient estimators that achieve optimal …