Formal Privacy Proof of Data Encoding: The Possibility and Impossibility of Learnable Encryption

H Xiao, GE Suh, S Devadas - Proceedings of the 2024 on ACM SIGSAC …, 2024 - dl.acm.org
We initiate a formal study on the concept of learnable obfuscation and aim to answer the
following question: is there a type of data encoding that maintains the" learnability" of …

Bounding the invertibility of privacy-preserving instance encoding using fisher information

K Maeng, C Guo, S Kariyappa… - Advances in Neural …, 2024 - proceedings.neurips.cc
Privacy-preserving instance encoding aims to encode raw data into feature vectors without
revealing their privacy-sensitive information. When designed properly, these encodings can …

A Split-and-Privatize Framework for Large Language Model Fine-Tuning

X Shen, Y Liu, H Liu, J Hong, B Duan, Z Huang… - arXiv preprint arXiv …, 2023 - arxiv.org
Fine-tuning is a prominent technique to adapt a pre-trained language model to downstream
scenarios. In parameter-efficient fine-tuning, only a small subset of modules are trained over …

CPSample: Classifier Protected Sampling for Guarding Training Data During Diffusion

J Kazdan, H Sun, J Han, F Petersen… - arXiv preprint arXiv …, 2024 - arxiv.org
Diffusion models have a tendency to exactly replicate their training data, especially when
trained on small datasets. Most prior work has sought to mitigate this problem by imposing …

Bayes-Nash Generative Privacy Protection Against Membership Inference Attacks

T Zhang, R Venkatesaraman, RK De, BA Malin… - arXiv preprint arXiv …, 2024 - arxiv.org
An ability to share data, even in aggregated form, is critical to advancing both conventional
and data science. However, insofar as such datasets are comprised of individuals, their …

PAC-Private Algorithms

M Sridhar, H Xiao, S Devadas - Cryptology ePrint Archive, 2024 - eprint.iacr.org
Provable privacy typically requires involved analysis and is often associated with
unacceptable accuracy loss. While many empirical verification or approximation methods …

Disentangling data distribution for Federated Learning

X Zhao, H Gu, L Fan, Q Yang, Y Han - arXiv preprint arXiv:2410.12530, 2024 - arxiv.org
Federated Learning (FL) facilitates collaborative training of a global model whose
performance is boosted by private data owned by distributed clients, without compromising …

Private Linear Regression with Differential Privacy and PAC Privacy

H Yang - arXiv preprint arXiv:2412.02578, 2024 - arxiv.org
Linear regression is a fundamental tool for statistical analysis, which has motivated the
development of linear regression methods that satisfy provable privacy guarantees so that …

PAC Privacy Preserving Diffusion Models

Q Xu - 2024 - search.proquest.com
Data privacy protection is garnering increased attention among researchers. Diffusion
models (DMs), particularly with strict differential privacy, can potentially produce images with …

A Novel Review of Stability Techniques for Improved Privacy-Preserving Machine Learning

C DuPlessie, A Gao - arXiv preprint arXiv:2406.00073, 2024 - arxiv.org
Machine learning models have recently enjoyed a significant increase in size and
popularity. However, this growth has created concerns about dataset privacy. To counteract …