[HTML][HTML] A survey on large language model (llm) security and privacy: The good, the bad, and the ugly

Y Yao, J Duan, K Xu, Y Cai, Z Sun, Y Zhang - High-Confidence Computing, 2024 - Elsevier
Abstract Large Language Models (LLMs), such as ChatGPT and Bard, have revolutionized
natural language understanding and generation. They possess deep language …

I know what you trained last summer: A survey on stealing machine learning models and defences

D Oliynyk, R Mayer, A Rauber - ACM Computing Surveys, 2023 - dl.acm.org
Machine-Learning-as-a-Service (MLaaS) has become a widespread paradigm, making
even the most complex Machine Learning models available for clients via, eg, a pay-per …

Data-free model extraction

JB Truong, P Maini, RJ Walls… - Proceedings of the …, 2021 - openaccess.thecvf.com
Current model extraction attacks assume that the adversary has access to a surrogate
dataset with characteristics similar to the proprietary data used to train the victim model. This …

Towards data-free model stealing in a hard label setting

S Sanyal, S Addepalli, RV Babu - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Abstract Machine learning models deployed as a service (MLaaS) are susceptible to model
stealing attacks, where an adversary attempts to steal the model within a restricted access …

Lion: Adversarial distillation of proprietary large language models

Y Jiang, C Chan, M Chen, W Wang - arXiv preprint arXiv:2305.12870, 2023 - arxiv.org
The practice of transferring knowledge from a sophisticated, proprietary large language
model (LLM) to a compact, open-source LLM has garnered considerable attention. Previous …

Black-box attacks on sequential recommenders via data-free model extraction

Z Yue, Z He, H Zeng, J McAuley - … of the 15th ACM conference on …, 2021 - dl.acm.org
We investigate whether model extraction can be used to 'steal'the weights of sequential
recommender systems, and the potential threats posed to victims of such attacks. This type …

Hermes attack: Steal {DNN} models with lossless inference accuracy

Y Zhu, Y Cheng, H Zhou, Y Lu - 30th USENIX Security Symposium …, 2021 - usenix.org
Deep Neural Network (DNN) models become one of the most valuable enterprise assets
due to their critical roles in all aspects of applications. With the trend of privatization …

Defending against data-free model extraction by distributionally robust defensive training

Z Wang, L Shen, T Liu, T Duan, Y Zhu… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Data-Free Model Extraction (DFME) aims to clone a black-box model without
knowing its original training data distribution, making it much easier for attackers to steal …

Data-free knowledge transfer: A survey

Y Liu, W Zhang, J Wang, J Wang - arXiv preprint arXiv:2112.15278, 2021 - arxiv.org
In the last decade, many deep learning models have been well trained and made a great
success in various fields of machine intelligence, especially for computer vision and natural …

Stolenencoder: stealing pre-trained encoders in self-supervised learning

Y Liu, J Jia, H Liu, NZ Gong - Proceedings of the 2022 ACM SIGSAC …, 2022 - dl.acm.org
Pre-trained encoders are general-purpose feature extractors that can be used for many
downstream tasks. Recent progress in self-supervised learning can pre-train highly effective …