A survey of machine unlearning

TT Nguyen, TT Huynh, Z Ren, PL Nguyen… - arXiv preprint arXiv …, 2022 - arxiv.org
Today, computer systems hold large amounts of personal data. Yet while such an
abundance of data allows breakthroughs in artificial intelligence, and especially machine …

Composable interventions for language models

A Kolbeinsson, K O'Brien, T Huang, S Gao… - arXiv preprint arXiv …, 2024 - arxiv.org
Test-time interventions for language models can enhance factual accuracy, mitigate harmful
outputs, and improve model efficiency without costly retraining. But despite a flood of new …

Position: LLM Unlearning Benchmarks are Weak Measures of Progress

P Thaker, S Hu, N Kale, Y Maurya, ZS Wu… - arXiv preprint arXiv …, 2024 - arxiv.org
Unlearning methods have the potential to improve the privacy and safety of large language
models (LLMs) by removing sensitive or harmful information post hoc. The LLM unlearning …

A Closer Look at Machine Unlearning for Large Language Models

X Yuan, T Pang, C Du, K Chen, W Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) may memorize sensitive or copyrighted content, raising
privacy and legal concerns. Due to the high cost of retraining from scratch, researchers …