A survey on federated unlearning: Challenges, methods, and future directions
In recent years, the notion of “the right to be forgotten”(RTBF) has become a crucial aspect of
data privacy for digital trust and AI safety, requiring the provision of mechanisms that support …
data privacy for digital trust and AI safety, requiring the provision of mechanisms that support …
Threats, attacks, and defenses in machine unlearning: A survey
Machine Unlearning (MU) has recently gained considerable attention due to its potential to
achieve Safe AI by removing the influence of specific data from trained Machine Learning …
achieve Safe AI by removing the influence of specific data from trained Machine Learning …
Privacy in large language models: Attacks, defenses and future directions
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …
effectively tackle various downstream NLP tasks and unify these tasks into generative …
Model merging in llms, mllms, and beyond: Methods, theories, applications and opportunities
Model merging is an efficient empowerment technique in the machine learning community
that does not require the collection of raw training data and does not require expensive …
that does not require the collection of raw training data and does not require expensive …
Negative preference optimization: From catastrophic collapse to effective unlearning
Large Language Models (LLMs) often memorize sensitive, private, or copyrighted data
during pre-training. LLM unlearning aims to eliminate the influence of undesirable data from …
during pre-training. LLM unlearning aims to eliminate the influence of undesirable data from …
Targeted latent adversarial training improves robustness to persistent harmful behaviors in llms
Large language models (LLMs) can often be made to behave in undesirable ways that they
are explicitly fine-tuned not to. For example, the LLM red-teaming literature has produced a …
are explicitly fine-tuned not to. For example, the LLM red-teaming literature has produced a …
Open problems in technical ai governance
AI progress is creating a growing range of risks and opportunities, but it is often unclear how
they should be navigated. In many cases, the barriers and uncertainties faced are at least …
they should be navigated. In many cases, the barriers and uncertainties faced are at least …
Machine unlearning in generative ai: A survey
Generative AI technologies have been deployed in many places, such as (multimodal) large
language models and vision generative models. Their remarkable performance should be …
language models and vision generative models. Their remarkable performance should be …
Ununlearning: Unlearning is not sufficient for content regulation in advanced generative ai
Exact unlearning was first introduced as a privacy mechanism that allowed a user to retract
their data from machine learning models on request. Shortly after, inexact schemes were …
their data from machine learning models on request. Shortly after, inexact schemes were …
Safe unlearning: A surprisingly effective and generalizable solution to defend against jailbreak attacks
LLMs are known to be vulnerable to jailbreak attacks, even after safety alignment. An
important observation is that, while different types of jailbreak attacks can generate …
important observation is that, while different types of jailbreak attacks can generate …