Promptcare: Prompt copyright protection by watermark injection and verification

H Yao, J Lou, Z Qin, K Ren - 2024 IEEE Symposium on Security …, 2024 - ieeexplore.ieee.org
Large language models (LLMs) have witnessed a meteoric rise in popularity among the
general public users over the past few months, facilitating diverse downstream tasks with …

Towards reliable and efficient backdoor trigger inversion via decoupling benign features

X Xu, K Huang, Y Li, Z Qin, K Ren - The Twelfth International …, 2024 - openreview.net
Recent studies revealed that using third-party models may lead to backdoor threats, where
adversaries can maliciously manipulate model predictions based on backdoors implanted …

Towards faithful xai evaluation via generalization-limited backdoor watermark

M Ya, Y Li, T Dai, B Wang, Y Jiang… - The Twelfth International …, 2023 - openreview.net
Saliency-based representation visualization (SRV)($ eg $, Grad-CAM) is one of the most
classical and widely adopted explainable artificial intelligence (XAI) methods for its simplicity …

Defending against data-free model extraction by distributionally robust defensive training

Z Wang, L Shen, T Liu, T Duan, Y Zhu… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Data-Free Model Extraction (DFME) aims to clone a black-box model without
knowing its original training data distribution, making it much easier for attackers to steal …

Towards stealthy backdoor attacks against speech recognition via elements of sound

H Cai, P Zhang, H Dong, Y Xiao… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Deep neural networks (DNNs) have been widely and successfully adopted and deployed in
various applications of speech recognition. Recently, a few works revealed that these …

Defenses in adversarial machine learning: A survey

B Wu, S Wei, M Zhu, M Zheng, Z Zhu, M Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Adversarial phenomenon has been widely observed in machine learning (ML) systems,
especially in those using deep neural networks, describing that ML systems may produce …

Pointncbw: Towards dataset ownership verification for point clouds via negative clean-label backdoor watermark

C Wei, Y Wang, K Gao, S Shao, Y Li… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Recently, point clouds have been widely used in computer vision, whereas their collection is
time-consuming and expensive. As such, point cloud datasets are the valuable intellectual …

A general framework for data-use auditing of ml models

Z Huang, NZ Gong, MK Reiter - Proceedings of the 2024 on ACM …, 2024 - dl.acm.org
Auditing the use of data in training machine-learning (ML) models is an increasingly
pressing challenge, as myriad ML practitioners routinely leverage the effort of content …

Towards reliable verification of unauthorized data usage in personalized text-to-image diffusion models

B Li, Y Wei, Y Fu, Z Wang, Y Li, J Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Text-to-image diffusion models are pushing the boundaries of what generative AI can
achieve in our lives. Beyond their ability to generate general images, new personalization …

M-to-N Backdoor Paradigm: A Multi-Trigger and Multi-Target Attack to Deep Learning Models

L Hou, Z Hua, Y Li, Y Zheng… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Deep neural networks (DNNs) are vulnerable to backdoor attacks, where a backdoored
model behaves normally with clean inputs but exhibits attacker-specified behaviors upon the …