Adding robustness to support vector machines against adversarial reverse engineering
Proceedings of the 23rd ACM international conference on conference on …, 2014•dl.acm.org
Many classification algorithms have been successfully deployed in security-sensitive
applications including spam filters and intrusion detection systems. Under such adversarial
environments, adversaries can generate exploratory attacks against the defender such as
evasion and reverse engineering. In this paper, we discuss why reverse engineering attacks
can be carried out quite efficiently against fixed classifiers, and investigate the use of
randomization as a suitable strategy for mitigating their risk. In particular, we derive a …
applications including spam filters and intrusion detection systems. Under such adversarial
environments, adversaries can generate exploratory attacks against the defender such as
evasion and reverse engineering. In this paper, we discuss why reverse engineering attacks
can be carried out quite efficiently against fixed classifiers, and investigate the use of
randomization as a suitable strategy for mitigating their risk. In particular, we derive a …
Many classification algorithms have been successfully deployed in security-sensitive applications including spam filters and intrusion detection systems. Under such adversarial environments, adversaries can generate exploratory attacks against the defender such as evasion and reverse engineering. In this paper, we discuss why reverse engineering attacks can be carried out quite efficiently against fixed classifiers, and investigate the use of randomization as a suitable strategy for mitigating their risk. In particular, we derive a semidefinite programming (SDP) formulation for learning a distribution of classifiers subject to the constraint that any single classifier picked at random from such distribution provides reliable predictions with a high probability. We analyze the tradeoff between variance of the distribution and its predictive accuracy, and establish that one can almost always incorporate randomization with large variance without incurring a loss in accuracy. In other words, the conventional approach of using a fixed classifier in adversarial environments is generally Pareto suboptimal. Finally, we validate such conclusions on both synthetic and real-world classification problems.
ACM Digital Library
以上显示的是最相近的搜索结果。 查看全部搜索结果