Understanding black-box predictions via influence functions

PW Koh, P Liang - International conference on machine …, 2017 - proceedings.mlr.press
How can we explain the predictions of a black-box model? In this paper, we use influence
functions—a classic technique from robust statistics—to trace a model's prediction through …

Survey on cyberspace security

H Zhang, W Han, X Lai, D Lin, J Ma, JH Li - Science China Information …, 2015 - Springer
Along with the rapid development and wide application of information technology, human
society has entered the information era. In this era, people live and work in cyberspace …

Dataset pruning: Reducing training data by examining generalization influence

S Yang, Z Xie, H Peng, M Xu, M Sun, P Li - arXiv preprint arXiv …, 2022 - arxiv.org
The great success of deep learning heavily relies on increasingly larger training data, which
comes at a price of huge computational and infrastructural costs. This poses crucial …

On the accuracy of influence functions for measuring group effects

PWW Koh, KS Ang, H Teo… - Advances in neural …, 2019 - proceedings.neurips.cc
Influence functions estimate the effect of removing a training point on a model without the
need to retrain. They are based on a first-order Taylor approximation that is guaranteed to …

Achieving fairness at no utility cost via data reweighing with influence

P Li, H Liu - International Conference on Machine Learning, 2022 - proceedings.mlr.press
With the fast development of algorithmic governance, fairness has become a compulsory
property for machine learning models to suppress unintentional discrimination. In this paper …

Fastif: Scalable influence functions for efficient model interpretation and debugging

H Guo, NF Rajani, P Hase, M Bansal… - arXiv preprint arXiv …, 2020 - arxiv.org
Influence functions approximate the" influences" of training data-points for test predictions
and have a wide variety of applications. Despite the popularity, their computational cost …

The detection of epileptic seizure signals based on fuzzy entropy

J Xiang, C Li, H Li, R Cao, B Wang, X Han… - Journal of neuroscience …, 2015 - Elsevier
Background Entropy is a nonlinear index that can reflect the degree of chaos within a
system. It is often used to analyze epileptic electroencephalograms (EEG) to detect whether …

Partial label learning via label influence function

X Gong, D Yuan, W Bao - International Conference on …, 2022 - proceedings.mlr.press
To deal with ambiguities in partial label learning (PLL), state-of-the-art strategies implement
disambiguations by identifying the ground-truth label directly from the candidate label set …

Refined Learning Bounds for Kernel and Approximate -Means

Y Liu - Advances in neural information processing systems, 2021 - proceedings.neurips.cc
Kernel $ k $-means is one of the most popular approaches to clustering and its theoretical
properties have been investigated for decades. However, the existing state-of-the-art risk …

Multi-class learning: From theory to algorithm

J Li, Y Liu, R Yin, H Zhang, L Ding… - Advances in Neural …, 2018 - proceedings.neurips.cc
In this paper, we study the generalization performance of multi-class classification and
obtain a shaper data-dependent generalization error bound with fast convergence rate …