Big data in public health: terminology, machine learning, and privacy
The digital world is generating data at a staggering and still increasing rate. While these “big
data” have unlocked novel opportunities to understand public health, they hold still greater …
data” have unlocked novel opportunities to understand public health, they hold still greater …
Generalized out-of-distribution detection: A survey
Abstract Out-of-distribution (OOD) detection is critical to ensuring the reliability and safety of
machine learning systems. For instance, in autonomous driving, we would like the driving …
machine learning systems. For instance, in autonomous driving, we would like the driving …
Combating noisy labels by agreement: A joint training method with co-regularization
Deep Learning with noisy labels is a practically challenging problem in weakly-supervised
learning. The state-of-the-art approaches" Decoupling" and" Co-teaching+" claim that the" …
learning. The state-of-the-art approaches" Decoupling" and" Co-teaching+" claim that the" …
How does disagreement help generalization against label corruption?
Learning with noisy labels is one of the hottest problems in weakly-supervised learning.
Based on memorization effects of deep neural networks, training on small-loss instances …
Based on memorization effects of deep neural networks, training on small-loss instances …
Co-teaching: Robust training of deep neural networks with extremely noisy labels
Deep learning with noisy labels is practically challenging, as the capacity of deep models is
so high that they can totally memorize these noisy labels sooner or later during training …
so high that they can totally memorize these noisy labels sooner or later during training …
Estimating noise transition matrix with label correlations for noisy multi-label learning
In label-noise learning, the noise transition matrix, bridging the class posterior for noisy and
clean data, has been widely exploited to learn statistically consistent classifiers. The …
clean data, has been widely exploited to learn statistically consistent classifiers. The …
Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels
Recent deep networks are capable of memorizing the entire data even when the labels are
completely random. To overcome the overfitting on corrupted labels, we propose a novel …
completely random. To overcome the overfitting on corrupted labels, we propose a novel …
Making deep neural networks robust to label noise: A loss correction approach
We present a theoretically grounded approach to train deep neural networks, including
recurrent networks, subject to class-dependent label noise. We propose two procedures for …
recurrent networks, subject to class-dependent label noise. We propose two procedures for …
Inferring the molecular and phenotypic impact of amino acid variants with MutPred2
Identifying pathogenic variants and underlying functional alterations is challenging. To this
end, we introduce MutPred2, a tool that improves the prioritization of pathogenic amino acid …
end, we introduce MutPred2, a tool that improves the prioritization of pathogenic amino acid …
Positive-unlabeled learning with non-negative risk estimator
From only positive (P) and unlabeled (U) data, a binary classifier could be trained with PU
learning, in which the state of the art is unbiased PU learning. However, if its model is very …
learning, in which the state of the art is unbiased PU learning. However, if its model is very …