The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation

B Plank - arXiv preprint arXiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …

Quark: Controllable text generation with reinforced unlearning

X Lu, S Welleck, J Hessel, L Jiang… - Advances in neural …, 2022 - proceedings.neurips.cc
Large-scale language models often learn behaviors that are misaligned with user
expectations. Generated text may contain offensive or toxic language, contain significant …

Rethinking calibration of deep neural networks: Do not be afraid of overconfidence

DB Wang, L Feng, ML Zhang - Advances in Neural …, 2021 - proceedings.neurips.cc
Capturing accurate uncertainty quantification of the prediction from deep neural networks is
important in many real-world decision-making applications. A reliable predictor is expected …

Locally typical sampling

C Meister, T Pimentel, G Wiher… - Transactions of the …, 2023 - direct.mit.edu
Today's probabilistic language generators fall short when it comes to producing coherent
and fluent text despite the fact that the underlying models perform well under standard …

Why do better loss functions lead to less transferable features?

S Kornblith, T Chen, H Lee… - Advances in Neural …, 2021 - proceedings.neurips.cc
Previous work has proposed many new loss functions and regularizers that improve test
accuracy on image classification tasks. However, it is not clear whether these loss functions …

When does data augmentation help with membership inference attacks?

Y Kaya, T Dumitras - International conference on machine …, 2021 - proceedings.mlr.press
Deep learning models often raise privacy concerns as they leak information about their
training data. This leakage enables membership inference attacks (MIA) that can identify …

Why do nearest neighbor language models work?

FF Xu, U Alon, G Neubig - International Conference on …, 2023 - proceedings.mlr.press
Abstract Language models (LMs) compute the probability of a text by sequentially computing
a representation of an already-seen context and using this representation to predict the next …

Fusemoe: Mixture-of-experts transformers for fleximodal fusion

X Han, H Nguyen, C Harris, N Ho, S Saria - arXiv preprint arXiv …, 2024 - arxiv.org
As machine learning models in critical fields increasingly grapple with multimodal data, they
face the dual challenges of handling a wide array of modalities, often incomplete due to …

Tailoring self-rationalizers with multi-reward distillation

S Ramnath, B Joshi, S Hallinan, X Lu, LH Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LMs) are capable of generating free-text rationales to aid question
answering. However, prior work 1) suggests that useful self-rationalization is emergent only …

Sparsing and smoothing for the seq2seq models

S Zhao, Z Liang, J Wen, J Chen - IEEE Transactions on Artificial …, 2022 - ieeexplore.ieee.org
Current neural language models are trained to minimize cross-entropy and use softmax to
compute the locally normalized probabilities over the target. While this setup provides solid …