Transformers as statisticians: Provable in-context learning with in-context algorithm selection

Y Bai, F Chen, H Wang, C Xiong… - Advances in neural …, 2024 - proceedings.neurips.cc
Neural sequence models based on the transformer architecture have demonstrated
remarkable\emph {in-context learning}(ICL) abilities, where they can perform new tasks …

Assessing generalization of SGD via disagreement

Y Jiang, V Nagarajan, C Baek, JZ Kolter - arXiv preprint arXiv:2106.13799, 2021 - arxiv.org
We empirically show that the test error of deep networks can be estimated by simply training
the same architecture on the same training set but with a different run of Stochastic Gradient …

Beyond confidence: Reliable models should also consider atypicality

M Yuksekgonul, L Zhang, JY Zou… - Advances in Neural …, 2024 - proceedings.neurips.cc
While most machine learning models can provide confidence in their predictions, confidence
is insufficient to understand a prediction's reliability. For instance, the model may have a low …

Early-stopped neural networks are consistent

Z Ji, J Li, M Telgarsky - Advances in Neural Information …, 2021 - proceedings.neurips.cc
This work studies the behavior of shallow ReLU networks trained with the logistic loss via
gradient descent on binary classification data where the underlying data distribution is …

On double-descent in uncertainty quantification in overparametrized models

L Clarté, B Loureiro, F Krzakala… - International …, 2023 - proceedings.mlr.press
Uncertainty quantification is a central challenge in reliable and trustworthy machine
learning. Naive measures such as last-layer scores are well-known to yield overconfident …

T-cal: An optimal test for the calibration of predictive models

D Lee, X Huang, H Hassani, E Dobriban - Journal of Machine Learning …, 2023 - jmlr.org
The prediction accuracy of machine learning methods is steadily increasing, but the
calibration of their uncertainty predictions poses a significant challenge. Numerous works …

Conformal prediction sets improve human decision making

JC Cresswell, Y Sui, B Kumar, N Vouitsis - arXiv preprint arXiv:2401.13744, 2024 - arxiv.org
In response to everyday queries, humans explicitly signal uncertainty and offer alternative
answers when they are unsure. Machine learning models that output calibrated prediction …

Multi-label intent detection via contrastive task specialization of sentence encoders

I Vulić, I Casanueva, G Spithourakis… - Proceedings of the …, 2022 - aclanthology.org
Deploying task-oriented dialog ToD systems for new domains and tasks requires natural
language understanding models that are 1) resource-efficient and work under low-data …

Federated inference with reliable uncertainty quantification over wireless channels via conformal prediction

M Zhu, M Zecchin, S Park, C Guo… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
In this paper, we consider a wireless federated inference scenario in which devices and a
server share a pre-trained machine learning model. The devices communicate statistical …

Analysis of bootstrap and subsampling in high-dimensional regularized regression

L Clarté, A Vandenbroucque, G Dalle… - arXiv preprint arXiv …, 2024 - arxiv.org
We investigate popular resampling methods for estimating the uncertainty of statistical
models, such as subsampling, bootstrap and the jackknife, and their performance in high …