The breakthrough of large language models release for medical applications: 1-year timeline and perspectives

M Cascella, F Semeraro, J Montomoli, V Bellini… - Journal of Medical …, 2024 - Springer
Within the domain of Natural Language Processing (NLP), Large Language Models (LLMs)
represent sophisticated models engineered to comprehend, generate, and manipulate text …

Scaling vision transformers to 22 billion parameters

M Dehghani, J Djolonga, B Mustafa… - International …, 2023 - proceedings.mlr.press
The scaling of Transformers has driven breakthrough capabilities for language models. At
present, the largest large language models (LLMs) contain upwards of 100B parameters …

[HTML][HTML] Large language models encode clinical knowledge

K Singhal, S Azizi, T Tu, SS Mahdavi, J Wei, HW Chung… - Nature, 2023 - nature.com
Large language models (LLMs) have demonstrated impressive capabilities, but the bar for
clinical applications is high. Attempts to assess the clinical knowledge of models typically …

Large language models encode clinical knowledge

K Singhal, S Azizi, T Tu, SS Mahdavi, J Wei… - arXiv preprint arXiv …, 2022 - arxiv.org
Large language models (LLMs) have demonstrated impressive capabilities in natural
language understanding and generation, but the quality bar for medical and clinical …

Larger language models do in-context learning differently

J Wei, J Wei, Y Tay, D Tran, A Webson, Y Lu… - arXiv preprint arXiv …, 2023 - arxiv.org
We study how in-context learning (ICL) in language models is affected by semantic priors
versus input-label mappings. We investigate two setups-ICL with flipped labels and ICL with …

Out-of-distribution detection and selective generation for conditional language models

J Ren, J Luo, Y Zhao, K Krishna, M Saleh… - The Eleventh …, 2022 - openreview.net
Machine learning algorithms typically assume independent and identically distributed
samples in training and at test time (IID). Much work has shown that high-performing ML …

Lightning Pose: improved animal pose estimation via semi-supervised learning, Bayesian ensembling and cloud-native open-source tools

D Biderman, MR Whiteway, C Hurwitz, N Greenspan… - Nature …, 2024 - nature.com
Contemporary pose estimation methods enable precise measurements of behavior via
supervised deep learning with hand-labeled video frames. Although effective in many cases …

Probvlm: Probabilistic adapter for frozen vison-language models

U Upadhyay, S Karthik, M Mancini… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale vision-language models (VLMs) like CLIP successfully find correspondences
between images and text. Through the standard deterministic mapping process, an image or …

Deup: Direct epistemic uncertainty prediction

S Lahlou, M Jain, H Nekoei, VI Butoi, P Bertin… - arXiv preprint arXiv …, 2021 - arxiv.org
Epistemic Uncertainty is a measure of the lack of knowledge of a learner which diminishes
with more evidence. While existing work focuses on using the variance of the Bayesian …

Deep ensembles work, but are they necessary?

T Abe, EK Buchanan, G Pleiss… - Advances in …, 2022 - proceedings.neurips.cc
Ensembling neural networks is an effective way to increase accuracy, and can often match
the performance of individual larger models. This observation poses a natural question …