H2o: Heavy-hitter oracle for efficient generative inference of large language models

Z Zhang, Y Sheng, T Zhou, T Chen… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Large Language Models (LLMs), despite their recent impressive accomplishments,
are notably cost-prohibitive to deploy, particularly for applications involving long-content …

Strategies for implementing machine learning algorithms in the clinical practice of radiology

A Chae, MS Yao, H Sagreiya, AD Goldberg… - Radiology, 2024 - pubs.rsna.org
Despite recent advancements in machine learning (ML) applications in health care, there
have been few benefits and improvements to clinical medicine in the hospital setting. To …

The emergence of essential sparsity in large pre-trained models: The weights that matter

A Jaiswal, S Liu, T Chen… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large pre-trained transformers are $\textit {show-stealer} $ in modern-day deep learning,
and it becomes crucial to comprehend the parsimonious patterns that exist within them as …

Retrieving evidence from ehrs with llms: Possibilities and challenges

H Ahsan, DJ McInerney, J Kim, C Potter… - arXiv preprint arXiv …, 2023 - arxiv.org
Unstructured Electronic Health Record (EHR) data often contains critical information
complementary to imaging data that would inform radiologists' diagnoses. However, time …

Instant soup: Cheap pruning ensembles in a single pass can draw lottery tickets from large models

AK Jaiswal, S Liu, T Chen, Y Ding… - … on Machine Learning, 2023 - proceedings.mlr.press
Large pre-trained transformers have been receiving explosive attention in the past few
years, due to their acculturation for numerous downstream applications via fine-tuning, but …

On the opportunities and risks of foundation models for natural language processing in radiology

WF Wiggins, AS Tejani - Radiology: Artificial Intelligence, 2022 - pubs.rsna.org
Ali S. Tejani, MD, is a radiology resident at the University of Texas Southwestern Medical
Center in Dallas, Tex, where he founded the Imaging Informatics and Business Intelligence …

Sparse moe as the new dropout: Scaling dense and self-slimmable transformers

T Chen, Z Zhang, A Jaiswal, S Liu, Z Wang - arXiv preprint arXiv …, 2023 - arxiv.org
Despite their remarkable achievement, gigantic transformers encounter significant
drawbacks, including exorbitant computational and memory footprints during training, as …

Performance of multiple pretrained BERT models to automate and accelerate data annotation for large datasets

AS Tejani, YS Ng, Y Xi, JR Fielding… - Radiology: Artificial …, 2022 - pubs.rsna.org
Purpose To develop and evaluate domain-specific and pretrained bidirectional encoder
representations from transformers (BERT) models in a transfer learning task on varying …

[HTML][HTML] Less likely brainstorming: Using language models to generate alternative hypotheses

L Tang, Y Peng, Y Wang, Y Ding, G Durrett… - Proceedings of the …, 2023 - ncbi.nlm.nih.gov
A human decision-maker benefits the most from an AI assistant that corrects for their biases.
For problems such as generating interpretation of a radiology report given findings, a system …

Information extraction from weakly structured radiological reports with natural language queries

A Dada, TL Ufer, M Kim, M Hasin, N Spieker… - European …, 2024 - Springer
Objectives Provide physicians and researchers an efficient way to extract information from
weakly structured radiology reports with natural language processing (NLP) machine …