A survey of knowledge enhanced pre-trained language models

L Hu, Z Liu, Z Zhao, L Hou, L Nie… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Pre-trained Language Models (PLMs) which are trained on large text corpus via self-
supervised learning method, have yielded promising performance on various tasks in …

Named entity recognition and relation extraction: State-of-the-art

Z Nasar, SW Jaffry, MK Malik - ACM Computing Surveys (CSUR), 2021 - dl.acm.org
With the advent of Web 2.0, there exist many online platforms that result in massive textual-
data production. With ever-increasing textual data at hand, it is of immense importance to …

BioGPT: generative pre-trained transformer for biomedical text generation and mining

R Luo, L Sun, Y Xia, T Qin, S Zhang… - Briefings in …, 2022 - academic.oup.com
Pre-trained language models have attracted increasing attention in the biomedical domain,
inspired by their great success in the general natural language domain. Among the two main …

Galactica: A large language model for science

R Taylor, M Kardas, G Cucurull, T Scialom… - arXiv preprint arXiv …, 2022 - arxiv.org
Information overload is a major obstacle to scientific progress. The explosive growth in
scientific literature and data has made it ever harder to discover useful insights in a large …

Linkbert: Pretraining language models with document links

M Yasunaga, J Leskovec, P Liang - arXiv preprint arXiv:2203.15827, 2022 - arxiv.org
Language model (LM) pretraining can learn various knowledge from text corpora, helping
downstream tasks. However, existing methods such as BERT model a single document, and …

A study of generative large language model for medical research and healthcare

C Peng, X Yang, A Chen, KE Smith… - NPJ digital …, 2023 - nature.com
There are enormous enthusiasm and concerns in applying large language models (LLMs) to
healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT …

Structured information extraction from scientific text with large language models

J Dagdelen, A Dunn, S Lee, N Walker… - Nature …, 2024 - nature.com
Extracting structured knowledge from scientific text remains a challenging task for machine
learning models. Here, we present a simple approach to joint named entity recognition and …

Domain-specific language model pretraining for biomedical natural language processing

Y Gu, R Tinn, H Cheng, M Lucas, N Usuyama… - ACM Transactions on …, 2021 - dl.acm.org
Pretraining large neural language models, such as BERT, has led to impressive gains on
many natural language processing (NLP) tasks. However, most pretraining efforts focus on …

Does synthetic data generation of llms help clinical text mining?

R Tang, X Han, X Jiang, X Hu - arXiv preprint arXiv:2303.04360, 2023 - arxiv.org
Recent advancements in large language models (LLMs) have led to the development of
highly potent models like OpenAI's ChatGPT. These models have exhibited exceptional …

Transfer learning in biomedical natural language processing: an evaluation of BERT and ELMo on ten benchmarking datasets

Y Peng, S Yan, Z Lu - arXiv preprint arXiv:1906.05474, 2019 - arxiv.org
Inspired by the success of the General Language Understanding Evaluation benchmark, we
introduce the Biomedical Language Understanding Evaluation (BLUE) benchmark to …