A comprehensive survey on pretrained foundation models: A history from bert to chatgpt
Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks with different data modalities. A PFM (eg, BERT, ChatGPT, and GPT-4) is …
downstream tasks with different data modalities. A PFM (eg, BERT, ChatGPT, and GPT-4) is …
[HTML][HTML] GPT models in construction industry: Opportunities, limitations, and a use case validation
Abstract Large Language Models (LLMs) trained on large data sets came into prominence in
2018 after Google introduced BERT. Subsequently, different LLMs such as GPT models from …
2018 after Google introduced BERT. Subsequently, different LLMs such as GPT models from …
A survey of large language models
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
Large language models for information retrieval: A survey
As a primary means of information acquisition, information retrieval (IR) systems, such as
search engines, have integrated themselves into our daily lives. These systems also serve …
search engines, have integrated themselves into our daily lives. These systems also serve …
Exploring the potential of large language models (llms) in learning on graphs
Learning on Graphs has attracted immense attention due to its wide real-world applications.
The most popular pipeline for learning on graphs with textual node attributes primarily relies …
The most popular pipeline for learning on graphs with textual node attributes primarily relies …
Text embeddings by weakly-supervised contrastive pre-training
This paper presents E5, a family of state-of-the-art text embeddings that transfer well to a
wide range of tasks. The model is trained in a contrastive manner with weak supervision …
wide range of tasks. The model is trained in a contrastive manner with weak supervision …
MTEB: Massive text embedding benchmark
Text embeddings are commonly evaluated on a small set of datasets from a single task not
covering their possible applications to other tasks. It is unclear whether state-of-the-art …
covering their possible applications to other tasks. It is unclear whether state-of-the-art …
Video pretraining (vpt): Learning to act by watching unlabeled online videos
Pretraining on noisy, internet-scale datasets has been heavily studied as a technique for
training models with broad, general capabilities for text, images, and other modalities …
training models with broad, general capabilities for text, images, and other modalities …
ChemCrow: Augmenting large-language models with chemistry tools
Over the last decades, excellent computational chemistry tools have been developed.
Integrating them into a single platform with enhanced accessibility could help reaching their …
Integrating them into a single platform with enhanced accessibility could help reaching their …
Augmenting large language models with chemistry tools
Large language models (LLMs) have shown strong performance in tasks across domains
but struggle with chemistry-related problems. These models also lack access to external …
but struggle with chemistry-related problems. These models also lack access to external …