A comprehensive survey on pretrained foundation models: A history from bert to chatgpt
Abstract Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
A comprehensive survey on applications of transformers for deep learning tasks
Abstract Transformers are Deep Neural Networks (DNN) that utilize a self-attention
mechanism to capture contextual relationships within sequential data. Unlike traditional …
mechanism to capture contextual relationships within sequential data. Unlike traditional …
A bibliometric review of large language models research from 2017 to 2023
Large language models (LLMs), such as OpenAI's Generative Pre-trained Transformer
(GPT), are a class of language models that have demonstrated outstanding performance …
(GPT), are a class of language models that have demonstrated outstanding performance …
[PDF][PDF] mt5: A massively multilingual pre-trained text-to-text transformer
L Xue - arXiv preprint arXiv:2010.11934, 2020 - fq.pkwyx.com
The recent" Text-to-Text Transfer Transformer"(T5) leveraged a unified text-to-text format and
scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this …
scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this …
Ammus: A survey of transformer-based pretrained models in natural language processing
KS Kalyan, A Rajasekharan, S Sangeetha - arXiv preprint arXiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …
almost every NLP task. The evolution of these models started with GPT and BERT. These …
ARBERT & MARBERT: Deep bidirectional transformers for Arabic
M Abdul-Mageed, AR Elmadany… - arXiv preprint arXiv …, 2020 - arxiv.org
Pre-trained language models (LMs) are currently integral to many natural language
processing systems. Although multilingual LMs were also introduced to serve many …
processing systems. Although multilingual LMs were also introduced to serve many …
[HTML][HTML] Pre-trained models for natural language processing: A survey
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …
IndicNLPSuite: Monolingual corpora, evaluation benchmarks and pre-trained multilingual language models for Indian languages
D Kakwani, A Kunchukuttan, S Golla… - Findings of the …, 2020 - aclanthology.org
In this paper, we introduce NLP resources for 11 major Indian languages from two major
language families. These resources include:(a) large-scale sentence-level monolingual …
language families. These resources include:(a) large-scale sentence-level monolingual …
Spanish pre-trained bert model and evaluation data
J Cañete, G Chaperon, R Fuentes, JH Ho… - arXiv preprint arXiv …, 2023 - arxiv.org
The Spanish language is one of the top 5 spoken languages in the world. Nevertheless,
finding resources to train or evaluate Spanish language models is not an easy task. In this …
finding resources to train or evaluate Spanish language models is not an easy task. In this …
[引用][C] Transformers: State-of-the-Art Natural Language Processing
T Wolf - arXiv preprint arXiv:1910.03771, 2020