A comprehensive study of ChatGPT: advancements, limitations, and ethical considerations in natural language processing and cybersecurity

M Alawida, S Mejri, A Mehmood, B Chikhaoui… - Information, 2023 - mdpi.com
This paper presents an in-depth study of ChatGPT, a state-of-the-art language model that is
revolutionizing generative text. We provide a comprehensive analysis of its architecture …

Chatgpt is not enough: Enhancing large language models with knowledge graphs for fact-aware language modeling

L Yang, H Chen, Z Li, X Ding, X Wu - arXiv preprint arXiv:2306.11489, 2023 - arxiv.org
Recently, ChatGPT, a representative large language model (LLM), has gained considerable
attention due to its powerful emergent abilities. Some researchers suggest that LLMs could …

Twhin-bert: A socially-enriched pre-trained language model for multilingual tweet representations at twitter

X Zhang, Y Malkov, O Florez, S Park… - Proceedings of the 29th …, 2023 - dl.acm.org
Pre-trained language models (PLMs) are fundamental for natural language processing
applications. Most existing PLMs are not tailored to the noisy user-generated text on social …

Oag-bench: a human-curated benchmark for academic graph mining

F Zhang, S Shi, Y Zhu, B Chen, Y Cen, J Yu… - Proceedings of the 30th …, 2024 - dl.acm.org
With the rapid proliferation of scientific literature, versatile academic knowledge services
increasingly rely on comprehensive academic graph mining. Despite the availability of …

Give us the facts: Enhancing large language models with knowledge graphs for fact-aware language modeling

L Yang, H Chen, Z Li, X Ding… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Recently, ChatGPT, a representative large language model (LLM), has gained considerable
attention. Due to their powerful emergent abilities, recent LLMs are considered as a possible …

A systematic review of transformer-based pre-trained language models through self-supervised learning

E Kotei, R Thirunavukarasu - Information, 2023 - mdpi.com
Transfer learning is a technique utilized in deep learning applications to transmit learned
inference to a different target domain. The approach is mainly to solve the problem of a few …

The effect of metadata on scientific literature tagging: A cross-field cross-model study

Y Zhang, B Jin, Q Zhu, Y Meng, J Han - Proceedings of the ACM Web …, 2023 - dl.acm.org
Due to the exponential growth of scientific publications on the Web, there is a pressing need
to tag each paper with fine-grained topics so that researchers can track their interested fields …

Oag: Linking entities across large-scale heterogeneous knowledge graphs

F Zhang, X Liu, J Tang, Y Dong, P Yao… - … on Knowledge and …, 2022 - ieeexplore.ieee.org
Different knowledge graphs for the same domain are often uniquely housed on the Web.
Effectively linking entities from different graphs is critical for building an open and …

Pretraining language models with text-attributed heterogeneous graphs

T Zou, L Yu, Y Huang, L Sun, B Du - arXiv preprint arXiv:2310.12580, 2023 - arxiv.org
In many real-world scenarios (eg, academic networks, social platforms), different types of
entities are not only associated with texts but also connected by various relationships, which …

Weakly supervised multi-label classification of full-text scientific papers

Y Zhang, B Jin, X Chen, Y Shen, Y Zhang… - Proceedings of the 29th …, 2023 - dl.acm.org
Instead of relying on human-annotated training samples to build a classifier, weakly
supervised scientific paper classification aims to classify papers only using category …