Gptscore: Evaluate as you desire

J Fu, SK Ng, Z Jiang, P Liu - arXiv preprint arXiv:2302.04166, 2023 - arxiv.org
Generative Artificial Intelligence (AI) has enabled the development of sophisticated models
that are capable of producing high-caliber text, images, and other outputs through the …

Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models

D Dai, C Deng, C Zhao, RX Xu, H Gao, D Chen… - arXiv preprint arXiv …, 2024 - arxiv.org
In the era of large language models, Mixture-of-Experts (MoE) is a promising architecture for
managing computational costs when scaling up model parameters. However, conventional …

Evaluating hallucinations in chinese large language models

Q Cheng, T Sun, W Zhang, S Wang, X Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we establish a benchmark named HalluQA (Chinese Hallucination Question-
Answering) to measure the hallucination phenomenon in Chinese large language models …

[PDF][PDF] Unifying the perspectives of nlp and software engineering: A survey on language models for code

Z Zhang, C Chen, B Liu, C Liao, Z Gong… - arXiv preprint arXiv …, 2023 - simg.baai.ac.cn
In this work we systematically review the recent advancements in code processing with
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …

Data contamination quiz: A tool to detect and estimate contamination in large language models

S Golchin, M Surdeanu - arXiv preprint arXiv:2311.06233, 2023 - arxiv.org
We propose the Data Contamination Quiz, a simple and effective approach to detect data
contamination in large language models (LLMs) and estimate the amount of it. Specifically …

Oceangpt: A large language model for ocean science tasks

Z Bi, N Zhang, Y Xue, Y Ou, D Ji, G Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Ocean science, which delves into the oceans that are reservoirs of life and biodiversity, is of
great significance given that oceans cover over 70% of our planet's surface. Recently …

Make pre-trained model reversible: From parameter to memory efficient fine-tuning

B Liao, S Tan, C Monz - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Parameter-efficient fine-tuning (PEFT) of pre-trained language models (PLMs) has emerged
as a highly successful approach, with training only a small number of parameters without …

CARE-MI: chinese benchmark for misinformation evaluation in maternity and infant care

T Xiang, L Li, W Li, M Bai, L Wei… - Advances in Neural …, 2023 - proceedings.neurips.cc
The recent advances in natural language processing (NLP), have led to a new trend of
applying large language models (LLMs) to real-world scenarios. While the latest LLMs are …

MedChatZH: A tuning LLM for traditional Chinese medicine consultations

Y Tan, Z Zhang, M Li, F Pan, H Duan, Z Huang… - Computers in Biology …, 2024 - Elsevier
Abstract Generative Large Language Models (LLMs) have achieved significant success in
various natural language processing tasks, including Question-Answering (QA) and …

Marinegpt: Unlocking secrets of ocean to the public

Z Zheng, J Zhang, TA Vu, S Diao, YHW Tim… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs), such as ChatGPT/GPT-4, have proven to be powerful tools
in promoting the user experience as an AI assistant. The continuous works are proposing …