Gptscore: Evaluate as you desire
Generative Artificial Intelligence (AI) has enabled the development of sophisticated models
that are capable of producing high-caliber text, images, and other outputs through the …
that are capable of producing high-caliber text, images, and other outputs through the …
Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models
In the era of large language models, Mixture-of-Experts (MoE) is a promising architecture for
managing computational costs when scaling up model parameters. However, conventional …
managing computational costs when scaling up model parameters. However, conventional …
Evaluating hallucinations in chinese large language models
In this paper, we establish a benchmark named HalluQA (Chinese Hallucination Question-
Answering) to measure the hallucination phenomenon in Chinese large language models …
Answering) to measure the hallucination phenomenon in Chinese large language models …
[PDF][PDF] Unifying the perspectives of nlp and software engineering: A survey on language models for code
Z Zhang, C Chen, B Liu, C Liao, Z Gong… - arXiv preprint arXiv …, 2023 - simg.baai.ac.cn
In this work we systematically review the recent advancements in code processing with
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …
Data contamination quiz: A tool to detect and estimate contamination in large language models
S Golchin, M Surdeanu - arXiv preprint arXiv:2311.06233, 2023 - arxiv.org
We propose the Data Contamination Quiz, a simple and effective approach to detect data
contamination in large language models (LLMs) and estimate the amount of it. Specifically …
contamination in large language models (LLMs) and estimate the amount of it. Specifically …
Oceangpt: A large language model for ocean science tasks
Ocean science, which delves into the oceans that are reservoirs of life and biodiversity, is of
great significance given that oceans cover over 70% of our planet's surface. Recently …
great significance given that oceans cover over 70% of our planet's surface. Recently …
Make pre-trained model reversible: From parameter to memory efficient fine-tuning
Parameter-efficient fine-tuning (PEFT) of pre-trained language models (PLMs) has emerged
as a highly successful approach, with training only a small number of parameters without …
as a highly successful approach, with training only a small number of parameters without …
CARE-MI: chinese benchmark for misinformation evaluation in maternity and infant care
The recent advances in natural language processing (NLP), have led to a new trend of
applying large language models (LLMs) to real-world scenarios. While the latest LLMs are …
applying large language models (LLMs) to real-world scenarios. While the latest LLMs are …
MedChatZH: A tuning LLM for traditional Chinese medicine consultations
Abstract Generative Large Language Models (LLMs) have achieved significant success in
various natural language processing tasks, including Question-Answering (QA) and …
various natural language processing tasks, including Question-Answering (QA) and …
Marinegpt: Unlocking secrets of ocean to the public
Large language models (LLMs), such as ChatGPT/GPT-4, have proven to be powerful tools
in promoting the user experience as an AI assistant. The continuous works are proposing …
in promoting the user experience as an AI assistant. The continuous works are proposing …