Qwen technical report

J Bai, S Bai, Y Chu, Z Cui, K Dang, X Deng… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have revolutionized the field of artificial intelligence,
enabling natural language processing tasks that were previously thought to be exclusive to …

Yi: Open foundation models by 01. ai

A Young, B Chen, C Li, C Huang, G Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
We introduce the Yi model family, a series of language and multimodal models that
demonstrate strong multi-dimensional capabilities. The Yi model family is based on 6B and …

Promptbench: A unified library for evaluation of large language models

K Zhu, Q Zhao, H Chen, J Wang, X Xie - Journal of Machine Learning …, 2024 - jmlr.org
The evaluation of large language models (LLMs) is crucial to assess their performance and
mitigate potential security risks. In this paper, we introduce PromptBench, a unified library to …

Chatkbqa: A generate-then-retrieve framework for knowledge base question answering with fine-tuned large language models

H Luo, Z Tang, S Peng, Y Guo, W Zhang, C Ma… - arXiv preprint arXiv …, 2023 - arxiv.org
Knowledge Base Question Answering (KBQA) aims to answer natural language questions
over large-scale knowledge bases (KBs), which can be summarized into two crucial steps …

Large language model agent in financial trading: A survey

H Ding, Y Li, J Wang, H Chen - arXiv preprint arXiv:2408.06361, 2024 - arxiv.org
Trading is a highly competitive task that requires a combination of strategy, knowledge, and
psychological fortitude. With the recent success of large language models (LLMs), it is …

Qilin-med: Multi-stage knowledge injection advanced medical large language model

Q Ye, J Liu, D Chong, P Zhou, Y Hua, F Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
Integrating large language models (LLMs) into healthcare holds great potential but faces
challenges. Pre-training LLMs from scratch for domains like medicine is resource-heavy and …

The dawn after the dark: An empirical study on factuality hallucination in large language models

J Li, J Chen, R Ren, X Cheng, WX Zhao, JY Nie… - arXiv preprint arXiv …, 2024 - arxiv.org
In the era of large language models (LLMs), hallucination (ie, the tendency to generate
factually incorrect content) poses great challenge to trustworthy and reliable deployment of …

Knowledge unlearning for llms: Tasks, methods, and challenges

N Si, H Zhang, H Chang, W Zhang, D Qu… - arXiv preprint arXiv …, 2023 - arxiv.org
In recent years, large language models (LLMs) have spurred a new research paradigm in
natural language processing. Despite their excellent capability in knowledge-based …

Model spider: Learning to rank pre-trained models efficiently

YK Zhang, TJ Huang, YX Ding… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Figuring out which Pre-Trained Model (PTM) from a model zoo fits the target task is
essential to take advantage of plentiful model resources. With the availability of numerous …

A Comprehensive Survey of Datasets, Theories, Variants, and Applications in Direct Preference Optimization

W Xiao, Z Wang, L Gan, S Zhao, W He, LA Tuan… - arXiv preprint arXiv …, 2024 - arxiv.org
With the rapid advancement of large language models (LLMs), aligning policy models with
human preferences has become increasingly critical. Direct Preference Optimization (DPO) …