A survey on evaluation of large language models
Large language models (LLMs) are gaining increasing popularity in both academia and
industry, owing to their unprecedented performance in various applications. As LLMs …
industry, owing to their unprecedented performance in various applications. As LLMs …
Pandalm: An automatic evaluation benchmark for llm instruction tuning optimization
Instruction tuning large language models (LLMs) remains a challenging task, owing to the
complexity of hyperparameter selection and the difficulty involved in evaluating the tuned …
complexity of hyperparameter selection and the difficulty involved in evaluating the tuned …
Understanding and mitigating the label noise in pre-training on downstream tasks
Pre-training on large-scale datasets and then fine-tuning on downstream tasks have
become a standard practice in deep learning. However, pre-training data often contain label …
become a standard practice in deep learning. However, pre-training data often contain label …
Coderujb: An executable and unified java benchmark for practical programming scenarios
In the evolving landscape of large language models (LLMs) tailored for software
engineering, the need for benchmarks that accurately reflect real-world development …
engineering, the need for benchmarks that accurately reflect real-world development …
A hard-to-beat baseline for training-free clip-based adaptation
Contrastive Language-Image Pretraining (CLIP) has gained popularity for its remarkable
zero-shot capacity. Recent research has focused on developing efficient fine-tuning …
zero-shot capacity. Recent research has focused on developing efficient fine-tuning …
Parameter-efficient long-tailed recognition
The" pre-training and fine-tuning" paradigm in addressing long-tailed recognition tasks has
sparked significant interest since the emergence of large vision-language models like the …
sparked significant interest since the emergence of large vision-language models like the …
Novelqa: A benchmark for long-range novel question answering
The rapid advancement of Large Language Models (LLMs) has introduced a new frontier in
natural language processing, particularly in understanding and processing long-context …
natural language processing, particularly in understanding and processing long-context …
Learning with noisy foundation models
Foundation models are usually pre-trained on large-scale datasets and then adapted to
downstream tasks through tuning. However, the large-scale pre-training datasets, often …
downstream tasks through tuning. However, the large-scale pre-training datasets, often …
ZooPFL: Exploring black-box foundation models for personalized federated learning
When personalized federated learning (FL) meets large foundation models, new challenges
arise from various limitations in resources. In addition to typical limitations such as data …
arise from various limitations in resources. In addition to typical limitations such as data …
Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts
The fine-tuning paradigm in addressing long-tail learning tasks has sparked significant
interest since the emergence of foundation models. Nonetheless, how fine-tuning impacts …
interest since the emergence of foundation models. Nonetheless, how fine-tuning impacts …