Small models are valuable plug-ins for large language models
Large language models (LLMs) such as GPT-3 and GPT-4 are powerful but their weights are
often publicly unavailable and their immense sizes make the models difficult to be tuned with …
often publicly unavailable and their immense sizes make the models difficult to be tuned with …
Multilingual machine translation with large language models: Empirical results and analysis
Large language models (LLMs) have demonstrated remarkable potential in handling
multilingual machine translation (MMT). In this paper, we systematically investigate the …
multilingual machine translation (MMT). In this paper, we systematically investigate the …
Os-copilot: Towards generalist computer agents with self-improvement
Autonomous interaction with the computer has been a longstanding challenge with great
potential, and the recent proliferation of large language models (LLMs) has markedly …
potential, and the recent proliferation of large language models (LLMs) has markedly …
MetaAdapt: Domain adaptive few-shot misinformation detection via meta learning
With emerging topics (eg, COVID-19) on social media as a source for the spreading
misinformation, overcoming the distributional shifts between the original training domain (ie …
misinformation, overcoming the distributional shifts between the original training domain (ie …
Forward-backward reasoning in large language models for mathematical verification
Self-Consistency samples diverse reasoning chains with answers and chooses the final
answer by majority voting. It is based on forward reasoning and cannot further improve …
answer by majority voting. It is based on forward reasoning and cannot further improve …
Chef: A comprehensive evaluation framework for standardized assessment of multimodal large language models
Multimodal Large Language Models (MLLMs) have shown impressive abilities in interacting
with visual content with myriad potential downstream tasks. However, even though a list of …
with visual content with myriad potential downstream tasks. However, even though a list of …
LLMeBench: A flexible framework for accelerating llms benchmarking
The recent development and success of Large Language Models (LLMs) necessitate an
evaluation of their performance across diverse NLP tasks in different languages. Although …
evaluation of their performance across diverse NLP tasks in different languages. Although …
Language versatilists vs. specialists: An empirical revisiting on multilingual transfer ability
Multilingual transfer ability, which reflects how well the models fine-tuned on one source
language can be applied to other languages, has been well studied in multilingual pre …
language can be applied to other languages, has been well studied in multilingual pre …
In-context demonstration selection with cross entropy difference
Large language models (LLMs) can use in-context demonstrations to improve performance
on zero-shot tasks. However, selecting the best in-context examples is challenging because …
on zero-shot tasks. However, selecting the best in-context examples is challenging because …
EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling
Neural language models are probabilistic models of human text. They are predominantly
trained using maximum likelihood estimation (MLE), which is equivalent to minimizing the …
trained using maximum likelihood estimation (MLE), which is equivalent to minimizing the …