Combating misinformation in the age of llms: Opportunities and challenges
Misinformation such as fake news and rumors is a serious threat for information ecosystems
and public trust. The emergence of large language models (LLMs) has great potential to …
and public trust. The emergence of large language models (LLMs) has great potential to …
Sora: A review on background, technology, limitations, and opportunities of large vision models
Sora is a text-to-video generative AI model, released by OpenAI in February 2024. The
model is trained to generate videos of realistic or imaginative scenes from text instructions …
model is trained to generate videos of realistic or imaginative scenes from text instructions …
Trustllm: Trustworthiness in large language models
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …
attention for their excellent natural language processing capabilities. Nonetheless, these …
Low-resource languages jailbreak gpt-4
AI safety training and red-teaming of large language models (LLMs) are measures to
mitigate the generation of unsafe content. Our work exposes the inherent cross-lingual …
mitigate the generation of unsafe content. Our work exposes the inherent cross-lingual …
Salad-bench: A hierarchical and comprehensive safety benchmark for large language models
In the rapidly evolving landscape of Large Language Models (LLMs), ensuring robust safety
measures is paramount. To meet this crucial need, we propose\emph {SALAD-Bench}, a …
measures is paramount. To meet this crucial need, we propose\emph {SALAD-Bench}, a …
[HTML][HTML] Position: TrustLLM: Trustworthiness in Large Language Models
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …
natural language processing capabilities. Nonetheless, these LLMs present many …
Factuality challenges in the era of large language models and opportunities for fact-checking
I Augenstein, T Baldwin, M Cha… - Nature Machine …, 2024 - nature.com
The emergence of tools based on large language models (LLMs), such as OpenAI's
ChatGPT and Google's Gemini, has garnered immense public attention owing to their …
ChatGPT and Google's Gemini, has garnered immense public attention owing to their …
On large language models' resilience to coercive interrogation
Abstract Large Language Models (LLMs) are increasingly employed in numerous
applications. It is hence important to ensure that their ethical standard aligns with humans' …
applications. It is hence important to ensure that their ethical standard aligns with humans' …
Sorry-bench: Systematically evaluating large language model safety refusal behaviors
Evaluating aligned large language models'(LLMs) ability to recognize and reject unsafe user
requests is crucial for safe, policy-compliant deployments. Existing evaluation efforts …
requests is crucial for safe, policy-compliant deployments. Existing evaluation efforts …
The art of saying no: Contextual noncompliance in language models
Chat-based language models are designed to be helpful, yet they should not comply with
every user request. While most existing work primarily focuses on refusal of" unsafe" …
every user request. While most existing work primarily focuses on refusal of" unsafe" …