Machine-generated text: A comprehensive survey of threat models and detection methods

EN Crothers, N Japkowicz, HL Viktor - IEEE Access, 2023 - ieeexplore.ieee.org
Machine-generated text is increasingly difficult to distinguish from text authored by humans.
Powerful open-source models are freely available, and user-friendly tools that democratize …

Neural machine translation for low-resource languages: A survey

S Ranathunga, ESA Lee, M Prifti Skenduli… - ACM Computing …, 2023 - dl.acm.org
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since
the early 2000s and has already entered a mature phase. While considered the most widely …

The flan collection: Designing data and methods for effective instruction tuning

S Longpre, L Hou, T Vu, A Webson… - International …, 2023 - proceedings.mlr.press
We study the design decision of publicly available instruction tuning methods, by
reproducing and breaking down the development of Flan 2022 (Chung et al., 2022) …

ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health

L De Angelis, F Baglivo, G Arzilli, GP Privitera… - Frontiers in public …, 2023 - frontiersin.org
Large Language Models (LLMs) have recently gathered attention with the release of
ChatGPT, a user-centered chatbot released by OpenAI. In this perspective article, we retrace …

Exploiting programmatic behavior of llms: Dual-use through standard security attacks

D Kang, X Li, I Stoica, C Guestrin… - 2024 IEEE Security …, 2024 - ieeexplore.ieee.org
Recent advances in instruction-following large language models (LLMs) have led to
dramatic improvements in a range of NLP tasks. Unfortunately, we find that the same …

Red teaming language models to reduce harms: Methods, scaling behaviors, and lessons learned

D Ganguli, L Lovitt, J Kernion, A Askell, Y Bai… - arXiv preprint arXiv …, 2022 - arxiv.org
We describe our early efforts to red team language models in order to simultaneously
discover, measure, and attempt to reduce their potentially harmful outputs. We make three …

Lamda: Language models for dialog applications

R Thoppilan, D De Freitas, J Hall, N Shazeer… - arXiv preprint arXiv …, 2022 - arxiv.org
We present LaMDA: Language Models for Dialog Applications. LaMDA is a family of
Transformer-based neural language models specialized for dialog, which have up to 137B …

Ignore previous prompt: Attack techniques for language models

F Perez, I Ribeiro - arXiv preprint arXiv:2211.09527, 2022 - arxiv.org
Transformer-based large language models (LLMs) provide a powerful foundation for natural
language tasks in large-scale customer-facing applications. However, studies that explore …

Truthfulqa: Measuring how models mimic human falsehoods

S Lin, J Hilton, O Evans - arXiv preprint arXiv:2109.07958, 2021 - arxiv.org
We propose a benchmark to measure whether a language model is truthful in generating
answers to questions. The benchmark comprises 817 questions that span 38 categories …

The gradient of generative AI release: Methods and considerations

I Solaiman - Proceedings of the 2023 ACM conference on fairness …, 2023 - dl.acm.org
As increasingly powerful generative AI systems are developed, the release method greatly
varies. We propose a framework to assess six levels of access to generative AI systems: fully …