An overview on language models: Recent developments and outlook

C Wei, YC Wang, B Wang, CCJ Kuo - arXiv preprint arXiv:2303.05759, 2023 - arxiv.org
Language modeling studies the probability distributions over strings of texts. It is one of the
most fundamental tasks in natural language processing (NLP). It has been widely used in …

Adversarial attacks and defenses in explainable artificial intelligence: A survey

H Baniecki, P Biecek - Information Fusion, 2024 - Elsevier
Explainable artificial intelligence (XAI) methods are portrayed as a remedy for debugging
and trusting statistical and deep learning models, as well as interpreting their predictions …

Beyond accuracy: Behavioral testing of NLP models with CheckList

MT Ribeiro, T Wu, C Guestrin, S Singh - arXiv preprint arXiv:2005.04118, 2020 - arxiv.org
Although measuring held-out accuracy has been the primary approach to evaluate
generalization, it often overestimates the performance of NLP models, while alternative …

Robust natural language processing: Recent advances, challenges, and future directions

M Omar, S Choi, DH Nyang, D Mohaisen - IEEE Access, 2022 - ieeexplore.ieee.org
Recent natural language processing (NLP) techniques have accomplished high
performance on benchmark data sets, primarily due to the significant improvement in the …

Benchmarking robustness of adaptation methods on pre-trained vision-language models

S Chen, J Gu, Z Han, Y Ma, P Torr… - Advances in Neural …, 2024 - proceedings.neurips.cc
Various adaptation methods, such as LoRA, prompts, and adapters, have been proposed to
enhance the performance of pre-trained vision-language models in specific domains. As test …

Are multimodal models robust to image and text perturbations?

J Qiu, Y Zhu, X Shi, F Wenzel, Z Tang, D Zhao… - arXiv preprint arXiv …, 2022 - arxiv.org
Multimodal image-text models have shown remarkable performance in the past few years.
However, evaluating their robustness against distribution shifts is crucial before adopting …

Can ChatGPT replace traditional KBQA models? An in-depth analysis of the question answering performance of the GPT LLM family

Y Tan, D Min, Y Li, W Li, N Hu, Y Chen, G Qi - International Semantic Web …, 2023 - Springer
ChatGPT is a powerful large language model (LLM) that covers knowledge resources such
as Wikipedia and supports natural language question answering using its own knowledge …

Information extraction from text intensive and visually rich banking documents

B Oral, E Emekligil, S Arslan, G Eryiǧit - Information Processing & …, 2020 - Elsevier
Document types, where visual and textual information plays an important role in their
analysis and understanding, pose a new and attractive area for information extraction …

An empirical study of testing machine learning in the wild

M Openja, F Khomh, A Foundjem, ZM Jiang… - ACM Transactions on …, 2024 - dl.acm.org
Background: Recently, machine and deep learning (ML/DL) algorithms have been
increasingly adopted in many software systems. Due to their inductive nature, ensuring the …

Understanding model robustness to user-generated noisy texts

J Náplava, M Popel, M Straka, J Straková - arXiv preprint arXiv …, 2021 - arxiv.org
Sensitivity of deep-neural models to input noise is known to be a challenging problem. In
NLP, model performance often deteriorates with naturally occurring noise, such as spelling …