Event knowledge in large language models: the gap between the impossible and the unlikely

C Kauf, AA Ivanova, G Rambelli, E Chersoni… - Cognitive …, 2023 - Wiley Online Library
Word co‐occurrence patterns in language corpora contain a surprising amount of
conceptual knowledge. Large language models (LLMs), trained to predict words in context …

Robust Pronoun Fidelity with English LLMs: Are they Reasoning, Repeating, or Just Biased?

V Gautam, E Bingert, D Zhu, A Lauscher… - Transactions of the …, 2024 - direct.mit.edu
Robust, faithful, and harm-free pronoun use for individuals is an important goal for language
model development as their use increases, but prior work tends to study only one or two of …

Back to square one: Artifact detection, training and commonsense disentanglement in the winograd schema

Y Elazar, H Zhang, Y Goldberg, D Roth - arXiv preprint arXiv:2104.08161, 2021 - arxiv.org
The Winograd Schema (WS) has been proposed as a test for measuring commonsense
capabilities of models. Recently, pre-trained language model-based approaches have …

BRAINTEASER: Lateral Thinking Puzzles for Large Language Model

Y Jiang, F Ilievski, K Ma - arXiv preprint arXiv:2310.05057, 2023 - arxiv.org
The success of language models has inspired the NLP community to attend to tasks that
require implicit and complex reasoning, relying on human-like commonsense mechanisms …

[HTML][HTML] Testing the limits of natural language models for predicting human language judgements

T Golan, M Siegelman, N Kriegeskorte… - Nature Machine …, 2023 - nature.com
Neural network language models appear to be increasingly aligned with how humans
process and generate language, but identifying their weaknesses through adversarial …

Adversarial attack against cross-lingual knowledge graph alignment

Z Zhang, Z Zhang, Y Zhou, L Wu, S Wu… - Proceedings of the …, 2021 - aclanthology.org
Recent literatures have shown that knowledge graph (KG) learning models are highly
vulnerable to adversarial attacks. However, there is still a paucity of vulnerability analyses of …

Robustness and adversarial examples in natural language processing

KW Chang, H He, R Jia, S Singh - Proceedings of the 2021 …, 2021 - aclanthology.org
Recent studies show that many NLP systems are sensitive and vulnerable to a small
perturbation of inputs and do not generalize well across different datasets. This lack of …

Hierarchical interpretation of neural text classification

H Yan, L Gui, Y He - Computational Linguistics, 2022 - direct.mit.edu
Recent years have witnessed increasing interest in developing interpretable models in
Natural Language Processing (NLP). Most existing models aim at identifying input features …

Generalized quantifiers as a source of error in multilingual NLU benchmarks

R Cui, D Hershcovich, A Søgaard - arXiv preprint arXiv:2204.10615, 2022 - arxiv.org
Logical approaches to representing language have developed and evaluated computational
models of quantifier words since the 19th century, but today's NLU models still struggle to …

A semantic-based method for unsupervised commonsense question answering

Y Niu, F Huang, J Liang, W Chen, X Zhu… - arXiv preprint arXiv …, 2021 - arxiv.org
Unsupervised commonsense question answering is appealing since it does not rely on any
labeled task data. Among existing work, a popular solution is to use pre-trained language …