Event knowledge in large language models: the gap between the impossible and the unlikely
Word co‐occurrence patterns in language corpora contain a surprising amount of
conceptual knowledge. Large language models (LLMs), trained to predict words in context …
conceptual knowledge. Large language models (LLMs), trained to predict words in context …
Robust Pronoun Fidelity with English LLMs: Are they Reasoning, Repeating, or Just Biased?
Robust, faithful, and harm-free pronoun use for individuals is an important goal for language
model development as their use increases, but prior work tends to study only one or two of …
model development as their use increases, but prior work tends to study only one or two of …
Back to square one: Artifact detection, training and commonsense disentanglement in the winograd schema
The Winograd Schema (WS) has been proposed as a test for measuring commonsense
capabilities of models. Recently, pre-trained language model-based approaches have …
capabilities of models. Recently, pre-trained language model-based approaches have …
BRAINTEASER: Lateral Thinking Puzzles for Large Language Model
The success of language models has inspired the NLP community to attend to tasks that
require implicit and complex reasoning, relying on human-like commonsense mechanisms …
require implicit and complex reasoning, relying on human-like commonsense mechanisms …
[HTML][HTML] Testing the limits of natural language models for predicting human language judgements
T Golan, M Siegelman, N Kriegeskorte… - Nature Machine …, 2023 - nature.com
Neural network language models appear to be increasingly aligned with how humans
process and generate language, but identifying their weaknesses through adversarial …
process and generate language, but identifying their weaknesses through adversarial …
Adversarial attack against cross-lingual knowledge graph alignment
Recent literatures have shown that knowledge graph (KG) learning models are highly
vulnerable to adversarial attacks. However, there is still a paucity of vulnerability analyses of …
vulnerable to adversarial attacks. However, there is still a paucity of vulnerability analyses of …
Robustness and adversarial examples in natural language processing
Recent studies show that many NLP systems are sensitive and vulnerable to a small
perturbation of inputs and do not generalize well across different datasets. This lack of …
perturbation of inputs and do not generalize well across different datasets. This lack of …
Hierarchical interpretation of neural text classification
Recent years have witnessed increasing interest in developing interpretable models in
Natural Language Processing (NLP). Most existing models aim at identifying input features …
Natural Language Processing (NLP). Most existing models aim at identifying input features …
Generalized quantifiers as a source of error in multilingual NLU benchmarks
Logical approaches to representing language have developed and evaluated computational
models of quantifier words since the 19th century, but today's NLU models still struggle to …
models of quantifier words since the 19th century, but today's NLU models still struggle to …
A semantic-based method for unsupervised commonsense question answering
Unsupervised commonsense question answering is appealing since it does not rely on any
labeled task data. Among existing work, a popular solution is to use pre-trained language …
labeled task data. Among existing work, a popular solution is to use pre-trained language …