Contextcite: Attributing model generation to context
How do language models use information provided as context when generating a
response? Can we infer whether a particular generated statement is actually grounded in …
response? Can we infer whether a particular generated statement is actually grounded in …
A primer on the inner workings of transformer-based language models
The rapid progress of research aimed at interpreting the inner workings of advanced
language models has highlighted a need for contextualizing the insights gained from years …
language models has highlighted a need for contextualizing the insights gained from years …
Unifying corroborative and contributive attributions in large language models
As businesses, products, and services spring up around large language models, the
trustworthiness of these models hinges on the verifiability of their outputs. However, methods …
trustworthiness of these models hinges on the verifiability of their outputs. However, methods …
Editable concept bottleneck models
Concept Bottleneck Models (CBMs) have garnered much attention for their ability to
elucidate the prediction process through a human-understandable concept layer. However …
elucidate the prediction process through a human-understandable concept layer. However …
Towards User-Focused Research in Training Data Attribution for Human-Centered Explainable AI
While Explainable AI (XAI) aims to make AI understandable and useful to humans, it has
been criticised for relying too much on formalism and solutionism, focusing more on …
been criticised for relying too much on formalism and solutionism, focusing more on …
Advances and open challenges in federated learning with foundation models
The integration of Foundation Models (FMs) with Federated Learning (FL) presents a
transformative paradigm in Artificial Intelligence (AI), offering enhanced capabilities while …
transformative paradigm in Artificial Intelligence (AI), offering enhanced capabilities while …
Reciprocal learning
We demonstrate that a wide array of machine learning algorithms are specific instances of
one single paradigm: reciprocal learning. These instances range from active learning over …
one single paradigm: reciprocal learning. These instances range from active learning over …
What is Your Data Worth to GPT? LLM-Scale Data Valuation with Influence Functions
Large language models (LLMs) are trained on a vast amount of human-written data, but data
providers often remain uncredited. In response to this issue, data valuation (or data …
providers often remain uncredited. In response to this issue, data valuation (or data …
Enhancing Data Quality in Federated Fine-Tuning of Foundation Models
In the current landscape of foundation model training, there is a significant reliance on public
domain data, which is nearing exhaustion according to recent research. To further scale up …
domain data, which is nearing exhaustion according to recent research. To further scale up …
Do Influence Functions Work on Large Language Models?
Influence functions aim to quantify the impact of individual training data points on a model's
predictions. While extensive research has been conducted on influence functions in …
predictions. While extensive research has been conducted on influence functions in …