How do Language Models Bind Entities in Context?

69citations
arXiv:2310.17191
69
citations
#417
in ICLR 2024
of 2297 papers
2
Top Authors
4
Data Points

Abstract

Language models (LMs) can recall facts mentioned in context, as shown by their performance on reading comprehension tasks. When the context describes facts about more than one entity, the LM has to correctly bind attributes to their corresponding entity. We show, via causal experiments, that LMs' internal activations represent binding information by exhibiting appropriate binding ID vectors at the entity and attribute positions. We further show that binding ID vectors form a subspace and often transfer across tasks. Our results demonstrate that LMs learn interpretable strategies for representing symbolic knowledge in context, and that studying context activations is a fruitful direction for understanding LM cognition.

Citation History

Jan 28, 2026
0
Feb 13, 2026
69+69
Feb 13, 2026
69
Feb 13, 2026
69