69
citations
#417
in ICLR 2024
of 2297 papers
2
Top Authors
4
Data Points
Top Authors
Abstract
Language models (LMs) can recall facts mentioned in context, as shown by their performance on reading comprehension tasks. When the context describes facts about more than one entity, the LM has to correctly bind attributes to their corresponding entity. We show, via causal experiments, that LMs' internal activations represent binding information by exhibiting appropriate binding ID vectors at the entity and attribute positions. We further show that binding ID vectors form a subspace and often transfer across tasks. Our results demonstrate that LMs learn interpretable strategies for representing symbolic knowledge in context, and that studying context activations is a fruitful direction for understanding LM cognition.
Citation History
Jan 28, 2026
0
Feb 13, 2026
69+69
Feb 13, 2026
69
Feb 13, 2026
69