Ask a Question

Prefer a chat interface with context about you and your work?

Retrieval Head Mechanistically Explains Long-Context Factuality

Retrieval Head Mechanistically Explains Long-Context Factuality

Despite the recent progress in long-context language models, it remains elusive how transformer-based models exhibit the capability to retrieve relevant information from arbitrary locations within the long context. This paper aims to address this question. Our systematic investigation across a wide spectrum of models reveals that a special type of …