Ask a Question

Prefer a chat interface with context about you and your work?

Farewell to Length Extrapolation, a Training-Free Infinite Context with Finite Attention Scope

Farewell to Length Extrapolation, a Training-Free Infinite Context with Finite Attention Scope

The maximum supported context length is a critical bottleneck limiting the practical application of the Large Language Model (LLM). Although existing length extrapolation methods can extend the context of LLMs to millions of tokens, these methods all have an explicit upper bound. In this work, we propose LongCache, a training-free …