Farewell to Length Extrapolation, a Training-Free Infinite Context with
Finite Attention Scope
Farewell to Length Extrapolation, a Training-Free Infinite Context with
Finite Attention Scope
The maximum supported context length is a critical bottleneck limiting the practical application of the Large Language Model (LLM). Although existing length extrapolation methods can extend the context of LLMs to millions of tokens, these methods all have an explicit upper bound. In this work, we propose LongCache, a training-free …