LookaheadKV: Fast and Accurate KV Cache Eviction by Glimpsing into the Future without Generation
Abstract
LookaheadKV is a lightweight framework that improves KV cache eviction in transformers by using parameter-efficient modules to predict importance scores without draft generation, achieving faster inference with reduced eviction costs.
Transformer-based large language models (LLMs) rely on key-value (KV) caching to avoid redundant computation during autoregressive inference. While this mechanism greatly improves efficiency, the cache size grows linearly with the input sequence length, quickly becoming a bottleneck for long-context tasks. Existing solutions mitigate this problem by evicting prompt KV that are deemed unimportant, guided by estimated importance scores. Notably, a recent line of work proposes to improve eviction quality by "glimpsing into the future", in which a draft generator produces a surrogate future response approximating the target model's true response, and this surrogate is subsequently used to estimate the importance of cached KV more accurately. However, these approaches rely on computationally expensive draft generation, which introduces substantial prefilling overhead and limits their practicality in real-world deployment. To address this challenge, we propose LookaheadKV, a lightweight eviction framework that leverages the strength of surrogate future response without requiring explicit draft generation. LookaheadKV augments transformer layers with parameter-efficient modules trained to predict true importance scores with high accuracy. Our design ensures negligible runtime overhead comparable to existing inexpensive heuristics, while achieving accuracy superior to more costly approximation methods. Extensive experiments on long-context understanding benchmarks, across a wide range of models, demonstrate that our method not only outperforms recent competitive baselines in various long-context understanding tasks, but also reduces the eviction cost by up to 14.5x, leading to significantly faster time-to-first-token. Our code is available at https://github.com/SamsungLabs/LookaheadKV.
Community
- Fast and Accurate KV Cache Eviction: LookaheadKV is a novel method employing lookahead modules to predict future token importance.
- LookaheadKV shows a performance drop less than 0.7% with 1K cache budget (~81% compression).
- LookaheadKV reduces eviction cost by up to 14.5×, leading to significantly faster TTFT.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Fast KVzip: Efficient and Accurate LLM Inference with Gated KV Eviction (2026)
- ForesightKV: Optimizing KV Cache Eviction for Reasoning Models by Learning Long-Term Contribution (2026)
- LongFlow: Efficient KV Cache Compression for Reasoning M (2026)
- Where Matters More Than What: Decoding-aligned KV Cache Compression via Position-aware Pseudo Queries (2026)
- Predicting Future Utility: Global Combinatorial Optimization for Task-Agnostic KV Cache Eviction (2026)
- ARKV: Adaptive and Resource-Efficient KV Cache Management under Limited Memory Budget for Long-Context Inference in LLMs (2026)
- DeltaKV: Residual-Based KV Cache Compression via Long-Range Similarity (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper