Q-Filters: Leveraging QK Geometry for Efficient KV Cache Compression
Abstract
Autoregressive language models rely on a Key-Value (KV) Cache, which avoids re-computing past hidden states during generation, making it faster. As model sizes and context lengths grow, the KV Cache becomes a significant memory bottleneck, which calls for compression methods that limit its size during generation. In this paper, we discover surprising properties of Query (Q) and Key (K) vectors that allow us to efficiently approximate attention scores without computing the attention maps. We propose Q-Filters, a training-free KV Cache compression method that filters out less crucial Key-Value pairs based on a single context-agnostic projection. Contrarily to many alternatives, Q-Filters is compatible with FlashAttention, as it does not require direct access to attention weights. Experimental results in long-context settings demonstrate that Q-Filters is competitive with attention-based compression methods such as SnapKV in retrieval tasks while consistently outperforming efficient compression schemes such as Streaming-LLM in generation setups. Notably, Q-Filters achieves a 99% accuracy in the needle-in-a-haystack task with a x32 compression level while reducing the generation perplexity drop by up to 65% in text generation compared to Streaming-LLM.
Community
An efficient KV cache compression technique that learns context-agnostic filters and improves on retrieval tasks while being compatible with FlashAttention. It outperforms efficient counterparts in both retrieval tasks and generation setup, reducing the perplexity gap induced by compression by up to 65%.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- TreeKV: Smooth Key-Value Cache Compression with Tree Structures (2025)
- RocketKV: Accelerating Long-Context LLM Inference via Two-Stage KV Cache Compression (2025)
- FastKV: KV Cache Compression for Fast Long-Context Processing with Token-Selective Propagation (2025)
- ChunkKV: Semantic-Preserving KV Cache Compression for Efficient Long-Context LLM Inference (2025)
- MPIC: Position-Independent Multimodal Context Caching System for Efficient MLLM Serving (2025)
- HACK: Homomorphic Acceleration via Compression of the Key-Value Cache for Disaggregated LLM Inference (2025)
- AKVQ-VL: Attention-Aware KV Cache Adaptive 2-Bit Quantization for Vision-Language Models (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper