Training-Free LLM Context Compression with Hybrid Graph Priors
So, there’s this new method for compressing context in large language models that doesn’t require any training and works with any model. It creates a unique sparse hybrid graph that combines semantic links with short-range connections. By clustering, it finds a topic skeleton and then ranks sentences using a clear scoring system that looks at how relevant they are to the task, how well they represent clusters, their centrality, and a cue for cycle coverage. The aim is to keep sentences relevant, cover the topic well, and ensure coherence, all while sticking to a strict token limit. This method is efficient and can be used with different long-context LLMs.
Key facts
- Framework is training-free and model-agnostic
- Uses hybrid sentence graph with k-NN and sequential edges
- Extracts topic skeleton via clustering
- Ranks sentences with interpretable score
- Addresses limitations of existing compression methods
- Preserves task relevance, topic coverage, and coherence
- Operates under strict token budget
- Designed for long-context LLMs
Entities
—