Fast KV Compaction via Attention Matching
63 points by cbracketdash 19 hours ago | 15 comments
cadamsdotcom 12 hours ago
Superficially it sounds like this could create a bit more of a move toward doing compaction on some continuous basis, or compacting in batches once you hit the context limit, rather than starting fresh with a summary and system prompt..
replyFeels like high fidelity, fast compaction could be a path to “solving” long context.
Also, I don't see why you couldn't patent this if you wanted to monetize it.
We all just saw the prior art published for the public. That will preclude patenting this work. Further reduction to practice is required.
(I am not a lawyer).
The reality is that the money being thrown = the time of humans. I guess compute as well, but in terms of people doing innovation - openly published things are the same thing, minus the money.
I don’t think it will last among researchers who think beyond production LLMs