Hackernews
new
show
ask
jobs
KV Cache Transform Coding for Compact Storage in LLM Inference
2 points
posted 3 hours ago
by walterbell
(arxiv.org)
No comments yet