Nvidia's KV Cache Transform Coding (KVTC) compresses LLM key-value cache by 20x without model changes, cutting GPU memory ...
Google has published TurboQuant, a KV cache compression algorithm that cuts LLM memory usage by 6x with zero accuracy loss, ...
Cache, in its crude definition, is a faster memory which stores copies of data from frequently used main memory locations. Nowadays, multiprocessor systems are supporting shared memories in hardware, ...