Akademska digitalna zbirka SLovenije - logo
E-viri
  • Buddy compression
    Choukse, Esha; Sullivan, Michael B.; O'Connor, Mike; Erez, Mattan; Pool, Jeff; Nellans, David; Keckler, Stephen W.

    2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA), 05/2020
    Conference Proceeding

    GPUs accelerate high-throughput applications, which require orders-of-magnitude higher memory bandwidth than traditional CPU-only systems. However, the capacity of such high-bandwidth memory tends to be relatively small. Buddy Compression is an architecture that makes novel use of compression to utilize a larger buddy-memory from the host or disaggregated memory, effectively increasing the memory capacity of the GPU. Buddy Compression splits each compressed 128B memory-entry between the high-bandwidth GPU memory and a slower-but-larger buddy memory such that compressible memory-entries are accessed completely from GPU memory, while incompressible entries source some of their data from off-GPU memory. With Buddy Compression, compressibility changes never result in expensive page movement or re-allocation. Buddy Compression achieves on average 1.9x effective GPU memory expansion for representative HPC applications and 1.5x for deep learning training, performing within 2% of an unrealistic system with no memory limit. This makes Buddy Compression attractive for performance-conscious developers that require additional GPU memory capacity.