A space-efficient on-chip compressed cache organization for high performance computing

Keun Soo Yim, Jang Soo Lee, Jihong Kim, Shin-Dug Kim, Kern Koh

Research output: Contribution to journalArticle

1 Citation (Scopus)

Abstract

In order to alleviate the ever-increasing processor-memory performance gap of high-end parallel computers, on-chip compressed caches have been developed that can reduce the cache miss count and off-chip memory traffic by storing and transferring cache lines in a compressed form. However, we observed that their performance gain is often limited due to their use of the coarse-grained compressed cache line management which incurs internally fragmented space. In this paper, we present the fine-grained compressed cache line management which addresses the fragmentation problem, while avoiding an increase in the metadata size such as tag field and VM page table. Based on the SimpleScalar simulator with the SPEC benchmark suite, we show that over an existing compressed cache system the proposed cache organization can reduce the memory traffic by 15%, as it delivers compressed cache lines in a fine-grained way, and the cache miss count by 23%, as it stores up to three compressed cache lines in a physical cache line.

Original languageEnglish
Pages (from-to)952-964
Number of pages13
JournalLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume3358
Publication statusPublished - 2004 Dec 1

Fingerprint

Cache
Chip
High Performance
Data storage equipment
Computing
Metadata
Line
Simulators
Count
Traffic
Parallel Computers
Fragmentation
Table
Simulator
Benchmark

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • Computer Science(all)

Cite this

@article{9d16e6f8b7dc48d9b8413d460aac2ba9,
title = "A space-efficient on-chip compressed cache organization for high performance computing",
abstract = "In order to alleviate the ever-increasing processor-memory performance gap of high-end parallel computers, on-chip compressed caches have been developed that can reduce the cache miss count and off-chip memory traffic by storing and transferring cache lines in a compressed form. However, we observed that their performance gain is often limited due to their use of the coarse-grained compressed cache line management which incurs internally fragmented space. In this paper, we present the fine-grained compressed cache line management which addresses the fragmentation problem, while avoiding an increase in the metadata size such as tag field and VM page table. Based on the SimpleScalar simulator with the SPEC benchmark suite, we show that over an existing compressed cache system the proposed cache organization can reduce the memory traffic by 15{\%}, as it delivers compressed cache lines in a fine-grained way, and the cache miss count by 23{\%}, as it stores up to three compressed cache lines in a physical cache line.",
author = "Yim, {Keun Soo} and Lee, {Jang Soo} and Jihong Kim and Shin-Dug Kim and Kern Koh",
year = "2004",
month = "12",
day = "1",
language = "English",
volume = "3358",
pages = "952--964",
journal = "Lecture Notes in Computer Science",
issn = "0302-9743",
publisher = "Springer Verlag",

}

TY - JOUR

T1 - A space-efficient on-chip compressed cache organization for high performance computing

AU - Yim, Keun Soo

AU - Lee, Jang Soo

AU - Kim, Jihong

AU - Kim, Shin-Dug

AU - Koh, Kern

PY - 2004/12/1

Y1 - 2004/12/1

N2 - In order to alleviate the ever-increasing processor-memory performance gap of high-end parallel computers, on-chip compressed caches have been developed that can reduce the cache miss count and off-chip memory traffic by storing and transferring cache lines in a compressed form. However, we observed that their performance gain is often limited due to their use of the coarse-grained compressed cache line management which incurs internally fragmented space. In this paper, we present the fine-grained compressed cache line management which addresses the fragmentation problem, while avoiding an increase in the metadata size such as tag field and VM page table. Based on the SimpleScalar simulator with the SPEC benchmark suite, we show that over an existing compressed cache system the proposed cache organization can reduce the memory traffic by 15%, as it delivers compressed cache lines in a fine-grained way, and the cache miss count by 23%, as it stores up to three compressed cache lines in a physical cache line.

AB - In order to alleviate the ever-increasing processor-memory performance gap of high-end parallel computers, on-chip compressed caches have been developed that can reduce the cache miss count and off-chip memory traffic by storing and transferring cache lines in a compressed form. However, we observed that their performance gain is often limited due to their use of the coarse-grained compressed cache line management which incurs internally fragmented space. In this paper, we present the fine-grained compressed cache line management which addresses the fragmentation problem, while avoiding an increase in the metadata size such as tag field and VM page table. Based on the SimpleScalar simulator with the SPEC benchmark suite, we show that over an existing compressed cache system the proposed cache organization can reduce the memory traffic by 15%, as it delivers compressed cache lines in a fine-grained way, and the cache miss count by 23%, as it stores up to three compressed cache lines in a physical cache line.

UR - http://www.scopus.com/inward/record.url?scp=34547325630&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=34547325630&partnerID=8YFLogxK

M3 - Article

VL - 3358

SP - 952

EP - 964

JO - Lecture Notes in Computer Science

JF - Lecture Notes in Computer Science

SN - 0302-9743

ER -