Non-referenced prefetch (NRP) cache for instruction prefetching

G. H. Park, O. Y. Kwon, Tack-Don Han, Shin-Dug Kim

Research output: Contribution to journalArticle

1 Citation (Scopus)

Abstract

A new conceptual cache, NRP (non-referenced prefetch) cache, is proposed to improve the performance of instruction prefetch mechanisms which try to prefetch both the sequential and nonsequential blocks under the limited memory bandwidth. The NRP cache is used in storing prefetched blocks that were not referenced by the CPU, while these blocks were discarded in other previous prefetch mechanisms. By storing the non-referenced prefetch blocks in the NRP cache, both cache misses and memory traffic are reduced. A prefetch method to prefetch both the sequential and the nonsequential instruction paths is designed to utilise the effectiveness of the NRP cache. The results from trace-driven simulation show that this approach provides an improvement in memory access time than other prefetch methods. Particularly, the NRP cache is more effective in a lookahead prefetch mechanism that can hide longer memory latency. Also, the NRP cache reduces 50 - 112% of the additional memory traffic required to prefetch both instruction paths. This approach can achieve both the improved memory access time and the reduced memory traffic as a cost-effective cache design.

Original languageEnglish
Pages (from-to)37-43
Number of pages7
JournalIEE Proceedings: Computers and Digital Techniques
Volume143
Issue number1
DOIs
Publication statusPublished - 1996 Jan 1

Fingerprint

Prefetching
Cache
Data storage equipment
Traffic
Path
Look-ahead
Program processors
Long Memory
Latency
Bandwidth
Trace
Costs

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • Hardware and Architecture
  • Computational Theory and Mathematics

Cite this

@article{5620a92f43bc468cb38cacb18518fd83,
title = "Non-referenced prefetch (NRP) cache for instruction prefetching",
abstract = "A new conceptual cache, NRP (non-referenced prefetch) cache, is proposed to improve the performance of instruction prefetch mechanisms which try to prefetch both the sequential and nonsequential blocks under the limited memory bandwidth. The NRP cache is used in storing prefetched blocks that were not referenced by the CPU, while these blocks were discarded in other previous prefetch mechanisms. By storing the non-referenced prefetch blocks in the NRP cache, both cache misses and memory traffic are reduced. A prefetch method to prefetch both the sequential and the nonsequential instruction paths is designed to utilise the effectiveness of the NRP cache. The results from trace-driven simulation show that this approach provides an improvement in memory access time than other prefetch methods. Particularly, the NRP cache is more effective in a lookahead prefetch mechanism that can hide longer memory latency. Also, the NRP cache reduces 50 - 112{\%} of the additional memory traffic required to prefetch both instruction paths. This approach can achieve both the improved memory access time and the reduced memory traffic as a cost-effective cache design.",
author = "Park, {G. H.} and Kwon, {O. Y.} and Tack-Don Han and Shin-Dug Kim",
year = "1996",
month = "1",
day = "1",
doi = "10.1049/ip-cdt:19960010",
language = "English",
volume = "143",
pages = "37--43",
journal = "IEE Proceedings: Computers and Digital Techniques",
issn = "1350-2387",
publisher = "Institute of Electrical Engineers",
number = "1",

}

Non-referenced prefetch (NRP) cache for instruction prefetching. / Park, G. H.; Kwon, O. Y.; Han, Tack-Don; Kim, Shin-Dug.

In: IEE Proceedings: Computers and Digital Techniques, Vol. 143, No. 1, 01.01.1996, p. 37-43.

Research output: Contribution to journalArticle

TY - JOUR

T1 - Non-referenced prefetch (NRP) cache for instruction prefetching

AU - Park, G. H.

AU - Kwon, O. Y.

AU - Han, Tack-Don

AU - Kim, Shin-Dug

PY - 1996/1/1

Y1 - 1996/1/1

N2 - A new conceptual cache, NRP (non-referenced prefetch) cache, is proposed to improve the performance of instruction prefetch mechanisms which try to prefetch both the sequential and nonsequential blocks under the limited memory bandwidth. The NRP cache is used in storing prefetched blocks that were not referenced by the CPU, while these blocks were discarded in other previous prefetch mechanisms. By storing the non-referenced prefetch blocks in the NRP cache, both cache misses and memory traffic are reduced. A prefetch method to prefetch both the sequential and the nonsequential instruction paths is designed to utilise the effectiveness of the NRP cache. The results from trace-driven simulation show that this approach provides an improvement in memory access time than other prefetch methods. Particularly, the NRP cache is more effective in a lookahead prefetch mechanism that can hide longer memory latency. Also, the NRP cache reduces 50 - 112% of the additional memory traffic required to prefetch both instruction paths. This approach can achieve both the improved memory access time and the reduced memory traffic as a cost-effective cache design.

AB - A new conceptual cache, NRP (non-referenced prefetch) cache, is proposed to improve the performance of instruction prefetch mechanisms which try to prefetch both the sequential and nonsequential blocks under the limited memory bandwidth. The NRP cache is used in storing prefetched blocks that were not referenced by the CPU, while these blocks were discarded in other previous prefetch mechanisms. By storing the non-referenced prefetch blocks in the NRP cache, both cache misses and memory traffic are reduced. A prefetch method to prefetch both the sequential and the nonsequential instruction paths is designed to utilise the effectiveness of the NRP cache. The results from trace-driven simulation show that this approach provides an improvement in memory access time than other prefetch methods. Particularly, the NRP cache is more effective in a lookahead prefetch mechanism that can hide longer memory latency. Also, the NRP cache reduces 50 - 112% of the additional memory traffic required to prefetch both instruction paths. This approach can achieve both the improved memory access time and the reduced memory traffic as a cost-effective cache design.

UR - http://www.scopus.com/inward/record.url?scp=0029775520&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=0029775520&partnerID=8YFLogxK

U2 - 10.1049/ip-cdt:19960010

DO - 10.1049/ip-cdt:19960010

M3 - Article

AN - SCOPUS:0029775520

VL - 143

SP - 37

EP - 43

JO - IEE Proceedings: Computers and Digital Techniques

JF - IEE Proceedings: Computers and Digital Techniques

SN - 1350-2387

IS - 1

ER -