File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

백웅기

Baek, Woongki
Intelligent System Software Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 94 -
dc.citation.startPage 81 -
dc.citation.title MICROPROCESSORS AND MICROSYSTEMS -
dc.citation.volume 43 -
dc.contributor.author Kim, Kyu Yeun -
dc.contributor.author Baek, Woongki -
dc.date.accessioned 2023-12-21T23:39:46Z -
dc.date.available 2023-12-21T23:39:46Z -
dc.date.created 2016-07-08 -
dc.date.issued 2016-06 -
dc.description.abstract To achieve higher performance and energy efficiency, GPGPU architectures have recently begun to employ hardware caches. Adding caches to GPGPUs, however, does not always guarantee improved performance and energy efficiency due to the thrashing in small caches shared by thousands of threads. While prior work has proposed warp-scheduling and cache-bypassing techniques to address this issue, relatively little work has been done in the context of advanced cache indexing (ACI). To bridge this gap, this work investigates the effectiveness of ACI for high-performance and energy efficient GPGPU computing. We discuss the design and implementation of static and adaptive cache indexing schemes for GPGPUs. We then quantify the effectiveness of the ACI schemes based on a cycle accurate GPGPU simulator. Our quantitative evaluation demonstrates that the ACI schemes are effective in that they provide significant performance and energy-efficiency gains over the conventional indexing scheme. Further, we investigate the performance sensitivity of ACI to key architectural parameters (e.g., indexing latency and cache associativity). Our experimental results show that the ACI schemes are promising in that they continue to provide significant performance gains even when additional indexing latency occurs due to the hardware complexity and the baseline cache is enhanced with high associativity or large capacity. (C) 2016 Elsevier B.V. All rights reserved -
dc.identifier.bibliographicCitation MICROPROCESSORS AND MICROSYSTEMS, v.43, pp.81 - 94 -
dc.identifier.doi 10.1016/j.micpro.2016.01.003 -
dc.identifier.issn 0141-9331 -
dc.identifier.scopusid 2-s2.0-84992310705 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/19984 -
dc.identifier.url http://www.sciencedirect.com/science/article/pii/S0141933116000053 -
dc.identifier.wosid 000377740500008 -
dc.language 영어 -
dc.publisher ELSEVIER SCIENCE BV -
dc.title Quantifying the performance and energy efficiency of advanced cache indexing for GPGPU computing -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Computer Science, Hardware & Architecture; Computer Science, Theory & Methods; Engineering, Electrical & Electronic -
dc.relation.journalResearchArea Computer Science; Engineering -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Advanced cache indexing -
dc.subject.keywordAuthor GPGPU computing -
dc.subject.keywordAuthor High performance -
dc.subject.keywordAuthor Energy efficiency -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.